Abstract?Objective: Key issues in the epilepsy seizure prediction research are (1) the reproducibility of results (2) the\r\ninability to compare multiple approaches directly. To overcome these problems, the Seizure Prediction Challenge was organized on Kaggle.com. It aimed at establishing benchmarks on a dataset with predefined train, validation and test sets. Our main objective is to analyse the competition format, and to propose improvements, which would facilitate a better comparison of algorithms. The second objective is to present a novel deep learning approach to seizure prediction and compare it to other commonly used methods using patient centered metrics. Methods: We used the competition?s datasets to illustrate the effects of data contamination. Having better data partitions, we compared three types of models in terms of different objectives. Results: We found that correct selection of test samples is crucial when evaluating\r\nthe performance of seizure forecasting models. Moreover, we showed that models, which achieve state-of-the-art performance with respect to commonly used AUC, sensitivity and specificity metrics, may not yet be suitable for practical usage because of low precision scores. Conclusion: Correlation between validation and test datasets used in the competition limited its scientific value. Significance: Our findings provide guidelines which allow for a more objective evaluation of seizure prediction models.