Rafi, Abdul Muntakim, Nogina, Daria, Penzar, Dmitry, Lee, Dohoon, Lee, Danyeong, Kim, Nayeon, Kim, Sangyeup, Kim, Dohyeon, Shin, Yeojin, Kwak, Il-Youp, Meshcheryakov, Georgy, Lando, Andrey, Zinkevich, Arsenii, Kim, Byeong-Chan, Lee, Juhyun, Kang, Taein, Vaishnav, Eeshit Dhaval, Yadollahpour, Payman, Random Promoter DREAM Challenge Consortium, Kim, Sun, Albrecht, Jake, Regev, Aviv, Gong, Wuming, Kulakovskiy, Ivan V, Meyer, Pablo, de Boer, Carl G (October 2024) A community effort to optimize sequence-based deep learning models of gene regulation. Nature Biotechnology. ISSN 1087-0156 (Public Dataset)
Preview |
PDF
10.1038.s41587-024-02414-w.pdf - Published Version Available under License Creative Commons Attribution. Download (4MB) | Preview |
Abstract
A systematic evaluation of how model architectures and training strategies impact genomics model performance is needed. To address this gap, we held a DREAM Challenge where competitors trained models on a dataset of millions of random promoter DNA sequences and corresponding expression levels, experimentally determined in yeast. For a robust evaluation of the models, we designed a comprehensive suite of benchmarks encompassing various sequence types. All top-performing models used neural networks but diverged in architectures and training strategies. To dissect how architectural and training choices impact performance, we developed the Prix Fixe framework to divide models into modular building blocks. We tested all possible combinations for the top three models, further improving their performance. The DREAM Challenge models not only achieved state-of-the-art results on our comprehensive yeast dataset but also consistently surpassed existing benchmarks on Drosophila and human genomic datasets, demonstrating the progress that can be driven by gold-standard genomics datasets.
Actions (login required)
Administrator's edit/view item |