Skip to content

Predicting gene expression using millions of random promoter sequences

Notifications You must be signed in to change notification settings

mtinti/DREAMChallenge2022

Repository files navigation

alt text

DREAM Challenge 2022, Model by mt (10th)

Short Description

I developed an end-to-end procedure to predict the expression of genes from random promoters using Jupyter Notebook and TensorFlow. My approach uses Recurrent (GRU) and Convolution Neural Networks to regress the strength of the targeted promoters using information encoded in the forward and reverse DNA strands. The starting point of my model, two bi-bidirectional GRUs, have been recently used in a machine learning competition at Kaggle to predict the stability of RNA vaccines. In this work, I expanded on this architecture and found that adding convolutions and fully connected layers can efficiently extract features from DNA sequences and predict gene expression. Surprisingly, the training of this model benefits from using Binary Crossentropy loss coupled with the sigmoid activation of the output layer.

Usage

Create a conda env using the environment.yml file. If the env creation fails, the python version and packages used are listed in the last cells of the jupyter notebook (250_pandas_final_1_bis_bc-0.745-Copy1.nbconvert.ipynb)

The jupyter notebook contains an end-to-end strategy to predict the expression of genes using random promoters. Click here to download the input files. If the input files and the notebook are present in the same folder, you can run all cells to train the model until epoch 11, used for the competition. The notebook will then produce the prediction on the test set in tsv format and json format.

Alternatively, you can directly use a helper function (make_prediction) to predict the test sequences with the sub_model.11.h5 model (used for the competition) or any other model created during training.

Read more about this model here

About

Predicting gene expression using millions of random promoter sequences

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published