This package contains the code to reproduce the work in this report.
The report is about improving upon the Enformer model by replacing its self-attention layers with the sparse-attention layers from the BigBird model. This change allows for longer DNA sequences to be used as input to the model, while using the relatively less memory. This is because memory scales quadratically with sequence length when using self-attention, compared to the linear scaling when sparse-attention is used.
The created model from this work is called sparse-enformer (s-enformer), which is also the name of this package.
To install the s-enformer package:
python3 setup.py install --user
More info about using this package is found at s_enformer/README.md.