Stars
MoBA: Mixture of Block Attention for Long-Context LLMs
Awesome-Biomolecule-Language-Cross-Modeling: a curated list of resources for paper "Leveraging Biomolecule and Natural Language through Multi-Modal Learning: A Survey"
The first large protein language model trained follows structure instructions.
A family of open-sourced Mixture-of-Experts (MoE) Large Language Models
My own attempt at a long context genomics model, leveraging recent advances in long context attention modeling (Flash Attention + other hierarchical methods)
Quantum computational chemistry based on TensorCircuit
Sequence-to-drug concept adds a perspective on drug design. It can serve as an alternative method to SBDD, particularly for proteins that do not yet have high-quality 3D structures available.
Tensor network based quantum software framework for the NISQ era
Quantum circuit on top of tensor network. This version is archived, latest open source version release is at https://github.com/tensorcircuit/tensorcircuit-ng
Replication attempt for the Protein Folding Model described in https://www.biorxiv.org/content/10.1101/2021.08.02.454840v1
Open source code for AlphaFold 2.
This package contains deep learning models and related scripts for RoseTTAFold
CogDL: A Comprehensive Library for Graph Deep Learning (WWW 2023)
a modified version of the trRosetta folding protocol
Reinforcement Learning for Molecular Design Guided by Quantum Mechanics
"Linear Algebraic Tools for Graph Computation," Minisymposium 142 from the 2020 SIAM Conference on Math of Data Science.
Grammars written for ANTLR v4; expectation that the grammars are free of actions.
A simple SMILES validator and parser using pyparsing
Open-Source Neural Machine Translation in Tensorflow
Python code for training models in the ACL paper, "Beyond BLEU:Training Neural Machine Translation with Semantic Similarity".
Data generation program for alchemy project
Resources for people running research groups.