Skip to content

Sauravroy34/Teaching_LLMS_Chemistry

Repository files navigation

OLMo Learns Chemistry

Hugging Face Model

Can a General LLM Learn Chemistry?

In this project, I test if a general large language model (LLM) can learn to handle chemistry tasks well.

I start with OLMo-7B, a general LLM pre-trained on the DOLMA dataset. I then do continued pre-training using QLoRA.

I test the model on the MoleculeNet benchmark. I skip big datasets like QM9 and SIDER to save time and resources.

Continued Pre-Training

I train the model on 2.1 million raw SMILES strings from the smiles-molecules-chembl dataset.

I use QLoRA for pre-training with these settings:

  • Target: all_linear

  • Rank: 64

  • Alpha: 128

  • Learning rate: 5e-5

  • Training Script: RawSmiles.py (for ChemOlmo-7B)

All benchmark code is in the Notebooks folder.

Results

Classification

Classification Results

Regression

Regression Results

References

  1. Dolma
  2. OLMo-7B
  3. ZINC20
  4. USPTO
  5. DeepChem
  6. ChemBERTa-3
  7. MoleculeNet

About

Pre-training OLMO-7B model on cheminformatics dataset

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors