An explainable and simplified version of OLMo model
-
Updated
Mar 5, 2025 - Jupyter Notebook
An explainable and simplified version of OLMo model
AI工具类(Midjourney / Notion / ChatGPT) 订阅教程类(Netflix / Spotify / Adobe) 虚拟卡支付类(Namecheap / OpenAI / Google)
RDKit-Guided Topological State Machine (TSM) for constrained SMILES generation with OLMo-7B. Solves BPE-tokenizer mismatch via RDKit-in-the-loop decoding. Achieved 100% validity on allenai/OLMo-7B-hf.
Pre-training a ~150M parameter code-specialized language model using OLMo 3 architecture (GQA, SWA, SwiGLU, RoPE) on PHP/JS/Python/C source code.
Mechanistic interpretability research for studying how instruction tuning restructures the computational pipeline of language models
Provide an open-source implementation of the OLMo language model for efficient and accurate natural language processing tasks.
Add a description, image, and links to the olmo topic page so that developers can more easily learn about it.
To associate your repository with the olmo topic, visit your repo's landing page and select "manage topics."