Arcee just dropped Trinity‑Large‑TrueBase, a raw 10‑trillion‑token checkpoint. This open‑source milestone pushes U.S. AI research forward, offering a massive base model for anyone to fine‑tune. Dive into the details of the training process and see how it can accelerate your ML projects. #OpenSourceAI #TrinityLarge #Arcee #BaseModel

🔗 https://aidailypost.com/news/arcee-releases-trinity-large-truebase-raw-10trilliontoken-checkpoint

Writing an LLM from scratch, part 28 -- training a base model from scratch on an RTX 3090

I felt like it should be possible to train a GPT-2 small level model on my own hardware using modern tools and open datasets from scratch. It was!

Giles' Blog
Tied Crosscoders: Explaining Chat Behavior from Base Model — LessWrong

Abstract We are interested in model-diffing: finding what is new in the chat model when compared to the base model. One way of doing this is training…

How To Setup Ollama LLM AI Chabot With Open WebUI, An Open-source AI User Interface On An Ubuntu VPS

YouTube