I built a tiny LLM to demystify how language models work
https://github.com/arman-bd/guppylm
#HackerNews #tinyLLM #languageModels #AI #demystify #coding #guppylm
I built a tiny LLM to demystify how language models work
https://github.com/arman-bd/guppylm
#HackerNews #tinyLLM #languageModels #AI #demystify #coding #guppylm
Emotion concepts and their function in a large language model
https://www.anthropic.com/research/emotion-concepts-function
#HackerNews #EmotionConcepts #LanguageModels #AIResearch #NaturalLanguageProcessing #MachineLearning
fly51fly (@fly51fly)
대형 언어모델이 가르치는 과정에서 상대방의 정신 상태를 추론하는지(mentalize) 분석한 2026년 논문이 공개되었습니다. 인간의 teaching 상호작용을 모사하는 LLM의 인지적 행동을 다뤄, 모델 해석과 인간유사 추론 능력 연구에 의미가 있습니다.
Implicator.ai released the AI Top 40, a weekly ranking that combines 10 benchmarks into one score per language model. The system weights contamination-resistant tests like SWE-bench 4x higher than Chatbot Arena. GPT-5.4 currently leads despite Claude topping Arena rankings. Updates every Saturday and offers free embedding for websites.
Simplifying AI (@simplifyinAI)
Tencent과 Tsinghua가 CALM(Continuous Autoregressive Language Models)을 공개했습니다. 기존 LLM의 next-token 예측 패러다임을 대체하는 방식으로, 이산적 단일 토큰 예측에 쓰이던 막대한 연산 낭비를 줄이는 새로운 언어모델 접근법을 제시합니다.

🚨 BREAKING: Tencent has killed the “next-token” paradigm. Tencent and Tsinghua has released CALM (Continuous Autoregressive Language Models), and it completely disrupts the next-token paradigm. LLMs currently waste massive amounts of compute predicting discrete, single tokens
Teaching Writing in the Age of AI: Assessment and “Cheating”
This is the fourth post in a series on Teaching Writing in the Age of AI. The first post provided an overview of some of the changes we're facing as the number of AI writing tools increases. Post two covered conversations about academic integrity, and the third post offered some practical advice on teaching students to be critical readers and writers. In this post, I'll be exploring the assessment of writing, and why AI is such an apparent threat to the way we currently teach and assess. In […]https://leonfurze.com/2023/02/18/teaching-writing-in-the-age-of-ai-assessment-and-cheating/

Large language models (LLMs) are growing increasingly capable, prompting recent interest in LLM teams. Yet, despite increased deployment of LLM teams at scale, we lack a principled framework for addressing key questions such as when a team is helpful, how many agents to use, how structure impacts performance -- and whether a team is better than a single agent. Rather than designing and testing these possibilities through trial-and-error, we propose using distributed systems as a principled foundation for creating and evaluating LLM teams. We find that many of the fundamental advantages and challenges studied in distributed computing also arise in LLM teams, highlighting the rich practical insights that can come from the cross-talk of these two fields of study.