Starting a reading list for only normcore content around LLMs. What would you add?

Normcore = no hype, no langchain, no AI is going to destroy us all, but practical, technical readings in navigating in this brave new world.

https://gist.github.com/veekaybee/be375ab33085102f9027853128dc5f0e

Normcore LLM Reads

Normcore LLM Reads. GitHub Gist: instantly share code, notes, and snippets.

Gist
@vicki I like Ethan mollick s substack https://www.oneusefulthing.org and @simon s blog as well. I have more written up in my vault but I would have to dig for it.
One Useful Thing | Ethan Mollick | Substack

Trying to understand the implications of AI for work, education, and life. By Prof. Ethan Mollick. Click to read One Useful Thing, by Ethan Mollick, a Substack publication with hundreds of thousands of subscribers.

Generative AI: What You Need To Know

Become an expert detector of AI bullshit

@vicki I think if I read all of it first I will forget reply with stuff. And this is just stuff, don't take it too seriously.
There's Stormtrooper that I am yet to use
https://centre-for-humanities-computing.github.io/stormtrooper/index.html
There's SetFit which really didn't work on my data, possibly because I have too many classes
https://github.com/huggingface/setfit
They feel normy because it's pretty fast to see if it works with your data.
stormtrooper

@vicki
The Intelligence Illusion by @baldur definitely checks the anti-hype box.

https://illusion.baldurbjarnason.com/

The Intelligence Illusion (Second Edition): Why generative models are bad for business

Available in PDF and EPUB

@brohrer @baldur this looks great! Going to stick to free resources for now but bookmarking for myself

@vicki Possibly a bit out of date:

Attention Is All You Need
The Illustrated Transformer
The Annotated Tranformer
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Generating Wikipedia by Summarizing Long Sequences
RoBERTa: A Robustly Optimized BERT Pretraining Approach
The Illustrated GPT-2
ELECTRA: Pre-Training Text Encoders as Discriminators Rather than Generators
Scaling Laws for Neural Language Models
Training Compute-Optimal Large Language Models

@vicki
Language Models are Few-Shot Learners
LLM Introduction: Learn Language Models

LLM Introduction: Learn Language Models. GitHub Gist: instantly share code, notes, and snippets.

Gist
@vicki I expected a lot more AWS and huggingface documentation on this list
@vicki I don’t know but here is something I saw that is pretty terrible 🍸😺