chris brockett

@chris_brockett
1.1K Followers
604 Following
201 Posts

Leftover linguist. Data janitor at Microsoft Research. Tsundoku expert. Transient epiphanies. "Talented underachiever." Interested in language, computation, cognition, and agency. All opinions are my own. Even when I agree with my employer.

#NaturalLanguageProcessing #NLProc #MachineLearning #ML #Dialog #ChatBots #Cognition #Neuroscience #Language #Linguistics #DigitalHumanities #HCI

#PugetSound #Washington #Earthquakes

#JapaneseLiterature #EastAsia

Web pagehttps://www.microsoft.com/en-us/research/people/chrisbkt/
Also at:@chrisbrockett.bsky.social@chris_brockett

The NY Times' chronic timidity in calling things what they are is visible in its story on Musk's latest boosting of extremism. The news org says he has "faced increasing criticism that he has tolerated and even encouraged antisemitic abuse on his social media platform."

Remove the "faced increasing criticism that" part of that sentence, and it's what the Times should have reported. He's a bigot and he promotes bigots and extremism, period.

Did Twitter not pay its Google Cloud bill? #twitterdown
Here's a fun paper about some interesting properties of GPT-3 models https://arxiv.org/abs/2303.14310 #GPT #turingmachine
GPT is becoming a Turing machine: Here are some ways to program it

We demonstrate that, through appropriate prompting, GPT-3 family of models can be triggered to perform iterative behaviours necessary to execute (rather than just write or recall) programs that involve loops, including several popular algorithms found in computer science curricula or software developer interviews. We trigger execution and description of Iterations by Regimenting Self-Attention (IRSA) in one (or a combination) of three ways: 1) Using strong repetitive structure in an example of an execution path of a target program for one particular input, 2) Prompting with fragments of execution paths, and 3) Explicitly forbidding (skipping) self-attention to parts of the generated text. On a dynamic program execution, IRSA leads to larger accuracy gains than replacing the model with the much more powerful GPT-4. IRSA has promising applications in education, as the prompts and responses resemble student assignments in data structures and algorithms classes. Our findings hold implications for evaluating LLMs, which typically target the in-context learning: We show that prompts that may not even cover one full task example can trigger algorithmic behaviour, allowing solving problems previously thought of as hard for LLMs, such as logical puzzles. Consequently, prompt design plays an even more critical role in LLM performance than previously recognized.

arXiv.org
Here's a fun paper about some interesting properties of GPT-3 models https://arxiv.org/abs/2303.14310 #GPT #turingmachine
GPT is becoming a Turing machine: Here are some ways to program it

We demonstrate that, through appropriate prompting, GPT-3 family of models can be triggered to perform iterative behaviours necessary to execute (rather than just write or recall) programs that involve loops, including several popular algorithms found in computer science curricula or software developer interviews. We trigger execution and description of Iterations by Regimenting Self-Attention (IRSA) in one (or a combination) of three ways: 1) Using strong repetitive structure in an example of an execution path of a target program for one particular input, 2) Prompting with fragments of execution paths, and 3) Explicitly forbidding (skipping) self-attention to parts of the generated text. On a dynamic program execution, IRSA leads to larger accuracy gains than replacing the model with the much more powerful GPT-4. IRSA has promising applications in education, as the prompts and responses resemble student assignments in data structures and algorithms classes. Our findings hold implications for evaluating LLMs, which typically target the in-context learning: We show that prompts that may not even cover one full task example can trigger algorithmic behaviour, allowing solving problems previously thought of as hard for LLMs, such as logical puzzles. Consequently, prompt design plays an even more critical role in LLM performance than previously recognized.

arXiv.org
Anyway, one thing I can tell you really definitively is that these guys are very, very afraid of anyone who is not at all afraid of them.
One thing I think about sometimes is the time a decade or so ago when the editor of the MIT Tech Review scolded me for being critical of Milo Yiannopolis because Milo “went to a good school” (some British shit that probably means a lot to them but I can’t even remember what it was) and therefore couldn’t actually be awful. That was one of those turning points where I realized I would never actually get elevated to a certain level of culture or visibility, and I was quite content with that.
On the reliability of inter-rater agreement scores. This is a serious problem when interpreting crowd-sourced annotations. https://interhumanagreement.substack.com/p/kappa-scores-considered-harmful #interraterreliability
Kappa scores considered harmful

While popular, this data quality measure has some critical flaws, that have been known for a long time, but are often ignored.

inter human agreement

I'm teaching "NLP for Cultural Analytics" at UW Linguistics this quarter, and I thought I'd share my reading list for the course.

https://maria-antoniak.github.io/teaching/2023-nlp-ca

Lots of usual suspects from the Journal of Cultural Analytics, but also pieces from NLP and computational social science. Thinking back to an earlier thread on "what even is cultural analytics," you can see from this list my perception of the field (or maybe just what I'm in the mood to read right now).

#DigitalHumanities #CulturalAnalytics

Maria Antoniak

My academic website / portfolio.

PyTorch, critical Machine Learning/AI library PyTorch breached, compromised with malicious software. If you run on the latest (unstable) versions, that requires immediate attention. https://pytorch.org/blog/compromised-nightly-dependency/
Compromised PyTorch-nightly dependency chain between December 25th and December 30th, 2022. – PyTorch

Matrices and graphs

The single most undervalued fact of linear algebra: matrices are graphs, and graphs are matrices

The Palindrome