Matthew Honnibal

323 Followers
21 Following
20 Posts
Style (and mindset) tips for less experienced developers working with AI https://honnibal.dev/blog/llm-style-tips
Why I don't think AI is a bubble · honnibal.dev

Matthew Honnibal is a computational linguist from Sydney based in Berlin, Germany. He's the author of the spaCy Natural Language Processing library and the co-founder of Explosion.

Just published part 3 of my blog post series on making beautiful slides for your talks 🎨✨

This one is about presenting technical content and making dry and abstract topics more interesting. Featuring many examples, including talks by Vitaly Meursault and @sofie!

https://ines.io/blog/beautiful-slides-talks-part-3-technical-content/

Making beautiful slides for your talks, part 3: Technical content · ines.io

In part 3 of this series, I share some tips and inspiration for presenting technical content in your slides and making “boring” topics more fun and engaging.

ines.io

spaCy and Prodigy started as indie projects, but in 2021 we decided to raise capital and have a larger team. We couldn’t make that configuration work, so we’re back to how we were before. I’ll be spending most of my time hands-on with spaCy again, and we have a lot of updates and improvements planned for Prodigy.

I hate how vaguely these things are usually discussed, so I also wrote a long post about it all: https://honnibal.dev/blog/back-to-our-roots

Why I don't think AI is a bubble · honnibal.dev

Matthew Honnibal is a computational linguist from Sydney based in Berlin, Germany. He's the author of the spaCy Natural Language Processing library and the co-founder of Explosion.

Company update: We're going back to our roots!

We're back to running Explosion as a smaller, independent-minded and self-sufficient company. spaCy and Prodigy will stay stable and sustainable and we'll keep updating our stack with the latest technologies, without changing its core identity or purpose 💙

https://explosion.ai/blog/back-to-our-roots-company-update

The ultimate guide to optimizing annotation workflows · Explosion

This blog post collects tips and advice for how to build efficient human-in-the-loop data development workflows, break down business problems into actionable annotation steps and make the most of automation and model assistance.

We are really excited to share that we have just released the alpha version of Prodigy v1.12! This includes LLM-assisted workflows for data annotation and prompt engineering as well as extended, fully customizable support for multi-annotator workflows.

https://support.prodi.gy/t/prodigy-1-12-alpha-release-llm-assisted-workflows-prompt-engineering-fully-custom-task-routing-for-multi-annotator-scenarios/6552

Prodigy 1.12 alpha release: LLM-assisted workflows, prompt engineering & fully custom task routing for multi-annotator scenarios.

Hey everyone! We are really excited to share that we have just released the alpha version of Prodigy v1.12! (v1.12a1). This release is available for download for all v.1.11.x license holders and includes: New recipes for LLM-assisted annotations and prompt engineering: the LLM assisted workflows we have announced a while ago are now fully integrated with Prodigy and available out of the box. For v1.12a1 you'd still be restricted to OpenAI API to use them, but by v1.12a2 we definitely want to...

Prodigy Support

We present a brand new workflow for prompt engineering that allows you to compare the quality of several prompts in a tournament. The algorithm uses the Glico ranking system [https://en.wikipedia.org/wiki/Glicko_rating_system] to select the best prompt.

https://future--prodi-gy.netlify.app/docs/large-language-models#tournaments

Glicko rating system - Wikipedia

Here are the slides for my #PyDataLondon keynote on LLMs from prototype to production ✨

Including:
◾ visions for NLP in the age of LLMS
◾ a case for LLM pragmatism
◾ solutions for structured data
◾ spaCy LLM + https://prodi.gy

https://speakerdeck.com/inesmontani/large-language-models-from-prototype-to-production

What will production NLP look like, once the dust settles around LLMs? One view is basically “prompts are all you need”. I disagree. I wrote a bit about this when we released #spaCy LLM last week, but the topic deserves its own post, so here it is.

https://explosion.ai/blog/against-llm-maximalism

Against LLM maximalism · Explosion

LLMs are not a direct solution to most of the NLP use-cases companies have been working on. They are extremely useful, but if you want to deliver reliable software you can improve over time, you can't just write a prompt and call it a day. Once you're past prototyping and want to deliver the best system you can, supervised learning will often give you better efficiency, accuracy and reliability.

Explosion
Hi #MastoCats! Let me introduce Rizhik and Alaska, our guest cats from Ukraine.

Machine learning is basically programming by example: instead of specifying a system's behaviour with code, you (imperfectly) specify the desired behaviour with training data.

Well, zero-shot learning is like that, but without the training data. That does have some advantages — you don't have to tell it much about what you want it to do. But it's also pretty limiting. You can't tell it much about what you want it to do.