9. TL;DR
Word2Vec was more than just a way to embed words.
It showed us that contrastive learning works — and it’s now everywhere in LLMs and beyond.

#LLM #AI #Embeddings #NLP #MachineLearning #Word2Vec #ContrastiveLearning

Just given my first technical (guest) lecture to a group of #undergraduate students about #ai and semi-supervised #ContrastiveLearning!

It was a great opportunity to practice technical teaching ahead of September.

Topic: Nowcasting #flood risk from social media data with #CLIP and semi-supervised learning.

Image source: AI generated with Microsoft Designer, prompt: A cat giving a lecture in a lecture theatre, stylised photo
Reasoning: I doubt this exists online and I'm short on time.

This AI learnt language by seeing the world through a baby’s eyes
#AI #learning #baby #camera #human #NeuralNetwork #ContrastiveLearning

#Baby trains #AI to #learn #English #language?

In Wai Keen Vong's poster, "Grounded language acquisition through the eyes and ears of a single child", we learned how a multimodal #neuralNetwork model learned words——sometimes multiple senses of a word (like "button": the one that secures clothes and the one you press)!

Another standout poster at the #Philosophy of #DeepLearning #conference at #NYU.

#ComputerScience #ComputerVision #ContrastiveLearning #MachineLearning

Our pick of the week by @sarapapi: Siqi Ouyang et al., "WACO: Word-Aligned Contrastive Learning for Speech Translation"
https://arxiv.org/abs/2212.09359
#NLP #NLProc #SpeechTranslation #ST #MachineTranslation #MT #ContrastiveLearning #AI
WACO: Word-Aligned Contrastive Learning for Speech Translation

End-to-end Speech Translation (E2E ST) aims to directly translate source speech into target text. Existing ST methods perform poorly when only extremely small speech-text data are available for training. We observe that an ST model's performance closely correlates with its embedding similarity between speech and source transcript. In this paper, we propose Word-Aligned COntrastive learning (WACO), a simple and effective method for extremely low-resource speech-to-text translation. Our key idea is bridging word-level representations for both speech and text modalities via contrastive learning. We evaluate WACO and other methods on the MuST-C dataset, a widely used ST benchmark, and on a low-resource direction Maltese-English from IWSLT 2023. Our experiments demonstrate that WACO outperforms the best baseline by 9+ BLEU points with only 1-hour parallel ST data. Code is available at https://github.com/owaski/WACO.

arXiv.org