Gerard de Melo

@gdm
455 Followers
294 Following
247 Posts

Professor for #AI at HPI / University of Potsdam (based in Berlin, Germany)
Previously: Rutgers (US), Tsinghua (Beijing).

#AI, #NLP, #DataScience, #ML, #tech

he/him

Webhttp://gerard.demelo.org/
Group Pagehttps://hpi.de/en/de-melo/home.html
LinkedInhttps://www.linkedin.com/in/gdemelo
Contacthttp://gerard.demelo.org/contact.html

🚨Which AI conferences to submit to in 2025?🚨

Put up an up-to-date version of aideadlin.es, which was originally created by Abhishek Das.

For 2025 deadlines, check out:
👉🏽 http://aideadlines.demelo.org/

AI Conference Deadlines

Countdowns to top AI/ML/CV/NLP/robotics conference deadlines

LLMs "hallucinate" by making up information instead of revealing their lack of knowledge 🤔

Our #NeurIPS2024 paper adds an [I-Don't-Know] 🤷🏾‍♂️ token to the LLM and fine-tunes it to predict this instead of hallucinating.

No extra annotation is needed for training! Our IDK objective divides the shifted probability mass between the gold token and I-Don't-Know token for wrong predictions.

Details in the paper👉🏾 https://arxiv.org/abs/2412.06676v1

w/ Roi Cohen, Konstantin Dobler, Eden Biran

I Don't Know: Explicit Modeling of Uncertainty with an [IDK] Token

Large Language Models are known to capture real-world knowledge, allowing them to excel in many downstream tasks. Despite recent advances, these models are still prone to what are commonly known as hallucinations, causing them to emit unwanted and factually incorrect text. In this work, we propose a novel calibration method that can be used to combat hallucinations. We add a special [IDK] ("I don't know") token to the model's vocabulary and introduce an objective function that shifts probability mass to the [IDK] token for incorrect predictions. This approach allows the model to express uncertainty in its output explicitly. We evaluate our proposed method across multiple model architectures and factual downstream tasks. We find that models trained with our method are able to express uncertainty in places where they would previously make mistakes while suffering only a small loss of encoded knowledge. We further perform extensive ablation studies of multiple variations of our approach and provide a detailed analysis of the precision-recall tradeoff of our method.

arXiv.org

I am chairing the
AI@HPI Conference: Responsible AI

December 3-4 in Potsdam (Berlin metropolitan area)

Discussing AI with regard to bias, elections/society, trustworthiness, copyright, the EU AI Act, and best practices.

Register at
https://hpi.de/en/ai-hpi-conference/

Please repost to spread the word!

Among our speakers are:
* Special guest Gale Anne Hurd @GunnerGale: Well-known Hollywood producer (The Terminator, Walking Dead, etc.) and Founder/CEO of Valhalla Entertainment
[continued in reply]

AI@HPI Conference | Hasso-Plattner-Institut

Hasso-Plattner-Institut

German Chancellor 🇩🇪 Olaf Scholz got to hear about our 🦍 GorillaTracker project!

Using computer vision for re-identification of Western Lowland Gorillas in the Congo rainforest.

Collaboration with primatologist Magdalena Bremen of SPAC + Dante Wasmuht from Conservation X Labs.

Our amazing @Hasso_Plattner_Institute team consists of Maximilian Schall (lead), Ben Meyer-Meisel, Bennet Kampe, Emirhan Dogan, Joscha Schroff, Kajo Kratzenstein, Liam van der Viven, Robert Weeke & Vincent Eberwein.

ArtQuest: Countering Hidden Language Biases in ArtVQA

◼ ArtVQA = AI that answers questions about visual art

◼ We find that prior benchmarks can be tackled without actually looking at the image

◼ We propose a new benchmark called ArtQuest and investigate different deep learning models.

Presented at #WACV2024 this week

w/ Tibor Bleidt + Sarah Eslami

📁 Code:
https://github.com/bletib/artquest

ⓘ Paper:
http://gerard.demelo.org/papers/artvqa-artquest.pdf

GitHub - BleTib/artquest

Contribute to BleTib/artquest development by creating an account on GitHub.

GitHub

🚀🚀🚀
How to best scale up LLM training to 32 NVIDIA DGX A100 nodes?

Check out our #NeurIPS2023 WANT paper with extensive comparisons regarding different parallelization strategies and how they interact with flash attention/fused kernels, activation checkpointing, etc.

https://arxiv.org/abs/2311.05610

Further details: https://twitter.com/johannes_hage/status/1724682703642226918

Efficient Parallelization Layouts for Large-Scale Distributed Model Training

Efficiently training large language models requires parallelizing across hundreds of hardware accelerators and invoking various compute and memory optimizations. When combined, many of these strategies have complex interactions regarding the final training efficiency. Prior work tackling this problem did not have access to the latest set of optimizations, such as FlashAttention or sequence parallelism. In this work, we conduct a comprehensive ablation study of possible training configurations for large language models. We distill this large study into several key recommendations for the most efficient training. For instance, we find that using a micro-batch size of 1 usually enables the most efficient training layouts. Larger micro-batch sizes necessitate activation checkpointing or higher degrees of model parallelism and also lead to larger pipeline bubbles. Our most efficient configurations enable us to achieve state-of-the-art training efficiency results over a range of model sizes, most notably a Model FLOPs utilization of 70.5% when training a Llama 13B model.

arXiv.org

Humans don't necessarily follow instructions as well as AI systems!

Important to consider when training your system on human responses.

Via https://twitter.com/nearcyan/status/1722067382930133364

near on X

fine-tuning cautionary tale

X (formerly Twitter)

Had fun giving a talk on

AI Ethics & Compliance

in our joint workshop on AI by HPI Academy and Merantix AI School/Merantix Momentum

featuring Ralf Herbrich of @Hasso_Plattner_Institute and Moritz Schröder of Merantix AI Campus

https://hpi-academy.de/en/workshops/ai-essentials-for-decision-makers-navigating-the-future-of-business/

AI Essentials for Decision Makers - Navigating the future of Business

Jährlich besuchen zahlreiche Fach- und Führungskräfte die Workshops und Programme der HPI Academy, um sich weiterzubilden oder um ihre erworbenen Kompetenzen in den Bereichen Design Thinking, Leadership, Innovation & Transformation sowie IT & Digitalisierung zertifizieren zu lassen. Wir garantieren Ihnen höchste Qualitätsstandards und beste Lernerlebnisse sowohl für unsere Präsenz- als auch unsere Online-Angebote. Lassen Sie sich also inspirieren, bei uns vor Ort oder von überall auf der Welt. Sehr gerne unterstützen wir Sie auf Ihrer persönlichen Lernreise.

HPI Academy

Full house and exciting discussions at Urania Potsdam where I gave a talk on opportunities and risks of current AI

https://www.urania-potsdam.de/veranstaltungen/2456749/2023/11/02/künstliche-intelligenz-chance-oder-bedrohung.html

#ai #genai #airisk

Urania Verein - Künstliche Intelligenz - Chance oder Bedrohung?

Turing award winners Geoff Hinton, Yoshua Bengio, Andrew Yao (my former boss), along with many others have published a short paper on AI risks.

Whether or not you agree with the arguments, we definitely need to keep thinking about these issues.

->
https://managing-ai-risks.com

Managing extreme AI risks amid rapid progress