Gabriele Sarti

@gsarti
26 Followers
41 Following
71 Posts
PhD Student in Explainable NMT at GroNLP. Interpretability ∩ NLG #NLProc. Prev: AWS, Aindo, ItaliaNLP Lab. he/him
LocationGroningen, Netherlands
Websitehttps://gsarti.com
Twitterhttps://twitter.com/gsarti_
Githubhttps://github.com/gsarti

Huge congrats @[email protected]! πŸš€ The future of creative MT will be "made in @[email protected]"! 😎

RT @[email protected]

We're proud of our colleague @[email protected] for becoming an @[email protected] #awardee for her project INCREC – Studying the creative translation process in the intersect with technology! πŸŽ‰ https://twitter.com/univgroningen/status/1620376478130049025

πŸ¦πŸ”—: https://twitter.com/GroNlp/status/1620395831663935488

University of Groningen on Twitter

β€œGreat news for @DocTinaK and @AnaGuerberof! They both got an @ERC_Research #consolidator grant of €2 million for research into, respectively, #parenthood and the creative process of #translating. Congrats!πŸŽ€ More πŸ‘‰ https://t.co/v5o0SBfzu5 @rug_gmw @FacultyofArtsUG”

Twitter

RT @[email protected]

πŸ₯³Thrilled to announce our paper got accepted to #EACL2023!
We introduce Value Zeroing, a new interpretability method for quantifying context mixing in Transformers.

A joint work w/ me, @[email protected], @[email protected], and @[email protected]

πŸ“‘Paper: https://arxiv.org/abs/2301.12971

#NLProc #InDeep

πŸ¦πŸ”—: https://twitter.com/hmohebbi75/status/1620351439855063040

Quantifying Context Mixing in Transformers

Self-attention weights and their transformed variants have been the main source of information for analyzing token-to-token interactions in Transformer-based models. But despite their ease of interpretation, these weights are not faithful to the models' decisions as they are only one part of an encoder, and other components in the encoder layer can have considerable impact on information mixing in the output representations. In this work, by expanding the scope of analysis to the whole encoder block, we propose Value Zeroing, a novel context mixing score customized for Transformers that provides us with a deeper understanding of how information is mixed at each encoder layer. We demonstrate the superiority of our context mixing score over other analysis methods through a series of complementary evaluations with different viewpoints based on linguistically informed rationales, probing, and faithfulness analysis.

arXiv.org
Shout-out to @[email protected] et al. from our #InDeep consortium for their awesome work "Quantifying Context Mixing in Transformers", introducing Value Zeroing as a new promising post-hoc interpretability approach for NLP! πŸŽ‰ Paper: https://arxiv.org/abs/2301.12971 https://t.co/RxX2LenHqv
Quantifying Context Mixing in Transformers

Self-attention weights and their transformed variants have been the main source of information for analyzing token-to-token interactions in Transformer-based models. But despite their ease of interpretation, these weights are not faithful to the models' decisions as they are only one part of an encoder, and other components in the encoder layer can have considerable impact on information mixing in the output representations. In this work, by expanding the scope of analysis to the whole encoder block, we propose Value Zeroing, a novel context mixing score customized for Transformers that provides us with a deeper understanding of how information is mixed at each encoder layer. We demonstrate the superiority of our context mixing score over other analysis methods through a series of complementary evaluations with different viewpoints based on linguistically informed rationales, probing, and faithfulness analysis.

arXiv.org

RT @[email protected]

As Gregor Samsa awoke one morning from uneasy dreams, he found himself rewritten in Rust for performance reasons

πŸ¦πŸ”—: https://twitter.com/vboykis/status/1620072731994914819

Vicki on Twitter

β€œAs Gregor Samsa awoke one morning from uneasy dreams, he found himself rewritten in Rust for performance reasons”

Twitter

Welcome to @[email protected] πŸ€— Looking forward to promising future collaborations!

RT @[email protected]

Happy to announce that, from February 1st, I'll start a new job as a lecturer πŸ‘¨β€πŸ« at @[email protected]
Excited to start this new adventure!

πŸ¦πŸ”—: https://twitter.com/marco_zul/status/1618977505632985089

Marco Zullich on Twitter

β€œHappy to announce that, from February 1st, I'll start a new job as a lecturer πŸ‘¨β€πŸ« at @univgroningen Excited to start this new adventure!”

Twitter
Prediction: LMaaS companies will use private RNG keys for their APIs and sell LM plagiarism detection tools on subscription for $$$ to schools and universities

RT @[email protected]

#OpenAI is planning to stop #ChatGPT users from making social media bots and cheating on homework by "watermarking" outputs. How well could this really work? Here's just 23 words from a 1.3B parameter watermarked LLM. We detected it with 99.999999999994% confidence. Here's how 🧡

πŸ¦πŸ”—: https://twitter.com/tomgoldsteincs/status/1618287665006403585

Tweet / Twitter

Twitter

RT @[email protected]

#OpenAI is planning to stop #ChatGPT users from making social media bots and cheating on homework by "watermarking" outputs. How well could this really work? Here's just 23 words from a 1.3B parameter watermarked LLM. We detected it with 99.999999999994% confidence. Here's how 🧡

πŸ¦πŸ”—: https://twitter.com/tomgoldsteincs/status/1618287665006403585

Tweet / Twitter

Twitter

Attention attribution is here! Thanks to all the awesome contributors! πŸš€

RT @[email protected]

Version 0.3.3 is finally out! πŸŽ‰ Highlights: attention attribution, new L2 norm default attribution aggregation, ruff linting (tip hat @[email protected]), improved save/reload of attributions. See release notes for usage examples: https://github.com/inseq-team/inseq/releases/tag/v0.3.3

πŸ¦πŸ”—: https://twitter.com/InseqDev/status/1616419539616694273

Release v0.3.3: Attention attribution, new aggregation, improved saving/reloading and more Β· inseq-team/inseq

What’s Changed Attention attribution (#148 ) This release introduces a new category of attention attribution methods and adds support for AttentionAttribution (id: attention). This method attribute...

GitHub
If you are interested in the topic, our new @[email protected] library greatly simplifies access to model internals with support for most recent gradient, attention, and (soon!) occlusion methods. Find it here: https://github.com/inseq-team/inseq
GitHub - inseq-team/inseq: Interpretability for sequence generation models πŸ› πŸ”

Interpretability for sequence generation models πŸ› πŸ” - GitHub - inseq-team/inseq: Interpretability for sequence generation models πŸ› πŸ”

GitHub