Huge congrats @[email protected]! π The future of creative MT will be "made in @[email protected]"! π
RT @[email protected]
We're proud of our colleague @[email protected] for becoming an @[email protected] #awardee for her project INCREC β Studying the creative translation process in the intersect with technology! π https://twitter.com/univgroningen/status/1620376478130049025
π¦π: https://twitter.com/GroNlp/status/1620395831663935488

University of Groningen on Twitter
βGreat news for @DocTinaK and @AnaGuerberof! They both got an @ERC_Research #consolidator grant of β¬2 million for research into, respectively, #parenthood and the creative process of #translating. Congrats!π
More π https://t.co/v5o0SBfzu5 @rug_gmw @FacultyofArtsUGβ
TwitterRT @[email protected]
π₯³Thrilled to announce our paper got accepted to #EACL2023!
We introduce Value Zeroing, a new interpretability method for quantifying context mixing in Transformers.
A joint work w/ me, @[email protected], @[email protected], and @[email protected]
πPaper: https://arxiv.org/abs/2301.12971
#NLProc #InDeep
π¦π: https://twitter.com/hmohebbi75/status/1620351439855063040
Quantifying Context Mixing in Transformers
Self-attention weights and their transformed variants have been the main
source of information for analyzing token-to-token interactions in
Transformer-based models. But despite their ease of interpretation, these
weights are not faithful to the models' decisions as they are only one part of
an encoder, and other components in the encoder layer can have considerable
impact on information mixing in the output representations. In this work, by
expanding the scope of analysis to the whole encoder block, we propose Value
Zeroing, a novel context mixing score customized for Transformers that provides
us with a deeper understanding of how information is mixed at each encoder
layer. We demonstrate the superiority of our context mixing score over other
analysis methods through a series of complementary evaluations with different
viewpoints based on linguistically informed rationales, probing, and
faithfulness analysis.
arXiv.orgShout-out to @
[email protected] et al. from our
#InDeep consortium for their awesome work "Quantifying Context Mixing in Transformers", introducing Value Zeroing as a new promising post-hoc interpretability approach for NLP! π Paper:
https://arxiv.org/abs/2301.12971 https://t.co/RxX2LenHqvQuantifying Context Mixing in Transformers
Self-attention weights and their transformed variants have been the main
source of information for analyzing token-to-token interactions in
Transformer-based models. But despite their ease of interpretation, these
weights are not faithful to the models' decisions as they are only one part of
an encoder, and other components in the encoder layer can have considerable
impact on information mixing in the output representations. In this work, by
expanding the scope of analysis to the whole encoder block, we propose Value
Zeroing, a novel context mixing score customized for Transformers that provides
us with a deeper understanding of how information is mixed at each encoder
layer. We demonstrate the superiority of our context mixing score over other
analysis methods through a series of complementary evaluations with different
viewpoints based on linguistically informed rationales, probing, and
faithfulness analysis.
arXiv.orgRT @[email protected]
As Gregor Samsa awoke one morning from uneasy dreams, he found himself rewritten in Rust for performance reasons
π¦π: https://twitter.com/vboykis/status/1620072731994914819
Vicki on Twitter
βAs Gregor Samsa awoke one morning from uneasy dreams, he found himself rewritten in Rust for performance reasonsβ
TwitterWelcome to @[email protected] π€ Looking forward to promising future collaborations!
RT @[email protected]
Happy to announce that, from February 1st, I'll start a new job as a lecturer π¨βπ« at @[email protected]
Excited to start this new adventure!
π¦π: https://twitter.com/marco_zul/status/1618977505632985089
Marco Zullich on Twitter
βHappy to announce that, from February 1st, I'll start a new job as a lecturer π¨βπ« at @univgroningen
Excited to start this new adventure!β
TwitterPrediction: LMaaS companies will use private RNG keys for their APIs and sell LM plagiarism detection tools on subscription for $$$ to schools and universities
RT @[email protected]
#OpenAI is planning to stop #ChatGPT users from making social media bots and cheating on homework by "watermarking" outputs. How well could this really work? Here's just 23 words from a 1.3B parameter watermarked LLM. We detected it with 99.999999999994% confidence. Here's how π§΅
π¦π: https://twitter.com/tomgoldsteincs/status/1618287665006403585
RT @[email protected]
#OpenAI is planning to stop #ChatGPT users from making social media bots and cheating on homework by "watermarking" outputs. How well could this really work? Here's just 23 words from a 1.3B parameter watermarked LLM. We detected it with 99.999999999994% confidence. Here's how π§΅
π¦π: https://twitter.com/tomgoldsteincs/status/1618287665006403585
Attention attribution is here! Thanks to all the awesome contributors! π
RT @[email protected]
Version 0.3.3 is finally out! π Highlights: attention attribution, new L2 norm default attribution aggregation, ruff linting (tip hat @[email protected]), improved save/reload of attributions. See release notes for usage examples: https://github.com/inseq-team/inseq/releases/tag/v0.3.3
π¦π: https://twitter.com/InseqDev/status/1616419539616694273

Release v0.3.3: Attention attribution, new aggregation, improved saving/reloading and more Β· inseq-team/inseq
Whatβs Changed
Attention attribution (#148 )
This release introduces a new category of attention attribution methods and adds support for AttentionAttribution (id: attention). This method attribute...
GitHubIf you are interested in the topic, our new @
[email protected] library greatly simplifies access to model internals with support for most recent gradient, attention, and (soon!) occlusion methods. Find it here:
https://github.com/inseq-team/inseq
GitHub - inseq-team/inseq: Interpretability for sequence generation models π π
Interpretability for sequence generation models π π - GitHub - inseq-team/inseq: Interpretability for sequence generation models π π
GitHub