"You can use the metric to raise an index" is what I found in one or the other physics text book. Surely I missed some context each time, so I couldn't help to swear: 🤢 , I like the indexes where they are, what is the point of raising or lowering them.

I found the answer in Gravitation, by Misner, Thorne, Wheeler but felt the need to write it down in a way I understand it best: https://miamao.de/blog/2024-06/04.Raising_or_Lowering_an_Index.html

#physics #tensor #metric #vectorspace #dotproduct #raisinganindex #mtw

Haralds Blog — Raising or Lowering an Index

The Double Helix inside the NLP Transformer
https://arxiv.org/abs/2306.13817

We introduce a framework for analyzing types of information in a NLP Transformer. We distinguish four layers of information: positional, syntactic, semantic, and contextual.

We show that the distilled positional components of the embedding vectors follow the path of a helix, both on the encoder side & on the decoder side.
...

#NLP #MachineLearning #transformers #DependencyParsing #PartsOfSpeech #VectorSpace #informatics

The Double Helix inside the NLP Transformer

We introduce a framework for analyzing various types of information in an NLP Transformer. In this approach, we distinguish four layers of information: positional, syntactic, semantic, and contextual. We also argue that the common practice of adding positional information to semantic embedding is sub-optimal and propose instead a Linear-and-Add approach. Our analysis reveals an autogenetic separation of positional information through the deep layers. We show that the distilled positional components of the embedding vectors follow the path of a helix, both on the encoder side and on the decoder side. We additionally show that on the encoder side, the conceptual dimensions generate Part-of-Speech (PoS) clusters. On the decoder side, we show that a di-gram approach helps to reveal the PoS clusters of the next token. Our approach paves a way to elucidate the processing of information through the deep layers of an NLP Transformer.

arXiv.org

Juan-Pablo Ortega form #NTU recommended a GTM by Serre. I was aware of its existence but somehow assumed that it would be harder. Maybe I'll go for that instead?

Anyway, I like in Isaacs book that it exposes what aspects of linear #vectorspace make #GroupTheory tick. Also it was really fun reading Schur-Frobenius theorem yesterday, there the #module framework came to shine. I appreciate that.

What's more trouble for me is that the notation is sometimes not introduces certain things or I'm used to \(f(x)\) as opposed to \(xf\) 🤷‍♂️. Im getting used to inline definitions made in passing but suddenly later coming into play.

I'm trying to educate myself in graduate level #mathematics and picked up this book by Isaacs on #GroupTheory characters. I managed to just so survive the first chapter which is a dense write up on #vectorspace #module theory. I'm able to read on, chapter 2 and 3 were still possible, but are there better entries into representations and characters?

Appreciate advice on:
1. easier but more specific exposition of central results?
2. advantages of developing everything in the A-module framework?
3. which chapters are key for subsequent developments?
4. which theorems are the ultimate highlights and which are 'just' key to the particular question at hand?
5. What's particularly worthwhile in the presentation approach taken in this book?

@remi Using this for multiple tags and users would turn #Mastodon into a #vectorspace (linear combination of different feeds) where you can dial in your exact position of interest. "I need to #mastoart a little to the left and #math a little more towards to Z axis"

#Meaning is usually described with #VectorSpace #Semantics as in the article below comparing the works from #CAShannon and #AMTuring:

https://www.journals.uchicago.edu/doi/full/10.1093/bjps/axx029

Basically, what vector space semantics says is that the meaning of a message depends on the #Context provided by the sender’s and the receiver’s #DynamicalSystem #Knowledge #State.

As they are two different physical entities they will obviously be in different states, so the two meaning can never be exactly the same.

Serious Abe Weissman Moment - Marvelous Mrs. Maisel

YouTube