W := (M₃ × T² × H_{P₅} × S¹) / ~_{BII} (@Obius_Maximus)

Claude의 출력과 오토인코더를 결합해 새로운 모델을 훈련했다는 주장에 대해, 이는 신뢰할 수 없는 마커에 기반한 확률적 설명일 뿐이라는 반박이 제기됐다. 모델 해석과 학습 신호의 신뢰성에 대한 기술적 논쟁으로, AI 연구자들에게 의미 있는 내용이다.

https://x.com/Obius_Maximus/status/2052904599213019200

#claude #autoencoders #modeltraining #llm #airesearch

W := (M₃ × T² × H_{P₅} × S¹) / ~_{BII} (@Obius_Maximus) on X

@WesRoth No they haven't. they just trained another model on Claude's outputs paired with the auto-encoders and it is a generalization of markers that are not reliable in the least for multiple reasons. it is only a probabilistic description of an assumption built on training pairing.

X (formerly Twitter)

Dan McAteer (@daniel_mac8)

Anthropic의 Natural Language Autoencoders가 LLM 메커니즘 해석 가능성 연구에서 매우 인상적인 성과로 언급됐다. 모델의 activation을 언어로 설명하게 하는 접근이 핵심이다.

https://x.com/daniel_mac8/status/2052812665613939066

#anthropic #llm #interpretability #research #autoencoders

Dan McAteer (@daniel_mac8) on X

Anthropic's Natural Language Autoencoders is the most amazing piece of LLM mechanistic interpretability research yet. It's founded on the ability for a language model to verbalize its activations. An activation is the numeric representation of computations inside a model. It's

X (formerly Twitter)
Natural Language Autoencoders

Turning Claude's thoughts into text

When Dimensionality Hurts: The Role of #LLM Embedding Compression for Noisy Regression Tasks https://d.repec.org/n?u=RePEc:arx:papers:2502.02199&r=&r=cmp
"… suggest that the optimal dimensionality is dependent on the signal-to-noise ratio, exposing the necessity of feature compression in high noise environments. The implication of the result is that researchers should consider the #noise of a task when making decisions about the dimensionality of text.

… findings indicate that sentiment and emotion-based representations do not provide inherent advantages over learned latent features, implying that their previous success in similar tasks may be attributed to #regularisation effects rather than intrinsic informativeness."
#ML #autoencoders #Overfitting

I just added some extra chapters on #ANN. Since we are using #autoencoders, I thought it could be useful to provide some general introduction on #NeuralNetworks and how they can be tuned.

'Manifold Learning by Mixture Models of VAEs for Inverse Problems', by Giovanni S. Alberti, Johannes Hertrich, Matteo Santacesaria, Silvia Sciutto.

http://jmlr.org/papers/v25/23-0396.html

#autoencoders #manifold #manifolds

Manifold Learning by Mixture Models of VAEs for Inverse Problems

'The Power of Contrast for Feature Learning: A Theoretical Analysis', by Wenlong Ji, Zhun Deng, Ryumei Nakada, James Zou, Linjun Zhang.

http://jmlr.org/papers/v24/21-1501.html

#autoencoders #supervised #generative

The Power of Contrast for Feature Learning: A Theoretical Analysis

'Be More Active! Understanding the Differences Between Mean and Sampled Representations of Variational Autoencoders', by Lisa Bonheme, Marek Grzes.

http://jmlr.org/papers/v24/21-1145.html

#autoencoders #disentangled #representations

Be More Active! Understanding the Differences Between Mean and Sampled Representations of Variational Autoencoders

New preprint from our group ! 🧠 💻

*Whole-brain modelling of low-dimensional manifold modes reveals organising principle of brain dynamics*
https://www.biorxiv.org/content/10.1101/2023.11.20.567824v1

#brain #modeling #autoEncoders #variationalAutoEncoder #restingStateNetworks #manifold

Real-Time Anomaly Detection of NAB Ambient Temperature Readings using the TensorFlow/Keras Autoencoder

Today we will discuss the anomaly detection in time series data using autoencoders. In this approach, anomalies are data points with considerable reconstruction errors. In the context of predictive…

Digital High Science