๐Ÿง  New comprehensive review on #LowDimensional #embeddings of #HighDimensional data. Discusses how #dimensionalityreduction helps visualizing, exploring, and #modeling #ComplexSystems. From #PCA to #tSNE, #UMAP & #NeuralNetworks: Excellent overview paper๐Ÿ‘Œ

๐ŸŒ https://arxiv.org/abs/2508.15929

#CompNeuro #MachineLearning #DataVisualization

We are excited to welcome Prof. Alejandro Rodriguez Garcia from the Abdus Salam International Centre for Theoretical Physics (ICTP) to Enabla! In his lecture, Alex continues the topic started by Marcello and explores the use of unsupervised machine learning techniques in many-body quantum systems, highlighting how dimensionality reduction can illuminate structure within complex data. Particular emphasis is placed on the Principal Component Analysis (PCA) as a key method to maximize variance while reducing dimensionality. This lecture sets the stage for future topics such as clustering and manifold learning.

๐ŸŽฅ Join us for this #OpenAccess lecture and take advantage of Enabla's unique features to ask questions directly to Prof. Rodriguez Garcia and engage in discussions with the community: https://enabla.com/pub/1112/about

Don't miss this opportunity to enhance your knowledge in the intersection of data mining and quantum physics!

#MachineLearning #UnsupervisedLearning #DimensionalityReduction #QuantumSystems #PCA #DataMining #OpenScience

Biologists, stop putting UMAP plots in your papers

#UMAP is a powerful tool for exploratory data analysis, but without a clear understanding of how it works, it can easily lead to confusion and misinterpretation.

https://simplystatistics.org/posts/2024-12-23-biologists-stop-including-umap-plots-in-your-papers/

#clustering #dimensionalityreduction #dataviz

Simply Statistics: Biologists, stop putting UMAP plots in your papers

UMAP is a powerful tool for exploratory data analysis, but without a clear understanding of how it works, it can easily lead to confusion and misinterpretation.

Simply Statistics

We just completed a new course on #DimensionalityReduction in #Neuroscience, and the full teaching material ๐Ÿ๐Ÿ’ป is now freely available (CC BY 4.0 license):

๐ŸŒ https://www.fabriziomusacchio.com/blog/2024-10-24-dimensionality_reduction_in_neuroscience/

The course is designed to provide an introductory overview of the application of dimensionality reduction techniques for neuroscientists and data scientists alike, focusing on how to handle the increasingly high-dimensional datasets generated by modern neuroscience research.

#PythonTutorial #CompNeuro

New teaching material: Dimensionality reduction in neuroscience

We just completed a new two-day course on Dimensionality Reduction in Neuroscience, and I am pleased to announce that the full teaching material is now freely available under a Creative Commons (CC BY 4.0) license. This course is designed to provide an introductory overview of the application of dimensionality reduction techniques for neuroscientists and data scientists alike, focusing on how to handle the increasingly high-dimensional datasets generated by modern neuroscience research.

Fabrizio Musacchio
Untangling complexity: harnessing PCA for data dimensionality reduction

This tutorial explores the use of Principal Component Analysis (PCA), a powerful tool for reducing the complexity of high-dimensional data. By delving into both the theoretical underpinnings and practical Python applications, we illuminate how PCA can reveal hidden structures within data and make it more manageable for analysis.

Fabrizio Musacchio
By the way this is the original article that presents t-SNE. Published 11/2008
https://jmlr.org/papers/volume9/vandermaaten08a/vandermaaten08a.pdf
T-distributed stochastic neighbor embedding (t-SNE) is a statistical method for visualizing high-dimensional data in 2 or 3 dimensions.
#DataVisualization #tSNE #MachineLearning #DimensionalityReduction #DataScience #AI #DataAnalysis #DataAnalytics
CSN - Collection Space Navigator

CSN - Collection Space Navigator: Interactive Visualization Interface for Multidimensional Datasets.

โ€œRegardless of how we do dimensionality reduction, if the assumptions and biases underlying a method are not understood then it can be possible to see things in the data that arenโ€™t there. โ€œ

#PCA #DimensionalityReduction #statistics

https://doi.org/10.1073/pnas.2319169120

Why the simplest explanation isnโ€™t always the best: #DimensionalityReduction such as #PCA can see structures that do not exist and miss structures that exist. The simplest explanation isnโ€™t always the best.

โœ๏ธ Dyer & Kording ( @kordinglab) (2023)
๐ŸŒ https://www.pnas.org/doi/10.1073/pnas.2319169120

#DataAnalysis #CompNeuro