We are excited to welcome Prof. Alejandro Rodriguez Garcia from the Abdus Salam International Centre for Theoretical Physics (ICTP) to Enabla! In his lecture, Alex continues the topic started by Marcello and explores the use of unsupervised machine learning techniques in many-body quantum systems, highlighting how dimensionality reduction can illuminate structure within complex data. Particular emphasis is placed on the Principal Component Analysis (PCA) as a key method to maximize variance while reducing dimensionality. This lecture sets the stage for future topics such as clustering and manifold learning.

πŸŽ₯ Join us for this #OpenAccess lecture and take advantage of Enabla's unique features to ask questions directly to Prof. Rodriguez Garcia and engage in discussions with the community: https://enabla.com/pub/1112/about

Don't miss this opportunity to enhance your knowledge in the intersection of data mining and quantum physics!

#MachineLearning #UnsupervisedLearning #DimensionalityReduction #QuantumSystems #PCA #DataMining #OpenScience

Biologists, stop putting UMAP plots in your papers

#UMAP is a powerful tool for exploratory data analysis, but without a clear understanding of how it works, it can easily lead to confusion and misinterpretation.

https://simplystatistics.org/posts/2024-12-23-biologists-stop-including-umap-plots-in-your-papers/

#clustering #dimensionalityreduction #dataviz

Simply Statistics: Biologists, stop putting UMAP plots in your papers

UMAP is a powerful tool for exploratory data analysis, but without a clear understanding of how it works, it can easily lead to confusion and misinterpretation.

Simply Statistics

We just completed a new course on #DimensionalityReduction in #Neuroscience, and the full teaching material πŸπŸ’» is now freely available (CC BY 4.0 license):

🌍 https://www.fabriziomusacchio.com/blog/2024-10-24-dimensionality_reduction_in_neuroscience/

The course is designed to provide an introductory overview of the application of dimensionality reduction techniques for neuroscientists and data scientists alike, focusing on how to handle the increasingly high-dimensional datasets generated by modern neuroscience research.

#PythonTutorial #CompNeuro

New teaching material: Dimensionality reduction in neuroscience

We just completed a new two-day course on Dimensionality Reduction in Neuroscience, and I am pleased to announce that the full teaching material is now freely available under a Creative Commons (CC BY 4.0) license. This course is designed to provide an introductory overview of the application of dimensionality reduction techniques for neuroscientists and data scientists alike, focusing on how to handle the increasingly high-dimensional datasets generated by modern neuroscience research.

Fabrizio Musacchio
Untangling complexity: harnessing PCA for data dimensionality reduction

This tutorial explores the use of Principal Component Analysis (PCA), a powerful tool for reducing the complexity of high-dimensional data. By delving into both the theoretical underpinnings and practical Python applications, we illuminate how PCA can reveal hidden structures within data and make it more manageable for analysis.

Fabrizio Musacchio
By the way this is the original article that presents t-SNE. Published 11/2008
https://jmlr.org/papers/volume9/vandermaaten08a/vandermaaten08a.pdf
T-distributed stochastic neighbor embedding (t-SNE) is a statistical method for visualizing high-dimensional data in 2 or 3 dimensions.
#DataVisualization #tSNE #MachineLearning #DimensionalityReduction #DataScience #AI #DataAnalysis #DataAnalytics
CSN - Collection Space Navigator

CSN - Collection Space Navigator: Interactive Visualization Interface for Multidimensional Datasets.

β€œRegardless of how we do dimensionality reduction, if the assumptions and biases underlying a method are not understood then it can be possible to see things in the data that aren’t there. β€œ

#PCA #DimensionalityReduction #statistics

https://doi.org/10.1073/pnas.2319169120

Why the simplest explanation isn’t always the best: #DimensionalityReduction such as #PCA can see structures that do not exist and miss structures that exist. The simplest explanation isn’t always the best.

✍️ Dyer & Kording ( @kordinglab) (2023)
🌍 https://www.pnas.org/doi/10.1073/pnas.2319169120

#DataAnalysis #CompNeuro

...
Addendae (cont'd)

Manifold hypothesis
https://en.wikipedia.org/wiki/Manifold_hypothesis

Many high-dimensional data sets (requiring many variables) in the real world actually lie along low-dimensional latent manifolds in that high-dimensional space (described by a smaller number of variables).

This principle may underpin the effectiveness of ML algorithms in describing high-dimensional data sets by considering a few common features.

#ManifoldHypothesis #manifolds #MachineLearning #DimensionalityReduction

Manifold hypothesis - Wikipedia

πŸŒŒπŸ”¬ BEP39: the Dimensionality Reduction-Based Networks proposal (https://docs.google.com/document/d/1GTWsj0MFQedXjOaNk6H0or6IDVFyMAysrJ9I4Zmpz2E/edit?usp=sharing)! Capture high-dimensional brain data complexity and explore their lower-dimensional representation with BIDS. #BrainDataAnalysis #DimensionalityReduction
BEP39 Dimensionality Reduction-based Networks

BIDS Extension Proposal 39 (BEP039): Dimensionality reduction-based networks BIDS extension proposal for dimensionality reductions that produce spatiotemporal components for fMRI/PET/EEG/MEG/iEEG data Available under the CC-BY 4.0 International license. Extension moderators/leads: Arianna Sa...

Google Docs