A lot of #MachineLearning and #PredictiveModelling in #statistics is based on minimisation of loss with respect to a training data set. This assumes that the training data set as a whole is representative of potential training sets. Consequently, this implies that loss minimisation is not an appropriate approach (or way of conceptualising the problem) in problems where the training data sets are not representative of the potential testing sets. (As a working title, let's call this issue "radical nonstationarity".)

I recently read Javed & Sutton 2024 "The Big World Hypothesis and its Ramifications for Artificial Intelligence" (https://web.archive.org/web/20250203053026/https://openreview.net/forum?id=Sv7DazuCn8) and think it describes a superset of this issue of radical nonstationarity. I strongly recommend this paper for motivating why loss minimisation with respect to a training data set might not always be appropriate.

Imagine an intelligent agent existing over time in a "big world" environment. Each observation records information about a single interaction of the agent with it's environment, and this observation only records the locally observable part of the environment. The agent may be moving between locations in the environment that are radically different with respect to the predictive relationships that exist and the variables that are predictive of the outcome of interest may vary between observations. Nonetheless, there is some predictive information that an intelligent agent could exploit. The case where everything is totally random and unpredictable is of no interest when the focus of research is an intelligent agent. In such a world minimising loss with respect to the history of all observations seen by the agent or even a sliding window of recent history seems irrelevant to the point of obtuseness.

One possible approach to this issue might be for the agent to determine, on a per-observation basis, the subset of past observations that are most relevant to making a prediction for the current observation. Then loss minimisation might play some role in determining or using that subset. However, that use of a dynamically determined training set is not the same thing as loss minimisation with respect to a statically given training set.

I am trying to find pointers to scholarly literature that discusses this issue (i.e. situations where minimisation of loss with respect to some "fixed" training set). My problem is that I am struggling to come up with search terms to find them. So:

* Please suggest search terms that might help me find this literature
* Please provide pointers to relevant papers

#PhilosophyOfStatistics #PhilosophyOfMachineLearning #CognitiveRobotics #MathematicalPsychology #MathPsych #CognitiveScience #CogSci #CognitiveNeuroscience #nonstationarity #LossMinimisation

The Big World Hypothesis and its Ramifications for Artificial...

The big world hypothesis says that for many learning problems, the world is multiple orders of magnitude larger than the agent. The agent neither fully perceives the state of the world nor can it...

OpenReview

I recently read @djnavarro 's 2021 paper "If Mathematical Psychology Did Not Exist We Might Need to Invent It: A Comment on Theory Building in Psychology" (https://doi.org/10.1177/1745691620974769).

It's a gem on the role and use of theory in cognitive psychology (and related fields, by extension) and the relation of theory to statistics. As expected, the footnotes are a joy. For my extra reading pleasure, I imagined the paper written in Danielle's sweary-blog style.

#theory #CognitivePsychology #CogPsych #MathematicalPsychology #MathPsych #MetaScience #paper #paper

If Mathematical Psychology Did Not Exist We Might Need to Invent It: A Comment on Theory Building in Psychology - Danielle J. Navarro, 2021

It is commonplace, when discussing the subject of psychological theory, to write articles from the assumption that psychology differs from the physical sciences...

Sage Journals

The next VSAonline webinar is at 17:00 UTC (not the usual time), Monday 27 January.

Zoom: https://ltu-se.zoom.us/j/65564790287

WEB: https://bit.ly/vsaonline

Speaker: Anthony Thomas from UC Davis, USA

Title: ”Sketching a Picture of Vector Symbolic Architectures”

Abstract : Sketching algorithms are a broad area of research in theoretical computer science and numerical analysis that aim to distil data into a simple summary, called a "sketch," that retains some essential notion of structure while being much more efficient to store, query, and transmit.

Vector-symbolic architectures (VSAs) are an approach to computing on data represented using random vectors, and provide an elegant conceptual framework for realizing a wide variety of data structures and algorithms in a way that lends itself to implementation in highly-parallel and energy-efficient computer hardware.

Sketching algorithms and VSA have a substantial degree of consonance in their methods, motivations, and applications. In this tutorial style talk, I will discuss some of the connections between these two fields, focusing, in particular, on the connections between VSA and tensor-sketches, a family of sketching algorithms concerned with the setting in which the data being sketched can be decomposed into Kronecker (tensor) products between more primitive objects. This is exactly the situation of interest in VSA and the two fields have arrived at strikingly similar solutions to this problem.

#VectorSymbolicArchitectures #VSA #HyperdimensionalComputing #HDC #AI #ML #ComputationalCognitiveScience #CompCogSci #MathematicalPsychology #MathPsych #CognitiveScience #CogSci @cogsci

Join our Cloud HD Video Meeting

Zoom is the leader in modern enterprise video communications, with an easy, reliable cloud platform for video and audio conferencing, chat, and webinars across mobile, desktop, and room systems. Zoom Rooms is the original software-based conference room solution used around the world in board, conference, huddle, and training rooms, as well as executive offices and classrooms. Founded in 2011, Zoom helps businesses and organizations bring their teams together in a frictionless environment to get more done. Zoom is a publicly traded company headquartered in San Jose, CA.

Zoom Video

The schedule for the next VSAonline webinar series (January to June 2025) is published at:

https://sites.google.com/view/hdvsaonline/spring-2025

There are 11 talks around #VectorSymbolicArchitecture / #HyperdimensionalComputing

The talks are (almost always) recorded and published online, in case you can't participate in the live session.

@cogsci
#VSA #HDC #CompCogScii #MathPsych #AI #neuromorphic #neurosymbolic #ComputationalNeuroscience #ComputationalCognitiveScience #MathematicalPsychology

VSAONLINE - Spring 2025

CHECK THE UPCOMING EVENTS TOWARDS THE END OF THIS PAGE!

@cian
If (a big if) we performed generalisation at retrieval (rather than at storage, as in almost all current artificial neural networks) then the episodic memories would be the essential input to the generalisation (and inference) process. You are best placed to know what dimensions to abstract over when you have a specific current task and goal to drive generalisation and inference.

(Of course, having arrived at some specific generalisation from the current retrieval, that generalisation might be stored as part of the current episodic memory and be available to guide future generalisations on retrieval.)

What are the implications if episodic memory is the primary form of memory and other (declarative/procedural/etc) memories are epiphenomena arising out of the episodic memories?

#CogSci #CognitiveScience #MathPsych #MathematicalPsychology @cogsci

Maths/CogSci/MathPsych lazyweb: Are there any algebras in which you have subtraction but don't have negative values? Pointers appreciated. I am hoping that the abstract maths might shed some light on a problem in cognitive modelling.

The context is that I am interested in formal models of cognitive representations and I want to represent things (e.g. cats), don't believe that we should be able to represent negated things (i.e. I don't think it should be able to represent anti-cats), but it makes sense to subtract representations (e.g. remove the representation of a cat from the representation of a cat and a dog, leaving only the representation of the dog).

This *might* also be related to non-negative factorisation: (https://en.wikipedia.org/wiki/Non-negative_matrix_factorization) in that we want to represent a situation in terms of parts and don't allow anti-parts.

#mathematics #algebra #AbstractAlgebra #CogSci @cogsci #CognitiveScience #MathPsych #MathematicalPsychology

Non-negative matrix factorization - Wikipedia

Most of the Artificial Neural Net simulation research I have seen (say, at venues like NeurIPS) seems to take a *very* simple conceptual approach to analysis of simulation results - just treat everything as independent observations with fixed effects conditions, when it might be better conceptualised as random effects and repeated measures. Do other people think this? Does anyone have views on whether it would be worthwhile doing more complex analyses and whether the typical publication venues would accept those more complex analyses? Are there any guides to appropriate analyses for simulation results, e.g what to do with the results coming from multi-fold cross-validation (I presume the results are not independent across folds because they share cases).

@cogsci #CogSci #CognitiveScience #MathPsych #MathematicalPsychology #NeuralNetworks #MachineLearning

New preprint:
“Algebras of actions in an agent’s representations of the world”

https://scholar.social/@E_Mondragon/111177742052625126

#CogSci #CompCogSci #MathPsych #preprint #NewPaper

Esther Mondragón (@E_Mondragon@scholar.social)

New preprint! “Algebras of actions in an agent’s representations of the world” with Alex Dean (first author) and Eduardo Alonso https://arxiv.org/abs/2310.01536. The paper (1) proposes a framework for deriving the algebras of the transformation of worlds due to the actions of an agent, (2) derives symmetry-based disentangled representations (Higgins et al., 2018) using our framework, (3) proposes the algebras of the transformations of worlds using our framework,

Scholar Social

Here is a new version of our tutorial on fitting joint models of M/EEG and behavior! This version is much improved to the previous one due to a ton of additional work by co-author Kianté Fernandez as well as helpful reviewer comments.
https://psyarxiv.com/vf6t5

#eeg @eeg #cognition #MathPsych

See also the associated R package by Kianté here:
https://github.com/kiante-fernandez/Rhddmjags
Or the original Python scripts here:
https://github.com/mdnunez/pyhddmjags