Now here's a cool use for AI!

"DeepMind’s GraphCast and WindBorne’s WeatherMesh can both run on consumer-grade computers, and can complete in a matter of seconds work that would usually require a supercomputer under traditional models."

#ai #DeepLearninig #weather

https://www.semafor.com/article/02/14/2024/windborne-takes-the-ai-weather-prediction-crown

Little-known startup takes the AI weather prediction crown | Semafor

WindBorne is using low-cost weather balloons to gather detailed data, giving it an edge in predictive capabilities.

Interpolation Can Provably Preclude Invariance https://arxiv.org/abs/2211.15724
#Overfitting to the point of #interpolation can hinder invariance-inducing objectives: One cannot assume a #DeepLearninig model with an invariance penalty will indeed achieve any form of #invariance… suggests that “benign overfitting,” in which models generalize well despite interpolating, might not favorably extend to settings in which #robustness or #fairness are desirable.
Malign Overfitting: Interpolation Can Provably Preclude Invariance

Learned classifiers should often possess certain invariance properties meant to encourage fairness, robustness, or out-of-distribution generalization. However, multiple recent works empirically demonstrate that common invariance-inducing regularizers are ineffective in the over-parameterized regime, in which classifiers perfectly fit (i.e. interpolate) the training data. This suggests that the phenomenon of "benign overfitting", in which models generalize well despite interpolating, might not favorably extend to settings in which robustness or fairness are desirable. In this work we provide a theoretical justification for these observations. We prove that -- even in the simplest of settings -- any interpolating learning rule (with arbitrarily small margin) will not satisfy these invariance properties. We then propose and analyze an algorithm that -- in the same setting -- successfully learns a non-interpolating classifier that is provably invariant. We validate our theoretical observations on simulated data and the Waterbirds dataset.

arXiv.org

Fascinating #DeepLearninig paper about how #LargeLanguageModel exhibit new (emergent) capabilities when they scale up in size. It seems that as the models get bigger they start to be able to tackle new problems that they weren't explicitly trained on (e.g. arithmetic).

https://arxiv.org/abs/2206.07682

Emergent Abilities of Large Language Models

Scaling up language models has been shown to predictably improve performance and sample efficiency on a wide range of downstream tasks. This paper instead discusses an unpredictable phenomenon that we refer to as emergent abilities of large language models. We consider an ability to be emergent if it is not present in smaller models but is present in larger models. Thus, emergent abilities cannot be predicted simply by extrapolating the performance of smaller models. The existence of such emergence implies that additional scaling could further expand the range of capabilities of language models.

arXiv.org
Hello 👋 I model human intelligence and underlying brain mechanisms using artificial neural networks. Focusing on vision and language. Incoming prof at EPFL, currently research scientist at MIT, looking to connect with people in #compneuro #neuroscience #NeuroAI #DeepLearninig #MachineLearming #vision #languageprocessing #NeuroscienceMigration #introduction
#Introduction : I'm Paul, a computational semanticist who is currently working in the #ClinicalNLP world. I teach at the #MedicalUniversityOfSouthCarolina (#MUSC) in a joint PhD program with #ClemsonUniversity . I run the #NLPCore , a service center for helping other researchers at MUSC gain access to the power of #NLP , #AI , #ML , #DeepLearninig , and #ShallowLearning for their own agenda. My tech tree: #Linux, #RStats, #Python, #Java, #emacs, #OrgMode, #LaTeX, #git.
MauMau (@[email protected])

Attached: 1 image my trouble brothers ❤️ #catsofmastodon

Mastodon 🐘

I'm so fascinated by the outcome of deep learning hyper networks, even a small 5k train in 256x256 gave me that result.
Tried to create some #cyberpunk inspired magazine covers with #DeepLearninig a bit of tweaking gave good surprisingly good results.

More at this insta account (will try to setup a better alternative next days) https://www.instagram.com/p/CkWshquoZdQ/?igshid=YmMyMTA2M2Y=