This #DeepRL paper from University of Alberta seems quite cool:

"Deep reinforcement learning without experience replay, target networks, or batch updates"

As the title says, they succeeded in training deep RL networks in streaming setting getting rid of replay buffers.
The main tricks for that to work seem to be signal normalization and bounding the step-size 🤯

💻Code: http://github.com/mohmdelsayed/streaming-drl
📄Paper: https://openreview.net/pdf?id=yqQJGTDGXN

#AI #RL #DeepLearning

GitHub - mohmdelsayed/streaming-drl: Deep reinforcement learning without experience replay, target networks, or batch updates.

Deep reinforcement learning without experience replay, target networks, or batch updates. - mohmdelsayed/streaming-drl

GitHub

Глибоке навчання з підкріпленням (Deep RL) поєднує в собі навчання з підкріпленням і глибоке навчання, демонструючи безпрецедентний успіх у вирішенні складних завдань, які колись вважалися недосяжними для машин.

#AI #DeepRL #ML #ШІ

https://thetransmitted.com/ai/oglyad-dosyagnen-u-galuzi-glybokogo-navchannya-z-pidkriplennyam/

Огляд досягнень у галузі глибокого навчання з підкріпленням | TheTransmitted

Глибоке навчання з підкріпленням (Deep RL) поєднує в собі навчання з підкріпленням і глибоке навчання, демонструючи безпрецедентний успіх у вирішенні складних завдань, які колись вважалися недосяжними для машин.

TheTransmitted
My friend Janarthanan Rajendran is recruiting PhD students in deep reinforcement learning at Dalhousie University for Fall 2024. Check out his webpage to learn more: https://sites.google.com/umich.edu/janarthanan-rajendran/prospective_students
#deeprl #deeplearning #dalhousie #phdposition #phd #job
Janarthanan Rajendran - Prospective Students

Thank you for your interest in working with me and sharing the joy of doing research. :) I am committed to building a research group that provides an inclusive environment for everyone. I strongly encourage students from underrepresented groups in Computer Science research to apply. This includes,

Today is the #NeurIPS2022 #DeepRL workshop, starting in just a few hours!

We have two papers there which may be of interest to #MARL #gametheory #RL researchers:

- ESCHER: Eschewing Importance Sampling in Games by Computing a History Value Function to Estimate Regret my McAleer et al. (https://openreview.net/forum?id=GMMdlnRYj4), and

- A Unified Approach to Reinforcement Learning, Quantal Response Equilibria, and Two-Player Zero-Sum Games by Sokota and D'Orazio et al. (https://openreview.net/forum?id=ndZ42T8iUmd),

OpenReview

OpenReview

#AI #MSc_AI #Python #DeepLearning #AIEthics
#computationalcognition #creativeindustries #deeprl #ReinforcementLearning

MSc in AI. Registration is open.

Requirements:
First (or, pending on CV, upper second class) BSc/BEng/BA in computer science, mathematics, physics, computer engineering, psychology or biology.
Competence in Python and Mathematics

Interested?
For more information and to download a booklet, please visit https://cit-ai.net/CitAI-MSc.html

To Apply:
https://www.city.ac.uk/prospective-students/courses/postgraduate/artificial-intelligence/2023#accordion513943-header513943

CitAI-MSc_AI

Last year, a French #ai Athénan won 11/23 of the games at the 24th ICGA Computer Olympiad. Nice.

Even nicer: it ran on a single GPU, yet it challenged cutting-edge #deepRL-based AIs requiring hundreds of GPUs!

Here is the paper about its algorithm (a minimax): https://arxiv.org/pdf/2012.10700.pdf

making my way downtown
collecting MNIST >= 5

#deepRL #Doom

👉 New Preprint available #arXiv
http://arxiv.org/abs/2205.09738
AIGenC is a model for creative problem-solving in a deep reinforcement learning agent as a step forward to solving the problem of generalisation in #AI #deepRL
AIGenC: AI generalisation via creativity

This paper introduces a computational model of creative problem solving in deep reinforcement learning agents, inspired by cognitive theories of creativity. The AIGenC model aims at enabling artificial agents to learn, use and generate transferable representations. AIGenC is embedded in a deep learning architecture that includes three main components: concept processing, reflective reasoning, and blending of concepts. The first component extracts objects and affordances from sensory input and encodes them in a concept space, represented as a hierarchical graph structure. Concept representations are stored in a dual memory system. Goal-directed and temporal information acquired by the agent during deep reinforcement learning enriches the representations creating a higher-level of abstraction in the concept space. In parallel, a process akin to reflective reasoning detects and recovers from memory concepts relevant to the task according to a matching process that calculates a similarity value between the current state and memory graph structures. Once an interaction is finalised, rewards and temporal information are added to the graph structure, creating a higher abstraction level. If reflective reasoning fails to offer a suitable solution, a blending process comes into place to create new concepts by combining past information. We discuss the model's capability to yield better out-of-distribution generalisation in artificial agents, thus advancing toward artificial general intelligence. To the best of our knowledge, this is the first computational model, beyond mere formal theories, that posits a solution to creative problem solving within a deep learning architecture.

arXiv.org