Search Results - "Moskovitz, Ted"
-
1
Understanding dual process cognition via the minimum description length principle
Published in PLoS computational biology (18-10-2024)“…Dual-process theories play a central role in both psychology and neuroscience, figuring prominently in domains ranging from executive control to reward-based…”
Get full text
Journal Article -
2
A State Representation for Diminishing Rewards
Published 07-09-2023“…A common setting in multitask reinforcement learning (RL) demands that an agent rapidly adapt to various stationary reward functions randomly sampled from a…”
Get full text
Journal Article -
3
A First-Occupancy Representation for Reinforcement Learning
Published 28-09-2021“…Both animals and artificial agents benefit from state representations that support rapid transfer of learning across tasks and which enable them to efficiently…”
Get full text
Journal Article -
4
Transfer RL via the Undo Maps Formalism
Published 25-11-2022“…Transferring knowledge across domains is one of the most fundamental problems in machine learning, but doing so effectively in the context of reinforcement…”
Get full text
Journal Article -
5
A Unified Theory of Dual-Process Control
Published 13-11-2022“…Dual-process theories play a central role in both psychology and neuroscience, figuring prominently in fields ranging from executive control to reward-based…”
Get full text
Journal Article -
6
What needs to go right for an induction head? A mechanistic study of in-context learning circuits and their formation
Published 10-04-2024“…In-context learning is a powerful emergent ability in transformer models. Prior work in mechanistic interpretability has identified a circuit element that may…”
Get full text
Journal Article -
7
Minimum Description Length Control
Published 17-07-2022“…We propose a novel framework for multitask reinforcement learning based on the minimum description length (MDL) principle. In this approach, which we term…”
Get full text
Journal Article -
8
The Transient Nature of Emergent In-Context Learning in Transformers
Published 14-11-2023“…Transformer neural networks can exhibit a surprising capacity for in-context learning (ICL) despite not being explicitly trained for it. Prior work has…”
Get full text
Journal Article -
9
Towards an Understanding of Default Policies in Multitask Policy Optimization
Published 04-11-2021“…Much of the recent success of deep reinforcement learning has been driven by regularized policy optimization (RPO) algorithms with strong performance across…”
Get full text
Journal Article -
10
ReLOAD: Reinforcement Learning with Optimistic Ascent-Descent for Last-Iterate Convergence in Constrained MDPs
Published 02-02-2023“…In recent years, Reinforcement Learning (RL) has been applied to real-world problems with increasing success. Such applications often require to put…”
Get full text
Journal Article -
11
Confronting Reward Model Overoptimization with Constrained RLHF
Published 06-10-2023“…Large language models are typically aligned with human preferences by optimizing $\textit{reward models}$ (RMs) fitted to human feedback. However, human…”
Get full text
Journal Article -
12
Efficient Wasserstein Natural Gradients for Reinforcement Learning
Published 11-10-2020“…A novel optimization approach is proposed for application to policy gradient methods and evolution strategies for reinforcement learning (RL). The procedure…”
Get full text
Journal Article -
13
Tactical Optimism and Pessimism for Deep Reinforcement Learning
Published 07-02-2021“…In recent years, deep off-policy actor-critic algorithms have become a dominant approach to reinforcement learning for continuous control. One of the primary…”
Get full text
Journal Article -
14
First-Order Preconditioning via Hypergradient Descent
Published 18-10-2019“…Standard gradient descent methods are susceptible to a range of issues that can impede training, such as high correlations and different scaling in parameter…”
Get full text
Journal Article