December 2025

New frameworks for biologically constrained learning: from similarities in humans and transformers and sparse, contrained RNNs to a unified theory of hippocampal replay

Shared sensitivity to data distribution during learning in humans and transformer networks

Shared sensitivity to data distribution during learning in humans and transformer networks
  • Core Discovery: Humans and transformer networks share a specific sensitivity to the statistical distribution of training data, trading off between “in-context” (inference-based) and “in-weights” (memory-based) learning strategies.
  • Method: The study compared humans (n=530n=530) and transformers on a rule-learning task, manipulating the diversity (Zipfian skewness) and redundancy of examples.
  • Findings:
    • Both learners switch strategies at a similar threshold (α1\alpha \approx 1): high diversity promotes in-context generalization, while high redundancy promotes rote memorization.
    • A composite distribution (balanced diversity and redundancy) allows both systems to acquire both strategies simultaneously.
    • Critical Divergence: Humans benefit from a curriculum that emphasizes diversity early, whereas transformers suffer from catastrophic interference, overwriting early strategies with later ones.
  • Conclusion: While humans and transformers share computational principles regarding data distribution, biological memory constraints allow for flexible curriculum learning that current transformer architectures lack.

Between planning and map building: Prioritizing replay when future goals are uncertain

Between planning and map building: Prioritizing replay when future goals are uncertain
  • Objective: To reconcile the “value” hypothesis (replay plans for current goals) and the “map” hypothesis (replay builds structure) of hippocampal function.
  • Approach: The authors extended a reinforcement learning planning model to include a Geodesic Representation (GR)—a map encoding distances to multiple candidate goals—prioritized by their expected future utility.
  • Results:
    • The model explains “paradoxical” lagged replay (focusing on past rather than current goals) observed in goal-switching tasks as a rational response to uncertainty about future goal statistics.
    • It simultaneously accounts for predictive replay in stable environments where the goal structure is well-learned.
    • Replay prioritization depends on the agent’s learned belief about goal stability and recurrence.
  • Implications: Replay functionally builds a cognitive map (GR) but prioritizes its construction based on future relevance, unifying planning and map-building under a single computational framework.

Constructing biologically constrained RNNs via Dale’s backpropagation and topologically informed pruning

Constructing biologically constrained RNNs via Dale's backpropagation and topologically informed pruning
  • Scope: A new framework for training Recurrent Neural Networks (RNNs) that rigorously adhere to biological constraints: Dale’s principle (separate E/I neurons) and sparse connectivity.
  • Key Themes:
    • Method: Introduces “Dale’s backpropagation” (a projected gradient method) and “top-prob pruning” (probabilistically retaining strong weights) to enforce constraints without performance loss.
    • Performance: Constrained models empirically match the learning capability of conventional, unconstrained RNNs.
    • Application: When trained on mouse visual cortex data, the models inferred connectivity patterns that support predictive coding: feedforward pathways signaling prediction errors and feedback pathways modulating processing based on context.
  • Framework Proposal: This approach provides a mathematically grounded toolkit for constructing anatomically faithful circuit models, bridging the gap between artificial network trainability and biological plausibility.