HiCL: Hippocampal-Inspired Continual Learning

arXiv — cs.LGMonday, November 24, 2025 at 5:00:00 AM
arXiv:2508.16651v2 Announce Type: replace Abstract: We propose HiCL, a novel hippocampal-inspired dual-memory continual learning architecture designed to mitigate catastrophic forgetting by using elements inspired by the hippocampal circuitry. Our system encodes inputs through a grid-cell-like layer, followed by sparse pattern separation using a dentate gyrus-inspired module with top-k sparsity. Episodic memory traces are maintained in a CA3-like autoassociative memory. Task-specific processing is dynamically managed via a DG-gated mixture-of-experts mechanism, wherein inputs are routed to experts based on cosine similarity between their normalized sparse DG representations and learned task-specific DG prototypes computed through online exponential moving averages. This biologically grounded yet mathematically principled gating strategy enables differentiable, scalable task-routing without relying on a separate gating network, and enhances the model's adaptability and efficiency in learning multiple sequential tasks. Cortical outputs are consolidated using Elastic Weight Consolidation weighted by inter-task similarity. Crucially, we incorporate prioritized replay of stored patterns to reinforce essential past experiences. Evaluations on standard continual learning benchmarks demonstrate the effectiveness of our architecture in reducing task interference, achieving near state-of-the-art results in continual learning tasks at lower computational costs. Our code is available here https://github.com/kushalk173-sc/HiCL.
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended apps based on your readingExplore all apps
Continue Readings
Attention Projection Mixing and Exogenous Anchors
NeutralArtificial Intelligence
A new study introduces ExoFormer, a transformer model that utilizes exogenous anchor projections to enhance attention mechanisms, addressing the challenge of balancing stability and computational efficiency in deep learning architectures. This model demonstrates improved performance metrics, including a notable increase in downstream accuracy and data efficiency compared to traditional internal-anchor transformers.
User-Oriented Multi-Turn Dialogue Generation with Tool Use at scale
NeutralArtificial Intelligence
A new framework for user-oriented multi-turn dialogue generation has been developed, leveraging large reasoning models (LRMs) to create dynamic, domain-specific tools for task completion. This approach addresses the limitations of existing datasets that rely on static toolsets, enhancing the interaction quality in human-agent collaborations.
Detecting Mental Manipulation in Speech via Synthetic Multi-Speaker Dialogue
NeutralArtificial Intelligence
A new study has introduced the SPEECHMENTALMANIP benchmark, marking the first exploration of mental manipulation detection in spoken dialogues, utilizing synthetic multi-speaker audio to enhance a text-based dataset. This research highlights the challenges of identifying manipulative speech tactics, revealing that models trained on audio exhibit lower recall compared to text.
RULERS: Locked Rubrics and Evidence-Anchored Scoring for Robust LLM Evaluation
PositiveArtificial Intelligence
The recent introduction of RULERS (Rubric Unification, Locking, and Evidence-anchored Robust Scoring) addresses challenges in evaluating large language models (LLMs) by transforming natural language rubrics into executable specifications, thereby enhancing the reliability of assessments.
Rescind: Countering Image Misconduct in Biomedical Publications with Vision-Language and State-Space Modeling
PositiveArtificial Intelligence
A new framework named Rescind has been introduced to combat image manipulation in biomedical publications, addressing the challenges of detecting forgeries that arise from domain-specific artifacts and complex textures. This framework combines vision-language prompting with state-space modeling to enhance the detection and generation of biomedical image forgeries.
Whose Facts Win? LLM Source Preferences under Knowledge Conflicts
NeutralArtificial Intelligence
A recent study examined the preferences of large language models (LLMs) in resolving knowledge conflicts, revealing a tendency to favor information from credible sources like government and newspaper outlets over social media. This research utilized a novel framework to analyze how these source preferences influence LLM outputs.
Predicting Region of Interest in Human Visual Search Based on Statistical Texture and Gabor Features
NeutralArtificial Intelligence
A recent study published on arXiv investigates the relationship between Gabor-based features and gray-level co-occurrence matrix (GLCM) texture features in modeling human visual search behavior. The research proposes two feature-combination pipelines to enhance predictions of human fixation regions using simulated digital breast tomosynthesis images.
Instance-Aligned Captions for Explainable Video Anomaly Detection
NeutralArtificial Intelligence
A new framework for explainable video anomaly detection (VAD) has been introduced, featuring instance-aligned captions that connect textual claims to specific object instances, enhancing the reliability of explanations in safety-critical applications. This approach addresses the limitations of existing methods that often produce incomplete or misaligned descriptions, particularly in scenarios involving multiple entities.

Ready to build your own newsroom?

Subscribe to unlock a personalised feed, podcasts, newsletters, and notifications tailored to the topics you actually care about