RESEARCH
Papers, ideas, and ongoing investigations.
             _.-;;-._
      '-..-'|   ||   |
      '-..-'|_.-;;-._|
      '-..-'|   ||   |
jgs   '-..-'|_.-''-._|

AREAS
Energy based self learning engines
Reasoning and symbolic augmentation for Spatial Systems
Computational Neuroscience and Neural Modeling of Human Like Long Context Memory
Agent-centric AI architecture integrating long-term memory, adaptive retrieval, and reasoning pipelines
Controlled semantic drift modeling in persistent LLM memory architectures

PAPERS
Priyam Ghosh, Krish Jasiwal, Sauhard Gupta ( MetaCognition Labs )
ICLR Workshop Submission 2026
This work introduces a unified, energy-based framework for building a self-evolving memory substrate for AI agents. Instead of treating memory as static storage, we model it as a continuous dynamical system that integrates associative recall, adaptive decay, semantic diffusion, competition, and replay into a single convergent formulation. The result is a scalable memory layer that reorganizes itself over time, enabling long-term stability, abstraction, and persistent identity in artificial systems.
Ananta Research
ICLR Workshop Submission 2026
We introduce a recursive, energy-based self-learning architecture for neuro-symbolic scientific reasoning that tightly integrates a Hybrid Math-Text Tokenizer, a formal Recursive Logic Subsystem, and an Energy-Based Transformer. By coupling symbolic step-wise verification with energy minimization over derivation traces, the system forms a self-reinforcing loop that prioritizes process-level correctness, stability, and sample efficiency over raw likelihood. This framework provides a principled path toward scalable, verifiable scientific intelligence capable of autonomously formalizing and refining its own reasoning.
Ananta Research
ResearchGate
This analysis examines the proposed Recursive Logic Subsystem (RLS), a neuro-symbolic architecture that recursively integrates self-learning, formal verification, and dynamic curriculum design to enhance the reliability of scientific reasoning in LLMs. By coupling generative models with a logic engine in a co-evolutionary loop, the system aims to provide process-level verification, mitigate hallucinations, and expand formal knowledge through autoformalization. Despite significant challenges in computation, brittleness, and alignment, the RLS outlines a concrete roadmap toward robust, autonomous “Agentic Science.”
Ananta Research
ResearchGate
This paper argues that LLM brittleness in formal reasoning stems from a fundamental “tokenization bottleneck,” where subword schemes fragment mathematically meaningful structures. We propose the Hybrid Math-Text Tokenizer (HMTT), a structure-aware dual-stream tokenizer that preserves atomic formal units using LaTeX and AST parsing, significantly improving reasoning performance across math and code benchmarks. Empirically, HMTT boosts MATH accuracy by over 12 points and introduces a Tokenization Fidelity Score (TFS), strongly correlated with downstream reasoning gains.
Ananta Research
ResearchGate
We introduce Ananta, a novel large language model (LLM) architecture designed for advanced symbolic and scientific reasoning. Ananta uniquely integrates self-supervised learning and reinforcement learning techniques to derive physical equations and generate algorithms from first principles. Its training leverages cu-rated datasets spanning mathematics and scientific derivations (e.g., DeepMind Mathematics , GSM8K, MathQA, and symbolic derivations from arXiv papers) to validate its reasoning abilities. Key innovations include symbolic curriculum learning for gradually increasing problem complexity, recursive logic model-ing to enable multi-step derivations, and hybrid math-text tokenization that treats mathematical symbols and expressions distinctly. We compare Ananta's architecture and fine-tuning (us-ing LoRA, RLHF, PPO) against leading models such as DeepSeek-R1, AlphaEvolve, and GPT-4. Extensive experiments on standard math and physics benchmarks demonstrate that Ananta achieves higher symbolic consistency and derivation completeness while maintaining competitive accuracy. We present architectural diagrams, training pipelines, and pseudocode for Ananta's algorithm derivation modules, along with evaluations on novel reasoning generation. Our results suggest that Ananta advances the state of the art in neural symbolic reasoning, opening avenues for AI-driven scientific discovery.
Priyam Ghosh
CVPRW 2024
We propose an Explicit Temporal Attention mechanism for video diffusion models that enhances temporal coherence and sample quality in video generation tasks. Our approach extends standard 3D U-Net diffusion architectures by inserting specialized attention blocks that explicitly compute attention across the temporal dimension. We derive the full forward and reverse diffusion equations and incorporate multi-frame conditioning via temporal attention. Experiments on UCF-101 and Kinetics-600 show that our method achieves superior Fréchet Video Distance (FVD) and lower Temporal-MSE compared to leading baselines (e.g., Video Diffusion [1], W.A.L.T [3]), demonstrating improved frame-to-frame consistency. We provide extensive implementation details (training time, hardware, hyperparameters) and ablation studies, and discuss limitations and future work. Pseudocode is given for the training loop, temporal attention block, and denoising forward pass.
Priyam Ghosh, Siddhant Shivam ( Elysium )
Earth Prize 2022, Global Talent Week 2022
Microplastics (< 5 mm) pose serious ecological and health risks by entering aquatic food chains and carrying toxic pollutants. This paper presents M2D2 (Mobile Microplastic Diminution Device), an autonomous system that (1) collects microplastics via a Bulk Acoustic Wave (BAW) microfluidic module, (2) identifies and quantifies them using a Digital In-line Holographic Microscope (DIHM) with 3D reconstruction, and (3) navigates toward high-concentration regions using onboard processing (Raspberry Pi) with a CNN+RIHVR feedback loop. In laboratory tests, the BAW module effectively focused and extracted microplastic particles into a central outlet, in agreement with theoretical models of acoustophoretic force (e.g., particles with positive acoustic contrast factor move to pressure nodes). The DIHM captured interference patterns of particles in water, enabling 3D reconstruction of particle fields. A convolutional neural network trained on RIHVR-processed holograms achieved rapid concentration mapping, with predictions closely correlating (≈ 0.95) with full RIHVR results. Overall, the integrated prototype demonstrated high collection efficiency for particles above a few microns and accurate identification of particle fields. This approach is cost-effective (uses inexpensive hardware) and scalable, offering a promising solution to reduce microplastic pollution in waters
Papers will be listed here as they are published.

NOTES
Technical notes and informal write-ups.
forgetting is doing the heavy lifting
Feb 2026 · ramble
was reading some old neuroscience notes at 2am and had this moment. we keep trying to make AI remember everything and it's kind of the wrong instinct? like your brain actively throws stuff away and that's WHY you can think clearly. that's literally what our memory decay paper is about. I kept arguing with krish about whether drift was a bug or a feature and somewhere around the third whiteboard session it clicked. it's the feature. the fact that memories blur and merge is what gives you abstractions. a KV cache that never forgets isn't memory, it's a pile.
the tokenizer was the problem the whole time
Jan 2026 · rant
ok so we were debugging this thing for weeks. model keeps getting basic integrals wrong, like embarrassingly wrong. tried everything. more data, better prompts, reward shaping, nothing. then one night I just dumped the raw token IDs and saw that BPE was splitting "∫" and "dx" into completely unrelated fragments. the model never even got a chance. it was literally solving a jigsaw puzzle where someone cut the pieces wrong. that's when we said screw it and built HMTT from scratch. honestly the whole paper came from staring at tokenizer outputs and going "wait, THAT'S what it sees?"
why I keep going back to energy models
Dec 2025 · thought
I don't know how to explain this without sounding pretentious but when we rewrote ananta's objective from cross-entropy to energy minimization, something shifted in how I think about all of this. loss functions feel like accounting. energy feels like physics. like you're not telling the model "get closer to this answer," you're saying "find the stable configuration." and suddenly things like correctness and robustness stop being separate goals, they're just what low-energy states look like. I know that sounds hand-wavy. I'm still trying to formalize the intuition properly. but it changed something for me.
your agent doesn't remember you
Nov 2025 · note
tried using a bunch of agent frameworks last month and it hit me how hollow they feel. you talk to them for an hour, close the tab, come back and they have no idea who you are. no continuity. no "oh yeah we were working on that thing." it's like groundhog day every session. and people want to fix this by making the context window bigger? that's not memory, that's cramming. real memory is messy. it decays, it reorganizes, it connects things that happened months apart. that's what I want to build. not a bigger notepad, an actual sense of "I've been here before."
the time we built a robot to eat microplastics
Oct 2025 · reflection
someone once looked at my publication list and was like "why is there a microplastics paper between two AI papers" lol fair. but that project with siddhant was probably the most fun I've had doing research. we literally strapped a holographic microscope and a raspberry pi onto a thing that floats, wrote a CNN to tell it where the plastic is, and used acoustic waves to suck it up. half the time we were debugging water leaks not code. presented it at global talent week and people kept asking if it was real. it was real. it also broke twice during the demo. anyway I think the best research training is building something physical that can fail in ways your simulator never warned you about.