Oshri Naparstek
Research Scientist · IBM Research Haifa
Leading the AI for Knowledge group. Exploring multimodal AI, complexity theory, cognitive offloading, and the patterns that connect fields.
The story
I trained my first neural network in high school, in Visual Basic, after reading Kurzweil. After three years in combat engineering, I spent 15 years doing what seemed like different things — applied math, signal processing, defense systems, distributed optimization. It took me a while to realize they were all the same thing.
During my postdoc at Washington University in St. Louis, I got curious about a reference in a widely-cited 2010 paper. It pointed to a 1967 paper nobody had read. I ordered a physical copy through the library. It turned out that 20 years of modern research had been unknowingly rediscovering what was already there. That experience shaped how I work — I always go to the original source, because fields forget.
At Rafael, I worked on reinforcement learning for defense — systems that had to learn in real time, with no room for error. At IBM, I co-authored Granite Vision, created the Real-mm-RAG benchmark, and manage the AI for Knowledge group. Currently a Principal RSM and Master Inventor.
Outside of research, I play bass, guitar, and piano — during my Master's I played bass in a wedding band to pay the bills. I'm a macro insect photographer and freediver. There's something about the small that interests me as much as the large — the complexity of a single compound eye, the patterns in coral. I'm also drawn to philosophy: questions about consciousness, understanding, and what it means to observe. I'm interested in a lot of things — probably too many — and I tend to spread across fields rather than dig into one. But I've found that the interesting ideas usually live at the intersections.
Selected publications
Closing the Modality Gap in Vision-Language Models
The modality gap in CLIP-style models hurts robustness. A few lines of linear algebra fix it — no retraining, drop-in for any VLM.
Token Maturation
Delayed token commitment for reducing hallucinations. Letting representations mature before making hard decisions.
Granite Vision
Co-authored IBM's compact 2B parameter vision-language model. Our team at IBM Haifa contributed significantly to its development.
Real-mm-RAG
An automatically generated benchmark for multimodal retrieval-augmented generation. The community needed it — adoption was faster than expected.
Col-Bandit
Zero-shot query-time pruning for late-interaction retrieval. Casts ColBERT reranking as a Top-K bandit problem — 5x FLOP reduction, no retraining.
Complexity as Advantage
Complexity as the performance gap between observers with different capabilities. Connects entropy, MDL, regret, and logical depth.
Deep Multi-User Reinforcement Learning for Dynamic Spectrum Access
One of the earlier papers applying deep RL to multi-user wireless networks. Distributed policy learning without centralized coordination.
Cognitive Offloading in Autonomous Agents
A system where LLM agents learn to replace their own reasoning with verified deterministic code. 67% offloaded, same accuracy, 4x cheaper.
Get in touch
If any of this resonated, or you're thinking about similar problems — I enjoy those conversations. I'm not always fast to reply, but I read everything.