Oshri Naparstek Research Scientist · IBM Research
Oshri Naparstek

Oshri Naparstek

Research Scientist · IBM Research Haifa

Leading the AI for Knowledge group. Exploring multimodal AI, complexity theory, cognitive offloading, and the patterns that connect fields.

1,200+ citations
28+ papers
Master Inventor IBM designation
CVPR 2026 latest publication

The story

I trained my first neural network in high school, in Visual Basic, after reading Kurzweil. After three years in combat engineering, I spent 15 years doing what seemed like different things — applied math, signal processing, defense systems, distributed optimization. It took me a while to realize they were all the same thing.

During my postdoc at Washington University in St. Louis, I got curious about a reference in a widely-cited 2010 paper. It pointed to a 1967 paper nobody had read. I ordered a physical copy through the library. It turned out that 20 years of modern research had been unknowingly rediscovering what was already there. That experience shaped how I work — I always go to the original source, because fields forget.

At Rafael, I worked on reinforcement learning for defense — systems that had to learn in real time, with no room for error. At IBM, I co-authored Granite Vision, created the Real-mm-RAG benchmark, and manage the AI for Knowledge group. Currently a Principal RSM and Master Inventor.

Outside of research, I play bass, guitar, and piano — during my Master's I played bass in a wedding band to pay the bills. I'm a macro insect photographer and freediver. There's something about the small that interests me as much as the large — the complexity of a single compound eye, the patterns in coral. I'm also drawn to philosophy: questions about consciousness, understanding, and what it means to observe. I'm interested in a lot of things — probably too many — and I tend to spread across fields rather than dig into one. But I've found that the interesting ideas usually live at the intersections.

Selected publications

CVPR 2026

Closing the Modality Gap in Vision-Language Models

The modality gap in CLIP-style models hurts robustness. A few lines of linear algebra fix it — no retraining, drop-in for any VLM.

Under review — ICML

Token Maturation

Delayed token commitment for reducing hallucinations. Letting representations mature before making hard decisions.

IBM · Published

Granite Vision

Co-authored IBM's compact 2B parameter vision-language model. Our team at IBM Haifa contributed significantly to its development.

ACL · 21 citations in first year

Real-mm-RAG

An automatically generated benchmark for multimodal retrieval-augmented generation. The community needed it — adoption was faster than expected.

Under review

Col-Bandit

Zero-shot query-time pruning for late-interaction retrieval. Casts ColBERT reranking as a Top-K bandit problem — 5x FLOP reduction, no retraining.

Published

Complexity as Advantage

Complexity as the performance gap between observers with different capabilities. Connects entropy, MDL, regret, and logical depth.

IEEE JSAC · 579 citations

Deep Multi-User Reinforcement Learning for Dynamic Spectrum Access

One of the earlier papers applying deep RL to multi-user wireless networks. Distributed policy learning without centralized coordination.

In progress

Cognitive Offloading in Autonomous Agents

A system where LLM agents learn to replace their own reasoning with verified deterministic code. 67% offloaded, same accuracy, 4x cheaper.

Get in touch

If any of this resonated, or you're thinking about similar problems — I enjoy those conversations. I'm not always fast to reply, but I read everything.