May 12, 2026

Search bar? Never heard of her

Beyond Semantic Similarity

AI skips the fancy search bar, and the comments are already fighting about it

TLDR: Researchers say AI can sometimes find better answers by directly combing through raw files instead of relying on a polished search system first. Commenters were split between “wow, simple tools are back” and “this falls apart in messy or multilingual real-world data,” which is why the debate matters.

A new paper basically says: what if artificial intelligence stopped politely asking a search engine for the “top results” and just dug through the files itself with blunt tools like keyword search and simple commands? The researchers claim this rough-and-ready approach can beat more fashionable systems on several tests, especially when the AI needs to hunt for clues step by step instead of grabbing one neat answer upfront. In plain English: sometimes the smartest move is not a magic shortcut, but letting the bot rummage through the closet.

And oh, the crowd had opinions. One camp was instantly like, “Sure, but good luck when the question is in one language and the documents are in another,” turning the comments into a global reality check. Another waved the “it depends!” flag, arguing that old-school text search is great for tidy worlds like code manuals, but could totally fall apart in messier fields like medicine or law where people say the same thing ten different ways. That sparked the big showdown: is this a brilliant rebellion against overhyped search tech, or just a clever trick that works only on the right data?

Then came the fun stuff. One commenter declared “map-reduce might be back”, which is the kind of nerd nostalgia that makes some readers cheer and others feel ancient. Another dreamed that the future “index” might just be a pile of markdown files and a tiny cheap AI model. Translation: the paper dropped a serious research claim, but the comments turned it into a spicy street fight over whether simple tools are secretly winning again.

Key Points

  • The article argues that fixed top-k lexical and semantic retrieval interfaces constrain agentic search workflows.
  • Agentic search tasks often require multi-step evidence gathering, intermediate entity discovery, weak clue combination, and plan revision based on partial evidence.
  • The proposed direct corpus interaction (DCI) approach lets agents search raw corpora directly with tools such as grep, file reads, shell commands, and lightweight scripts.
  • DCI is presented as requiring no embedding model, vector index, retrieval API, or offline indexing, making it adaptable to evolving local corpora.
  • The article reports that DCI outperforms strong sparse, dense, and reranking baselines on several BRIGHT and BEIR datasets and achieves strong results on BrowseComp-Plus and multi-hop QA.

Hottest takes

"Map-reduce as a pattern might be on its way back" — nivekney
"cheap methods like grepping and BM25 just are not going to work well" — HarHarVeryFunny
"Maybe the best 'index' will just be markdown files fed into a tiny LLM model" — 2001zhaozhao
Made with <3 by @siedrix and @shesho from CDMX. Powered by Forge&Hive.