May 1, 2026

Citation hunting eats more hours than most researchers admit. A normal submission burns 4 to 12 hours, and almost all of it is spent on a few specific tasks that compound poorly. This post breaks down where the hours go, why traditional keyword search keeps failing, and a workflow that works well in practice. Paste a paragraph from your draft into Underleaf's Find Citations tool, and ranked results come back in seconds. Triage them, capture BibTeX, move on. Half a day of citation hunting becomes minutes.
When you sit down to find citations for a paragraph in your draft, the clock starts. Time tends to disappear into four buckets:
Buckets one and two are the biggest. They're also the ones that scale worst as a paper grows. A 12-page conference submission with 30 citations easily burns half a day if you do it the traditional way.
Google Scholar, Semantic Scholar's legacy keyword index, and library catalogs all match on tokens. That works when you already know the field's vocabulary. It breaks when you don't.
Two researchers can describe the same idea with completely different words. “In-context learning,” “few-shot prompting,” “prompt-based adaptation:” pick the wrong term and you miss half the relevant literature. For cross-disciplinary work it's worse: “graph neural network” in CS overlaps heavily with concepts called “message passing” in chemistry, but the search engines treat them as unrelated.
The tell-tale symptom: you find the perfect paper, look at its reference list, and discover three other relevant works you never would have surfaced from your keywords. That's wasted hours, every time.
Paste a paragraph from your draft and Underleaf's semantic search ranks the most relevant arXiv papers in seconds. Free, no signup required for the first three searches.
Try the AI Citation FinderA semantic citation finder embeds your paragraph as a vector in a high-dimensional space, then finds the nearest papers by meaning rather than overlapping tokens. Two practical consequences:
Underleaf's Find Citations tool runs this matching across 3 million+ arXiv preprints and returns the ten most semantically similar papers in seconds, with abstracts, authors, publication dates, and a ready-to-paste BibTeX entry per result. The whole round-trip, from pasted paragraph to ranked results, takes about as long as a single keyword query on Google Scholar.
Here's a sequence that works well in practice. Try it on a section of your next draft and adapt from there:
Write the paragraph or section without breaking flow to find sources. Mark spots with [CITE] placeholders. Citation hunting in the middle of a writing session destroys momentum and produces worse prose.
When you finish a section, do all the citation searches in one sitting. Same context, same vocabulary, same mental model. Far less context-switching cost than searching ad hoc throughout the day.
For each [CITE] placeholder, copy the surrounding 2–4 sentences (not just the claim) into a semantic search tool. The extra context dramatically improves match quality. A common mistake is pasting a one-line claim, since embeddings need signal. With a full paragraph, Underleaf returns ranked results in seconds.
For each result, give yourself 30 seconds: read the title, scan the abstract, open the PDF only if it's a likely keeper. Resist the urge to read top-to-bottom on every result. You're building a candidate set, not committing yet.
When you find a keeper, copy the BibTeX entry to your .bib file before moving on. Don't leave it for later. The cost of re-finding a paper when you can't remember its title is high, and unnecessary.
For claims you're not sure are well-supported, run the same semantic search but read the 10th result, not the 1st. If your claim is well-established it shows up everywhere; if only a few results vaguely support it, that's a signal to soften your prose.
For a typical 30-citation paper, the workflow above lands roughly here:
The core task, finding the right paper for each citation, drops from hours of keyword fiddling to about 25 minutes total. Most of that 25 minutes is reading, not searching, since each search returns ranked results in seconds. The optional chasing and cleanup steps are research work that happens regardless of which tool you use to find the initial paper.
The savings come from cutting query reformulation to near-zero (paste a paragraph, get matches by meaning) and skipping most of the skim-and-discard step (semantically ranked results have a much higher keeper rate at the top).
.bib file with consistent keys is worth the discipline. See our BibTeX guide for the conventions.Citation hunting is a workflow problem, not a research problem. The time disappears into query reformulation and skim-and-discard, and keyword search keeps both buckets large because you can never quite guess the right vocabulary. Semantic search collapses both. Paste a paragraph, get matches by meaning in seconds, capture BibTeX, move on.
For a 30-citation paper, that's the difference between half a day of citation hunting and minutes. Multiply across a year of writing and the compounding savings are substantial.
Paste a paragraph from your draft and see the workflow above in action. Free to try, no signup, no credit card. Works on any browser.
Open Find CitationsEmpowering students and researchers with AI-powered tools for academic writing.
Go to appContact us© 2026 Underleaf. All rights reserved.