On this page
- TL;DR
- The basic shape of "chat with a PDF"
- Step 1: pick the right tool for the job
- Single document, casual use
- Single document, serious analysis
- A document so big nothing else can hold it
- Many documents, ongoing project
- Confidential documents
- Step 2: ingest the PDF cleanly
- Step 3: prompt like you mean it
- Set the role
- Constrain the scope
- Ask for citations explicitly
- Allow "I don't know"
- Use multi-step prompts for hard questions
- Step 4: verify everything that matters
- Practical recipes
- Recipe: skim a paper in 5 minutes
- Recipe: extract structured data from a contract
- Recipe: compare two documents
- Recipe: build a study guide
- The four failure modes to watch for
- 1. Hallucinated quotes
- 2. Mangled tables
- 3. Lost figures
- 4. Confident wrong page numbers
- Going beyond a single chat
- Privacy: the question to actually ask
- A workflow that scales
- The summary
TL;DR
- Pick the tool by the job: NotebookLM for fast cited Q&A, Claude Opus 4.7 for hard reasoning, Gemini 2.5 Pro for very long PDFs, NovaKit if you want all of them with your own keys.
- The single biggest quality lift is good prompting, not a fancier tool. Ask for citations, define scope, ask for "I don't know" as a valid answer.
- For multi-document work, build a knowledge base instead of re-uploading every time.
- Watch for the four classic failure modes: hallucinated quotes, mangled tables, lost figures, and confident wrong page numbers.
- Your PDFs are data. Treat them like data — know where they're going.
The basic shape of "chat with a PDF"
Every tool that does this works the same way under the hood:
- Ingest — parse the PDF into text (and ideally images of pages).
- Embed or load — either chunk and embed for retrieval, or stuff the whole thing into the context window.
- Retrieve — pull the relevant pieces for your question (or, with long context, just send everything).
- Generate — the model answers using the retrieved pieces, ideally with citations.
In 2026, the long-context approach (skip retrieval, send the whole document) works for most documents under ~300 pages. Above that, retrieval still matters.
Step 1: pick the right tool for the job
Single document, casual use
NotebookLM. Drag in the PDF, start asking. Free. Best citations of any consumer tool.
Single document, serious analysis
Claude Projects or NovaKit + Claude Opus 4.7. Opus 4.7 is the strongest at deep multi-step reasoning over long PDFs in 2026. Pay for the quality.
A document so big nothing else can hold it
Gemini 2.5 Pro (2M tokens). A 600-page government RFP fits in a single context. Or NovaKit pointed at Gemini 2.5 Pro if you want one chat app for everything.
Many documents, ongoing project
A real knowledge base. NotebookLM handles this up to ~50 sources well. Beyond that, build one yourself — see the knowledge base guide.
Confidential documents
Do not upload to a free third-party wrapper. Use Claude.ai (with appropriate plan), ChatGPT Enterprise, or a BYOK tool like NovaKit where the document stays in your browser.
Step 2: ingest the PDF cleanly
Garbage in, garbage out. A few tips:
- Use the original PDF, not a scan of a print of a PDF. Each generation of "print to PDF" loses fidelity.
- If it's a scan, OCR it first. Most modern tools OCR automatically, but quality varies. For dense scans, a dedicated OCR pass (Adobe Acrobat, ABBYY, or a vision model like GPT-5 / Gemini 2.5 Pro reading page images) gives you cleaner text.
- Strip junk pages. Cover pages, blank pages, ads, and table-of-contents pages waste context. Trim them when possible.
- Watch for two-column layouts. Some parsers read across columns instead of down them. If answers come back garbled, try a different tool.
Step 3: prompt like you mean it
Most "the AI hallucinated" complaints are really "I asked a vague question and got a vague answer." Be specific.
Set the role
Open with what kind of expert you want. This nudges the model into the right register.
You are a careful research assistant. Answer only from the document.
If the document doesn't contain the answer, say "Not in the document."
Cite page numbers for every claim.
Constrain the scope
Tell the model what not to do.
Do not bring in outside knowledge. Do not summarize.
Answer the specific question only.
Ask for citations explicitly
Most tools cite by default, but the format varies. Force the format you want:
For each fact, append [p. NN] where NN is the page number.
If a claim spans multiple pages, list all of them.
Allow "I don't know"
This single line saves you from 90% of hallucinations:
If the answer is not in the document, respond exactly with:
"The document does not contain this information."
Do not guess.
Use multi-step prompts for hard questions
Don't ask "summarize this 300-page report" in one shot. Break it down:
- "List the main sections."
- "For each section, summarize in 3 bullets."
- "What are the three biggest risks called out?"
- "Quote the exact passages where each risk is described."
You'll get better answers and you can verify each step.
Step 4: verify everything that matters
The model can be wrong. The citations can be wrong. The page numbers can be wrong.
For anything load-bearing — a contract clause, a study finding, a financial number — open the cited page and check. Two-second sanity check, saves your reputation.
A useful pattern: after the answer, ask the model to quote the exact source passage verbatim. If it can't, it probably made it up.
Practical recipes
Recipe: skim a paper in 5 minutes
1. "What is the main claim of this paper, in one sentence?"
2. "What is the strongest evidence for that claim?"
3. "What are the limitations the authors acknowledge?"
4. "What's the most important table or figure, and what does it show?"
5. "What would a skeptic say is missing?"
This gives you 80% of what a careful reading would, in 5 minutes.
Recipe: extract structured data from a contract
"Extract the following fields as JSON:
- counterparty name
- effective date
- termination clause (full text + page)
- payment terms (amount, frequency, currency)
- governing law
- liability cap
If a field is missing, use null. Cite page numbers."
GPT-5 and Claude Opus 4.7 both nail this. Verify the cap clause manually.
Recipe: compare two documents
Put both in the same context (or in NotebookLM as two sources):
"Compare these two documents on:
1. Pricing terms — list every price difference.
2. Termination conditions — what's different about how each side can exit?
3. IP assignment — who owns what under each?
Format as a table. Cite page numbers for both documents."
Recipe: build a study guide
"Create a study guide from this textbook chapter:
- 10 key terms with definitions
- 5 example problems with worked solutions
- 3 conceptual questions for self-quizzing
- A 5-bullet summary at the end
Cite chapter sections."
NotebookLM is excellent at this and will even generate audio.
The four failure modes to watch for
1. Hallucinated quotes
The model gives you a quote that's not in the document. Always verify direct quotes.
Mitigation: ask for "exact verbatim quote" and check.
2. Mangled tables
Tables in PDFs are notoriously brittle. Numbers swap columns, headers get misread.
Mitigation: for any number that matters, look at the actual page. Vision-capable models (GPT-5, Claude Opus 4.7, Gemini 2.5 Pro) handle tables much better than text-only extraction.
3. Lost figures
If your answer depends on a chart or diagram, make sure the tool actually saw the image. Some tools strip images during ingestion.
Mitigation: use a multimodal model and explicitly ask "describe figure 3 on page 14."
4. Confident wrong page numbers
The model says "page 47" but the citation is actually on page 51. Common when the PDF has front-matter with roman numerals.
Mitigation: check the citation. Don't rely on page numbers alone for legal or academic work.
Going beyond a single chat
When you find yourself uploading the same PDFs over and over, you've outgrown chat. You want a knowledge base — a persistent, searchable, cited collection of documents you can query forever.
Two paths:
- Use NotebookLM as a long-lived notebook. Works for tens of sources. Free.
- Build a real knowledge base. Embed once, query forever. See build an AI knowledge base from PDFs for the full walkthrough.
For team workflows, a shared knowledge base in NovaKit (or self-hosted) beats every "everyone uploads the same contract to ChatPDF" disaster.
Privacy: the question to actually ask
Before you upload, ask: would I be okay if this document leaked to the public tomorrow?
If the answer is "yes" — anything goes. NotebookLM, ChatGPT, ChatPDF, whatever.
If the answer is "no" — your options narrow to:
- A paid tier with a real DPA (Claude Pro/Team/Enterprise, ChatGPT Enterprise, Gemini for Workspace).
- A BYOK setup where the document stays in your browser and only the AI provider sees the API call (NovaKit, similar tools).
- A fully local setup with a local model (slower, weaker, but private).
Don't upload your unsigned NDA to a free random tool. People do this every day.
A workflow that scales
For people who do this often, here's what works:
- Triage with NotebookLM. Drop a batch of papers in. Ask the high-level questions. Find the 2-3 worth deep reading.
- Deep-dive with Claude Opus 4.7. In Claude Projects or NovaKit. Multi-step prompting, careful citation.
- Persist what matters. Either notes back to your own system, or a knowledge base for future query.
- Iterate. As you learn what questions you ask repeatedly, build prompt templates. See 25 prompt templates that work.
The summary
- The tool matters less than the model and the prompt.
- Ask for citations. Allow "I don't know." Verify anything load-bearing.
- For very large documents, use long-context models (Gemini 2.5 Pro, Claude Opus 4.7) instead of fighting with chunking.
- For ongoing work, build a knowledge base instead of re-uploading.
- For confidential work, know where the bytes are going.
That's it. Drop a PDF, ask a question, check the citation. The hard part is the discipline, not the technology.
NovaKit supports PDF uploads against every major model — pick Opus 4.7 for hard questions, Haiku 4.5 for quick ones, Gemini 2.5 Pro for huge documents. Your keys, your documents, your call.