...The pipeline ingests PDFs, splits them into chapters, cleans and preprocesses text, then constructs vector stores for fine-grained chunks, chapter summaries, and book quotes to support nuanced queries. At query time, it anonymizes entities, creates a high-level plan, de-anonymizes and expands that plan into concrete retrieval or reasoning tasks, and executes them in sequence while continuously revising the plan. A key focus is hallucination control: each answer is verified against retrieved context, and responses are reworked when they are not sufficiently grounded in the source documents.