Anthropic has unveiled Citations, a new feature for its developer API designed to enhance the accuracy and transparency of responses from its Claude AI models.
Announced on Thursday, Citations aims to address one of AI’s persistent issues: hallucinations—when models generate inaccurate or fabricated information.
By grounding AI responses in specific source documents such as emails or reports, Citations marks a good step in Anthropic’s efforts to reduce misinformation and increase user trust in AI systems.
How Citations Enhances Transparency
The Citations feature enables developers to upload source files, such as PDFs or plain text, directly into the Claude API. This allows AI models to generate responses that reference specific sections or sentences from these documents.
For example, when summarizing documents or answering queries, the AI can provide detailed references, making its outputs more accountable and easier to verify.
This capability is particularly valuable for tasks like document summarization, customer support, and question-answering, where reliability is non-negotiable.
By embedding these references within responses, Anthropic is setting a new standard for accountability in AI applications, empowering users to verify claims directly against the original source material.
Availability and Model Compatibility
Citations is currently accessible via Anthropic’s API and Google’s Vertex AI platform. However, the feature is exclusive to two advanced models in the Claude family: Claude 3.5 Sonnet and Claude 3.5 Haiku.
This selective rollout ensures the feature is implemented in Anthropic’s most advanced systems, likely to deliver the highest performance for complex use cases.
Pricing and Cost Efficiency
The Citations feature follows a token-based pricing model, with costs varying by the model and document size.
- Claude 3.5 Sonnet: Processing a 100-page document costs approximately $0.30.
- Claude 3.5 Haiku: The same task costs about $0.08.
Cited text does not count toward output token limits, making the feature a cost-effective choice for developers handling large volumes of documents. While the pricing might seem high for large-scale applications, it offers significant value for developers prioritizing accuracy and transparency.
Real-World Applications and Early Success
Thomson Reuters
The global legal and tax advisory service, Thomson Reuters, integrated Citation into its CoCounsel AI platform.
This improved trust in AI outputs by reducing hallucination risks. Jake Heller, Head of Product at CoCounsel, highlighted that Citations’ ability to ground responses in primary sources simplified system maintenance and enhanced legal assistance reliability.
Endex
Endex, an AI platform for financial research, achieved remarkable results with Citations, reducing source hallucinations from 10% to 0% while boosting reference accuracy by 20%.
CEO Tarun Amasa praised the feature for eliminating the complexities of prompt engineering and increasing multi-stage financial research accuracy.
Competing in a Crowded AI Landscape
Citations enters the market amid growing competition, including OpenAI’s Operator, a feature focused on web-based and GUI navigation tasks.
fUnlike Operator, which is currently limited to U.S.-based ChatGPT Pro users, Citations emphasizes source accountability and is available globally through Anthropic’s API and Google Cloud.
This broader accessibility positions Citations as a practical solution for developers working on document-centric workflows, including legal case analysis, financial reporting, and customer support.
Boosting Accuracy with Automation
Internal evaluations reveal that Citations improves recall accuracy by 15%, compared to traditional implementations. This makes it an invaluable tool for industries that demand precision, such as legal, medical, and financial services.
By automating the referencing process, Citations eliminates the need for intricate prompt engineering.
Developers can now upload source documents directly, allowing Claude models to generate responses tied explicitly to the source material. This automation reduces errors, enhances accountability, and simplifies workflows.
Challenges and Limitations
Despite its potential, Citations has limitations –
- Restricted Model Compatibility: The feature is only available for Claude 3.5 Sonnet and Haiku, potentially limiting adoption among users reliant on other Claude models.
- Scalability Concerns: Anthropic has yet to provide performance data for real-time or high-demand applications, raising questions about its scalability.
- Limited Metrics: While promising, real-world performance metrics and additional use cases for Citations are still awaited.
A Game-Changer for High-Trust Industries
Citation has the potential to transform workflows in industries where accuracy and reliability are paramount.
- Legal Services: Summarizing case files with references to rulings or statutes.
- Financial Analysis: Answering complex queries with citations from financial reports.
- Customer Support: Providing accurate responses with references to user manuals or product FAQs.
By meeting the demands of these high-trust industries, Citations underscores Anthropic’s commitment to building trustworthy AI applications.
Future Implications
Anthropic has hinted that Citations is just the beginning of its efforts to enhance AI accountability. As adoption grows, the company is expected to expand the feature’s availability across other Claude models and provide more real-world insights into its performance.
With events like Mobile World Congress 2025 approaching, it’s likely that Anthropic will share additional updates, keeping developers and users eager for more advancements in AI transparency.