Skip to main content
Document usage analytics show you which pieces of content in your knowledge base are actually being used by your AI—and which are sitting unused.

Viewing Document Usage

1

Navigate to your datasets

Go to Data → Datasets
2

Select a dataset

Choose the dataset you want to analyze
3

View document list

Browse the list of documents in your dataset
4

Access detailed usage

Click on any document to see comprehensive usage metrics and analytics

Key Metrics per Document

Understanding these metrics helps you identify which documents are performing well and which need attention.

Retrieval Count

How many times this document was retrieved across all traces.
Count RangeWhat It MeansWhat To Do
High (100+)Core document that’s frequently neededKeep this up to date—it’s critical to your AI’s performance
Medium (10-100)Regularly used and importantMonitor for accuracy and keep content fresh
Low (1-10)Occasionally useful for niche topicsVerify the content is still needed and relevant
ZeroNever retrieved, potentially irrelevantConsider removing or improving embeddings

Unique Queries

How many different queries retrieved this document.
A document with 100 retrievals from only 2 unique queries is less valuable than one with 100 retrievals from 50 unique queries.
Uniqueness LevelWhat It MeansUse Case
High UniquenessBroadly applicable contentAnswers many different questions and serves diverse use cases
Low UniquenessNarrow use caseSame questions repeatedly, limited applicability
Real-world comparison:
DocumentRetrievalsUnique QueriesAssessment
Document A20080Versatile content serving many use cases
Document B2005Narrow focus, repetitive queries

Average Relevance Score

Mean relevance score when this document was retrieved.
Score RangeMeaningWhat It Tells You
> 0.8Strong matchHighly relevant, excellent semantic alignment
0.6 - 0.8Good matchRelevant and useful for the queries
0.4 - 0.6Weak matchMarginally relevant, may not be helpful
< 0.4Very weak matchLikely not helpful, poor semantic match
Red Flag Alert: High retrieval count + low average relevance = your retrieval system is finding this document, but it’s not actually a good match.Consider these fixes:
  • Improve the document content to be more focused
  • Fix or regenerate embeddings
  • Adjust retrieval parameters or similarity thresholds

Last Retrieved

When this document was most recently used.
Time RangeStatusWhat It Indicates
Recent (< 7 days)Actively usedCurrently in use for ongoing operations
Moderate (7-30 days)Regularly usedStill relevant and useful
Old (30-90 days)Infrequently usedWorth reviewing for continued relevance
Very Old (> 90 days)Rarely/never usedPossibly outdated or no longer relevant

User Feedback Correlation

How users rated traces that used this document. This metric shows you the quality of your documents from your users’ perspective:

Thumbs Up Count

Traces using this document that received positive feedback

Thumbs Down Count

Traces using this document that received negative feedback

Satisfaction Rate

Percentage of positive feedback overall
Feedback PatternWhat It Might MeanRecommended Action
High Thumbs DownDocument contains wrong/outdated information, is unclear or confusing, or doesn’t answer the questions users are askingReview and update immediately
High Thumbs UpDocument is valuable and helpful to usersKeep it well-maintained, consider creating similar high-quality content, and use it as a template for other documents

Document Details Page

Click on any document to access comprehensive information and insights.

Full Content

The complete text of this document chunk, exactly as stored in your vector database. This lets you see precisely what your AI has access to when this document is retrieved.

Metadata

All fields synced from your vector database provide context about the document:
  • Source — Original file or URL where this content came from
  • Last updated — When the source was last modified
  • Tags — Categories or labels for organization
  • Custom fields — Any other metadata you’ve configured

Usage Timeline

A graph showing document retrievals over time reveals important patterns:
Pattern TypeWhat It Looks LikeWhat To Investigate
Trending DocumentsSudden increase in usageSomething made this document more relevant—find out what changed
Declining DocumentsUsed to be popular, now ignoredInvestigate why usage dropped—may be outdated or replaced
Seasonal PatternsSpikes at certain timesPlan updates around predictable usage patterns

Recent Traces

List of most recent traces that retrieved this document:
  • Trace ID — Link to the full trace for detailed analysis
  • Query — What the user asked
  • Timestamp — When this retrieval happened
  • User feedback — Thumbs up/down if available
Click through to see the full trace and understand exactly how this document was used in context.
Documents that are frequently retrieved alongside this one reveal important relationships:
Insight TypeWhat It RevealsAction To Take
Document ClustersRelated topics that are often retrieved togetherEnsure these documents are well-maintained as a group
Coverage GapsMissing related documents that users needAdd new content to fill the gaps
Redundancy CheckMultiple docs with similar contentConsolidate duplicates for clarity

Sorting and Filtering Documents

Find exactly the documents you need to review with powerful sorting and filtering options.

Sort Options

Sort ByWhat It Shows YouWhen To Use
Most RetrievedYour core documents that power your AIFinding high-impact content to maintain
Least RetrievedUnused content that may need attentionIdentifying candidates for removal
Highest RelevanceBest semantic matches across retrievalsFinding well-matched, quality documents
Lowest RelevancePoorly matched documents needing improvementIdentifying documents to fix or remove
Most RecentRecently added or updated documentsTracking new content performance
OldestLong-standing content that may need refreshFinding potentially outdated documents

Filter Options

Filter TypeExample UseWhat It Helps You Find
Retrieval Count RangeDocuments with 10-100 retrievalsMedium-usage documents for review
Relevance ThresholdOnly docs with >0.7 average relevanceHigh-quality, well-matched documents
Date RangeDocuments retrieved in last 30 daysRecently active content
User FeedbackFilter by positive or negative feedbackDocuments linked to user satisfaction
Source FilterFilter by original file or URLContent from specific sources
Custom MetadataFilter using your custom metadata fieldsDocuments with specific tags or attributes

Use Cases

Here’s how to apply document usage analytics to solve real problems.

Find High-Impact Documents to Update

Goal: Focus your limited time on updating what matters most.
1

Sort by Most Retrieved

Identify your most-used documents
2

Check last updated dates

Review the “Last Updated” date for your top 20 documents
3

Identify outdated high-impact docs

Find old documents with high retrieval counts
4

Update source files

Make improvements to those documents in your original source files
5

Re-sync data source

Sync the updates to your vector database
Result: Maximum impact from your update efforts—you’re improving the documents that matter most.

Remove Unused Documents

Goal: Clean up your vector database for better performance.
1

Filter for zero retrievals

Find documents with zero retrievals in the last 60 days
2

Review each document

Examine the content and metadata
3

Determine the root cause

Ask: “Is this truly irrelevant, or just poorly embedded?”
4

Remove irrelevant docs

Delete truly irrelevant documents from your vector database
5

Re-embed valuable docs

For documents that should be findable, regenerate embeddings
Result: Leaner database, faster retrievals, and better-quality results.

Investigate Low-Quality Documents

Goal: Fix documents that are retrieved but not helpful.
1

Filter for problematic docs

Find documents with high retrieval count + low average relevance (< 0.6)
2

Read the document content

Review what’s actually in the document
3

Check recent traces

Look at traces that recently used this document
4

Diagnose the issue

Determine if the problem is:
  • Content is wrong/outdated → Update the content
  • Content is fine but embeddings are bad → Re-embed the document
  • Retrieval parameters are too loose → Adjust similarity threshold
5

Apply the fix

Make the necessary changes based on your diagnosis
Result: Better retrieval quality and more relevant documents in your results.

Track Content Gaps

Goal: Find topics where you need more documentation.
1

Find traces with poor retrieval

Look at traces with no relevant documents retrieved
2

Group by topic

Organize by topic or query type to find patterns
3

Identify patterns

Example: “All pricing questions have no good docs”
4

Create new content

Write new documentation to fill the identified gaps
5

Verify improvement

Add to vector database and confirm retrieval improves
Result: Comprehensive knowledge base with fewer “I don’t know” responses.

Measure Update Impact

Goal: Verify that updating a document improved performance.
1

Record baseline metrics

Note retrieval metrics before update (count, relevance, feedback)
2

Update the document

Make improvements in your source files
3

Re-sync to Arcbeam

Push the updated document to your vector database
4

Wait for data collection

Allow 1-2 weeks for meaningful data
5

Check new metrics

Review the same metrics again
6

Compare before/after

Analyze the impact of your changes
Result: Data-driven confirmation that your updates actually helped.

Debug Wrong Answers

Goal: Fix AI responses that provide incorrect information.
1

View the trace

When AI gives a wrong answer, view the trace in Arcbeam
2

Check retrieved documents

See which documents were used to generate the response
3

Review document content

Click through to see the full content of each document
4

Diagnose the issue

Determine if:
  • Documents contain incorrect information
  • Correct documents weren’t retrieved
  • Wrong documents were prioritized
5

Fix the root cause

  • If documents are wrong, update the source files
  • If documents are missing, add new content
  • If retrieval is broken, adjust parameters
6

Re-sync and verify

Update your dataset and confirm improved outputs
Result: Trace errors back to source data and fix them systematically.

Compliance and Audit Trail

Goal: Demonstrate data provenance for compliance requirements.
1

Generate trace report

Create a report for the required time period
2

Export document usage

Download data showing which documents were accessed
3

Trace to original files

Use source attribution to link back to original files
4

Document the audit trail

Show complete chain from AI output to source document
5

Record access logs

Document which users accessed which information
Result: Full provenance chain for regulatory compliance and auditing.

Exporting Document Usage Data

Export document metrics for external analysis and reporting.
1

Navigate to datasets

Go to Data → Datasets
2

Select your dataset

Choose the dataset you want to export
3

Click Export

Click Export Usage Data
4

Choose format

Select CSV or JSON format
5

Download

Save the file for analysis

CSV Export Contents

The exported CSV includes all key metrics:
  • Document ID
  • Content (truncated or full based on your selection)
  • Retrieval count
  • Unique queries
  • Average relevance score
  • Last retrieved date
  • User feedback statistics

Common Use Cases for Exports

Use CaseWhat You Can DoWho Benefits
Custom DashboardsCreate visualizations in Excel, Tableau, or other BI toolsData analysts and leadership teams
Team CollaborationShare metrics with content teams and stakeholdersContent managers, product teams, and editors
Archival & ReportingMaintain historical records and generate periodic reportsCompliance teams and auditors
Trend AnalysisTrack changes in document performance over timeData scientists and ML engineers
Content PlanningIdentify gaps and prioritize content creation effortsDocumentation teams and content strategists

Setting Up Alerts

Get notified about important changes to stay proactive.

Critical Document Alerts

Monitor your most important documents for issues:
1

Tag critical documents

Mark your essential documents with a “critical” tag
2

Set retrieval drop alerts

Get notified if retrieval count drops suddenly
3

Monitor relevance scores

Alert when relevance score declines below threshold
4

Track negative feedback

Receive alerts on negative user feedback
Example Alert: “Pricing Policy Doc” retrieval count dropped 80% in the last week → investigate potential retrieval issue or content problem.

New Content Alerts

Track newly added documents to ensure they’re working properly:
1

Enable new document alerts

Get notified when a new document is synced
2

Monitor initial usage

Track its first 30 days of usage automatically
3

Verify retrieval

Confirm it’s being retrieved as expected

Best Practices

Follow these practices to get the most value from document usage analytics.

Review Top 20 Documents Monthly

Review TaskWhat To CheckWhy It Matters
Check AccuracyVerify your most-used documents contain correct informationErrors in frequently-used docs impact many users
Verify FreshnessConfirm documents are up to date with current informationOutdated content leads to wrong AI responses
Review FeedbackRead user feedback on traces that used these documentsUser feedback reveals quality issues
Update When NeededMake improvements based on your findingsContinuous improvement maintains quality

Investigate Zero-Retrieval Documents

Every month, check documents that are never retrieved:
Question To AskWhat To Look ForPossible Action
Are they truly irrelevant?Does this content apply to your AI’s use cases?Remove if genuinely not needed
Are embeddings broken?Is the content findable with semantic search?Regenerate embeddings if broken
Should they be removed?Is this taking up space without value?Delete to optimize database

Correlate with User Feedback

When users give thumbs down to AI responses:
1

View the trace

Open the trace with negative feedback
2

Check retrieved documents

See which documents were used in that response
3

Look for patterns

If the same document appears in many negative traces, you’ve found the problem
4

Update immediately

Fix that document as soon as possible

Track Seasonal Patterns

Some documents may have seasonal usage patterns that affect your update planning:
  • Tax documents — Retrieved heavily in April
  • Holiday policy docs — Spike in November-December
  • Budget documents — Active at fiscal year end
Plan your content updates and reviews around these patterns.
When creating new content:
TaskWhat To DoBenefit
Check Related DocsReview related documents for similar topicsUnderstand existing content landscape
Avoid DuplicationEnsure new content doesn’t duplicate existing docsPrevent confusion and redundancy
Fill GapsCreate content that bridges gaps between related documentsProvide comprehensive coverage

Next Steps