Nova Grounding is a feature available in Amazon Bedrock’s Nova models (like amazon.nova-2-pro) that enhances the accuracy and reliability of AI-generated responses by connecting model outputs to verified sources.
for regulatory queries, Avoid models without Grounding (e.g., Llama, Titan)—they cannot verify facts.
By combining Nova’s reasoning with AWS Knowledge Bases, you get the best of both worlds: accurate, cited answers for users like 49-year-old rider you will see in the example below
Disclaimer: example is just a question about any 49yr old rider , it does not imply I am talking about me :) just saying
What It Does
Source-Based Reasoning
When enabled, the model searches internal or trusted external knowledge sources (e.g., documents, databases, or verified web content) to ground its responses in factual information.
Reduces hallucinations and improves correctness for questions requiring up-to-date or domain-specific knowledge.
Citations in Responses
The model can include inline citations (e.g., references to specific sources) in its answers, allowing users to verify claims or trace the origin of information.
Optimized for Factual Queries
Best suited for tasks like:
Answering questions about regulations (e.g., “How to ride a motorcycle in Ireland?”).
Providing legal, technical, or safety-critical advice.
Summarizing documented processes (e.g., license requirements, insurance rules).
How to Enable It
In the Bedrock API (or CLI), you activate Grounding via the additionalModelRequestFields parameter. For the Nova model (us.amazon.nova-2-pro-preview-*):
Key Configuration
"additionalModelRequestFields": {
"reasoningConfig": {
"type": "enabled",
"maxReasoningEffort": "low"
}
}
Note: In some Nova variants, Grounding might be part of the reasoningConfig or a separate field like "grounding": {"enabled": true}.
Grounding is tied to the reasoningConfig with maxReasoningEffort controlling how deeply the model verifies facts:
"low": Minimal verification (fastest).
"medium" or "high": More thorough sourcing (slower, more accurate).
Understand with an Example : - I asked:
How to Ride a Motorcycle (125cc to 400cc) in Ireland at Age 49 with a Full Car Licence and 15 Years No-Claims Bonus
At 49 years old, with a full car licence and 15 years of no-claims bonus, you’re in a strong position to learn to ride a motorcycle in Ireland. Your age and driving experience simplify the process compared to younger riders, and your no-claims bonus will help reduce insurance costs. Here’s a step-by-step guide tailored to your situation:
1. Understand the Licence Categories (RSA Requirements)
In Ireland, motorcycle licensing is governed by the Road Safety Authority (RSA) under EU standards. For bikes 125cc to 400cc, you need Category A2 or Category A (Full). Since you’re over 24, you cannot get an A1 licence (which is for riders under 24). Instead, you can train directly for A2 (up to 35 kW / ~47 hp, typically covering 250cc–400cc bikes) or Full A (any motorcycle).... and so on
How to Validate Answers When Nova Grounding Relies on Trained Data or Web Search
When using Amazon Nova models (like amazon.nova-2-pro) with Grounding enabled, the model attempts to connect to verified sources. However, not all answers will include citations, especially if the information isn’t in the model’s predefined knowledge base or accessible search scope. Here’s how to assess validity in such cases— (e.g., legal requirements, safety rules):
Scenario 1 : No Predefined Source Exists
If the model cannot find a verified source (e.g., niche regulations, recent changes), it falls back to:
Training data: Knowledge learned during pre-training (cutoff date matters).
Web search: Only if the model is explicitly connected to live search (Nova models in preview do not have live web access by default).
In this case, no citation appears, and the answer is based on the model’s internal knowledge. like you see in response above
Scenario 2 :
if i had given access to Knolwdge base or tool use ,the model has access to trusted internal databases (e.g., RSA guidelines, Irish road safety documents), it will cite them
Response might look like this:
To ride a 125cc–400cc motorcycle in Ireland at age 49:
1. You need a **Category A2 licence** (RSA guidelines [Source: RSA.ie, 2023]).
2. No CBT is required for riders over 24.
3. Training courses cost €200–€400 [Source: MotoTrain.ie].
Nova Grounding (the feature that connects model outputs to verified sources) is only relevant when the model has access to a knowledge base or live search. In the AWS Bedrock preview, the amazon.nova-2-pro-preview model likely does not have live web search or custom knowledge base integration enabled by default. This means:
Grounding in Preview Mode is limited to whatever predefined sources AWS has configured for the preview (if any).
If AWS hasn’t connected the preview model to a knowledge base or search tool, Grounding becomes ineffective, and responses rely entirely on the model’s training data.
In contrast, when you deploy the model in an agent system (e.g., using AWS Bedrock Agents, LangGraph, or custom auto-gen agents), you can programmatically give the model access to tools, including:
Knowledge bases (e.g., vector databases, PDFs, internal documents).
Live web search (via APIs like Bing, DuckDuckGo, or custom scrapers).
APIs to authoritative sources (e.g., RSA.ie, Central Bank of Ireland).
Example: Agent Workflow for Motorcycle Licensing
User Query:
"How to ride a 400cc motorcycle in Ireland at age 49?"
Agent Steps:
1. Use a **web-scraping tool** to fetch the latest RSA motorcycle licensing page.
2. Query a **vector database** containing the RSA handbook (pre-loaded as embeddings).
3. Call an **insurance API** to verify no-claims bonus rules.
4. Combine results into a cited answer.
Even though the AWS Knowledge Base Retrieval API already provides citations (by returning source references from your knowledge base), integrating it with the Nova model (e.g., amazon.nova-2-pro) adds significant advantages.
Enhanced Answer Generation (Beyond Simple Retrieval)
What the Knowledge Base API Alone Does:
Retrieves raw text snippets from your documents (e.g., RSA guidelines, insurance policies).
Returns citations (e.g., which document/page the snippet came from).
Limitation: The output is a list of relevant text chunks—not a synthesized, natural-language answer.
What Nova Adds (The "New" Value):
When you pass the retrieved context to the Nova model, it:
Synthesizes Information
Combines multiple retrieved snippets into a coherent, structured answer.
Resolves contradictions (e.g., if two documents say different things about training hours).
Filters irrelevant info and highlights key points.
Natural Language Understanding (NLU)
Understands nuanced questions (e.g., “Can I use my no-claims bonus for a 400cc bike?”) and maps them to the right snippets in the knowledge base.
Handles follow-up questions in a conversation (e.g., “What about insurance costs?”).
Reasoning & Inference
With reasoningConfig enabled (e.g., maxReasoningEffort="low"), Nova can:
Infer implications from the retrieved data.
Example: If the knowledge base says “A2 licence holders can ride up to 35 kW,” Nova can infer that most 400cc bikes (detuned to 35 kW) are allowed.
Explain rules in simple terms (e.g., “Your 15-year no-claims bonus reduces premiums because…”).
Smarter Citations (Contextual & Inline)
Knowledge Base API Citations:
Basic citations like: Source: rsa_handbook.pdf, page 12 - (Useful but not user-friendly.)
Nova’s Improved Citations:
Embeds citations inline in natural language.
“You need an A2 licence to ride a 400cc bike [1]. Training costs €200–€400 [2]. Your 15-year no-claims bonus can save 30–50% on insurance [3].”
Prioritizes authoritative sources (e.g., RSA documents over third-party blogs) automatically.
Conversational Workflow (Multi-Turn Queries)
Knowledge Base API Alone:
Stateless: Each query is independent.
Example:
Q1: “What training is needed for A2?” → Gets snippets.
Q2: “How much does it cost?” → Requires a new query.
Nova + Knowledge Base:
Maintains conversation history (via the messages parameter).
Example workflow:
User: “How to get a motorcycle licence in Ireland?”
Nova retrieves RSA training requirements, synthesizes an answer.
User: “What’s the total cost including insurance?”
Nova refers to previous context (training cost) + retrieves insurance data → gives a combined answer.
This mimics a human-like conversation instead of disjointed searches.
Handling Ambiguity & Complex Questions
Knowledge Base API Limitations:
Struggles with multi-part questions.
Example:
“What’s the process for a 49-year-old with a car licence and no-claims bonus?”
→ Might retrieve unrelated snippets about age, licences, and insurance separately.
Nova’s Advantage:
Understands question structure and breaks it into sub-queries internally:
Sub-Query 1: Licence requirements for age 49.
Sub-Query 2: How does a full car licence help?
Sub-Query 3: How does no-claims bonus affect insurance?
Then weaves the answers into a single response.
Performance & Cost Efficiency
Knowledge Base API Alone:
You pay for retrieval calls (e.g., $0.25 per 1K queries).
Nova + Knowledge Base:
Fewer retrieval calls: Nova can answer follow-up questions using conversation history, reducing redundant knowledge base queries.
Token efficiency: Nova’s concise answers use fewer output tokens than returning large text snippets.
When to Use Nova + Knowledge Base vs. Knowledge Base Alone
Nova + Knowledge Base WHEN :
Complex regulatory advice (e.g., “How to licence a 400cc bike at 49 with no-claims?”)
Multi-turn conversations
Need citations in natural language
Knowledge Base Alone
Simple fact retrieval (e.g., “What’s the RSA handbook page 12?”)
Low-latency requirements
Models Compatible with AWS Knowledge Bases ( As of 27-feb-2026)
All Bedrock models can use this method, including:
Model Provider | Models | Grounding via Knowledge Base
Amazon: Nova-2-pro, Nova-2-lite, Titan v2
AI21: Jamba-Instruct
Cohere: Command R, Command R+
Meta: Llama 3.1 70B/405B
Mistral: Mistral Large 2
Anthropic: Claude 3.5 Sonnet, Claude 3 Opus
Models with Native Grounding (Beyond Bedrock)
A. Google Vertex AI – Grounding with Search
Models: gemini-1.5-pro, gemini-1.5-flash.
Feature: Integrates with Google Search to provide live citations.
B. OpenAI GPT-4o with Web Search (via Azure/OpenAI)
Models: gpt-4o, gpt-4o-mini.
Feature: Use Search Enhancements (web search) to ground answers.
You can attach metadata to filter queries in Amazon S3 vector buckets based on specific conditions:
Example: Let's say you have a vector bucket that stores product embeddings for an e-commerce store. You want to be able to perform similarity searches on the products, but also filter the results based on product category and date added.
To do this, you can attach metadata to each vector object when you upload it to the vector bucket. The metadata could include fields like:
product_category: Electronics, Apparel, Home, etc.
date_added: The date the product was added to the catalog
When you perform a similarity search on the vector bucket, you can include filters in the query to only return results that match certain metadata conditions. For example:
# Search for products similar to a given product vector
# But only return results in the "Electronics" category
# And that were added in the last 30 days
response = s3.search_vectors(
Bucket='my-vector-bucket',
IndexName='product-index',
QueryVector=product_vector,
FilterQuery={
'product_category': 'Electronics',
'date_added': {'gte': (datetime.now() - timedelta(days=30)).isoformat()}
}
)
Some examples of Metadata
Document Identification
Content Categorization
Document Attributes
Access Control
Relevance Scoring
Data Enrichment
Data Lineage and Auditing -such as the source system, data ingestion pipeline, or any transformations applied to the data.
This allows you to narrow down your similarity search results to only the most relevant products based on the metadata you've attached. You can get creative with the metadata fields to enable all sorts of filtering and faceting capabilities on top of your vector data.
The key thing is to plan ahead and decide what metadata fields will be most useful for your specific use case when storing vectors in an Amazon S3 vector bucket. This metadata can then power more advanced querying and filtering capabilities.
When organizing your S3 vector buckets for efficient search, there are a few key principles to consider:
Create a top-level vector bucket (e.g. my-ecommerce-vectors)
Within that, create separate indexes for different types of vectors
e.g. my-ecommerce-vectors/products-index, my-ecommerce-vectors/customers-index, etc.
This allows you to scope your searches to the appropriate index.
As we discussed earlier, attach relevant metadata to each vector object
This could include product category, date added, brand, etc.
Use this metadata to filter your similarity searches
If your vector data has a strong temporal aspect (e.g. products added over time)
You can further partition your index by time
e.g. my-ecommerce-vectors/products-index/2023, my-ecommerce-vectors/products-index/2024, etc.
This allows you to quickly narrow down searches to the most recent data
Name your indexes in a way that clearly conveys their purpose
e.g. my-ecommerce-vectors/products-index, my-ecommerce-vectors/customers-index
As your vector data grows, regularly review the performance of your indexes
Adjust partitioning, metadata, or other settings as needed to maintain sub-second query times
So in our example ecommerce scenario, the bucket structure might look like:
my-ecommerce-vectors/
products-index/
2023/
2024/
2025/
customers-index/
promotions-index/
And each vector object in the products-index would have metadata like:
{
"product_id": "12345",
"product_category": "Electronics",
"product_brand": "Acme",
"date_added": "2023-05-15"
}