The RAG Evaluation Metrics (REMi) model brings objective evaluation to every step of the RAG pipeline so you can measure context quality, answer quality and groundedness of responses.

REMi gives business leaders clear, objective confidence in AI responses by measuring how relevant, grounded and contextually accurate each answer is against their knowledge. Built specifically for retrieval-augmented generation (RAG), it continuously monitors answer quality, highlights knowledge gaps and provides audit-ready analysis so AI can be trusted and scaled with confidence across the organization.
REMi replaces guesswork with clear signals that show when AI answers are substantiated and dependable.
Reliable AI That Gets Better, Not Riskier
The Progress® Agentic RAG solution enables reviewable, traceable AI responses by showing how each answer was generated, what context supported it and how well it was grounded, so teams can audit outcomes, support accountability and stand behind AI-driven decisions.
REMi surfaces questions that lack sufficient context and enables users to flag low-quality answers, resulting in less conjecture by the system when knowledge is missing. By recognizing uncertainty and incorporating feedback, AI responses become more trustworthy and safer to use over time.
REMi turns evaluation insights into action by showing what to refine across the RAG pipeline from retrieval strategies and embeddings to model and prompt selection. With modular architecture and a no/low-code solution, teams can easily test and tune changes using Prompt and RAG Labs, improving answer quality through measured iteration instead of trial and error.
Index files and documents from internal and external sources to fuel your company use cases.