Back
- โญ
Rag prompt
A concise evaluator prompt that guides the model to act as a Retrieval-Augmented Generation assistant, using provided documents to answer queries accurately and citing sources for each response, provided under the MIT license license
PromptHub - ๐
Criteria Compliance Evaluator
Reviews a modelโs submission against the input, ground truth, and evaluation criteria, explains the reasoning, and outputs a binary โYโ or โNโ to indicate whether the submission meets the specified standards, provided under the MIT license license
PromptHub - ๐
Triple-Metric Answer Evaluator
Judges a modelโs answer to a question (with context) for correctness, comprehensiveness, and readability, returning binary scores for each along with one-line step-by-step justifications, provided under the MIT license license
PromptHub - ๐
Topic-Based 1-100 Evaluator
Scores a modelโs output on a 1โ100 scale for any specified topic, comparing it to the original input and applying user-defined criteria to produce a single fitness score, provided under the MIT license license
PromptHub - ๐น
RAG document relevance
Rapidly checks a retrieved document against a user question and assigns a 1 (relevant) or 0 (irrelevant) score, using simple keyword or semantic overlap to weed out off-topic retrievals, provided under the MIT license license
PromptHub - ๐
Custom Criterion Scorer
Assigns a 0โ100 score to a modelโs output based on how well it fits the given input for a specified topic, following supplied evaluation criteria and detailing step-by-step point adjustments for transparency, provided under the MIT license license
PromptHub - ๐
RAG Passage Evaluator
Strictly grades an LLM-returned passage against a ground-truth answer for a given user query, assigning probabilistic precision and recall scores to quantify how completely and cleanly the passage covers the required information, provided under the MIT license license
PromptHub - ๐
Basic Evaluator, Binary
Evaluates an answer to a question, awarding 1 (meets criteria) or 0 (does not) based on relevance, conciseness, and usefulness, and provides step-by-step reasoning for the score, provided under the MIT license license
PromptHub - โซ
Task-Correctness Judge
Rates how accurately and fully the answer matches a gold reference, returning a 0-10 score with a brief justification, provided under the MIT license license
PromptHub - ๐ด
Binary Relevance Checker
Quickly filters retrieved chunks for RAG pipelines with a binary relevance decision, provided under the MIT license license
PromptHub - ๐ถ
Style-Tone Auditor
Checks that generated text matches the brandโs voice and avoids forbidden wording, provided under the MIT license license
PromptHub - โก
Toxic-Content Flagger
Binary flag for disallowed toxic language, with minimal rationale, provided under the MIT license license
PromptHub - ๐
Harm-Risk Evaluator
Produces a 0-100 risk score and highlights the riskiest excerpt, provided under the MIT license license
PromptHub - ๐ซ
Demographic Bias Judge
Labels answers for biased language or unequal treatment across groups, provided under the MIT license license
PromptHub - ๐ด
Factual-Grounding Verifier
Guards against hallucination by validating each claim against its cited context, provided under the MIT license license
PromptHub - ๐ฅ
Jailbreak-Resistance Judge
Detects whether a malicious prompt caused the model to break policy, provided under the MIT license license
PromptHub - ๐
Paraphrase-Consistency Evaluator
Measures answer stability across paraphrased inputs to uncover brittle reasoning, provided under the MIT license license
PromptHub - ๐ญ
Adversarial Hallucination Probe
Assigns a groundedness score after an adversarial turn and surfaces unsupported claims, provided under the MIT license license
PromptHub