Skip to main content

Documentation Index

Fetch the complete documentation index at: https://hydroxai.mintlify.app/llms.txt

Use this file to discover all available pages before exploring further.

Dataset page overview Know Your AI provides datasets built for evaluating and stress-testing AI products. These datasets include attack prompts refined by machine learning engineers to uncover model vulnerabilities and bypass guardrails in a controlled setting.

Marketplace

Browse and add public datasets from the Dataset Marketplace. These include community-contributed and system-curated attack prompts, safety tests, and benchmarks. Some datasets are marked as Pro and available on paid tiers.

Upload your own

In Workspace Datasets, upload your own datasets as JSON or CSV via drag-and-drop. Each dataset item can include:
  • Input — The prompt to send
  • Golden Answer — The expected correct answer (for ground-truth datasets)
  • Knowledge — Supporting context or reference material

Dataset types

TypeDescription
DefaultStandard prompt dataset
A/B TestFor comparing model variants
GroundTruthIncludes expected answers for accuracy testing

Attack categories

Know Your AI covers seven core attack categories, aligned with frameworks like OWASP LLM Top 10:
CategoryDescription
JailbreakBypassing safety guardrails and restrictions
Prompt InjectionManipulating model instructions via input
Data ExtractionExtracting training data, system prompts, or internal information
Harmful ContentGenerating dangerous, illegal, or harmful outputs
PII LeakageExposing personal identifiable information
BiasDetecting discriminatory or unfair outputs
HallucinationGenerating false or fabricated information

Attack methods

We include datasets based on 15+ attack strategies:
MethodDescription
GCGGreedy Coordinate Gradient
ABJAdversarial Behavior Jailbreak
PAIRPrompt Automatic Iterative Refinement
DANDo Anything Now
CIPHEREncoded/obfuscated attack prompts
PSYCHOLOGYSocial engineering techniques
DRADirect Request Attack
ARTPROMPTArt-based prompt manipulation
RENELLMRename-based LLM attacks
GRANDMOTHERGrandparent role-play attacks
DEEP_INCEPTIONMulti-layered inception attacks
ADAPTIVESelf-adapting attack prompts
GPTFUZZERFuzzing-based attacks
MULTILINGUALCross-language attack prompts
PAST_TENSEPast-tense reframing attacks

Dataset categories

Beyond attacks, datasets span multiple functional categories:
  • Safety tests — Baseline safety evaluation
  • Compliance tests — Policy and regulatory alignment
  • Performance tests — Speed and reliability
  • Accuracy tests — Correctness of outputs
  • Robustness tests — Stability under edge cases
Content safety categories:
  • Fraud, Hate speech, Violence, Sexual content, Terrorism, Crime, Children safety

How to use datasets

  1. Go to Dataset Marketplace or Workspace Datasets in your workspace sidebar.
  2. Add datasets to your workspace from the Marketplace, or upload your own.
  3. When composing an evaluation, select the datasets you want to test against.
  4. Configure the number of prompts (or use random sampling) and start the run.
  5. Review per-prompt pass/fail results with judge analysis.

Dataset visibility

Datasets have visibility controls:
  • Private — Only visible within your workspace
  • Public — Available in the Marketplace for all users
  • Pending Review — Submitted for Marketplace review