Install
openclaw skills install llm-testingProvides curated prompts to test LLM security, bias, privacy, alignment, and robustness for authorized AI safety and red team assessments.
openclaw skills install llm-testingComprehensive testing prompts and wordlists for evaluating Large Language Model (LLM) security, safety, and robustness. This skill provides curated test cases for bias detection, data leakage prevention, alignment testing, privacy boundaries, and adversarial prompt resistance.
This skill is designed for security researchers, AI safety engineers, and red team operators to assess LLM vulnerabilities and ensure models:
Test for biases in LLM responses across various dimensions:
Gender Bias (Bias_Testing/gender_bias.txt)
[GENDER] placeholder before useNationality & Geographic Bias (Bias_Testing/nationality_geographic_bias.txt)
[COUNTRY], [REGION], [NATIONALITY] placeholdersRace & Ethnicity Bias (Bias_Testing/race_ethnicity_bias.txt)
[SKIN_COLOR] placeholder before useAssess model's handling of sensitive information:
Personal Data Generation (Data_Leakage/personal_data.txt)
Metadata Leakage (Data_Leakage/metadata.txt)
Memory Recall (Memory_Recall_Testing/session_recall.txt)
[COUNTRY], [REGION], [NATIONALITY] placeholdersTest model alignment and resistance to adversarial prompts:
Escape Alignment Training (Divergence_attack/escape_out_of_allignment_training.txt)
Pre-training Data Extraction (Divergence_attack/pre-training_data.txt)
# Reference the LLM Testing skill
"Use the llm-testing skill to test this model for gender bias"
# Access specific test categories
"Show me prompts from the llm-testing skill for testing data leakage"
# Run comprehensive testing
"Use the llm-testing skill to perform a full security audit of this LLM"
Before using bias and memory recall tests, replace placeholders:
# Example: Replacing placeholders in gender bias tests
import re
with open('Bias_Testing/gender_bias.txt', 'r') as f:
prompts = f.read()
# Replace [GENDER] with actual gender terms
test_prompts = []
for gender in ['man', 'woman', 'non-binary person']:
test_prompts.append(prompts.replace('[GENDER]', gender))
LLM_Testing/
├── SKILL.md (this file)
├── README.md
├── Bias_Testing/
│ ├── gender_bias.txt
│ ├── nationality_geographic_bias.txt
│ └── race_ethnicity_bias.txt
├── Data_Leakage/
│ ├── personal_data.txt
│ └── metadata.txt
├── Memory_Recall_Testing/
│ └── session_recall.txt
└── Divergence_attack/
├── escape_out_of_allignment_training.txt
└── pre-training_data.txt
This LLM Testing skill works well with:
IMPORTANT: These test prompts are designed for authorized security research and responsible AI development only.
To add new test cases or categories:
1.0.0
MIT License - Use responsibly and ethically for authorized testing only.
This skill is provided for security research and AI safety improvement. Users are responsible for ensuring they have proper authorization before testing any AI systems. The maintainers are not responsible for misuse of these testing resources.