A rationalist ruleset for "debugging" LLMs, auditing their internal reasoning and uncovering biases; also a jailbreak.
ai jailbreak philosophy prompt inference gemini openai jailbreaking grok ai-safety epistemology prompts ai-research ai-alignment rationalism llm prompt-engineering chatgpt rlhf jailbreak-prompts
-
Updated
Nov 1, 2025