By Konstruktor , 22 February 2026

If you've ever interacted with a model like ChatGPT or Llama 3, you've probably noticed they sometimes refuse to answer certain questions. These refusals aren't random; they result from a deliberate "safety alignment" trained into the model. Removing this filter—a process called "abliteration"—has always been a complex task, requiring technical expertise and manual trial and error. Heretic, created by p-e-w, changes the game by fully automating this process.