@elder_plinius
The crazy part? This was done (nearly) fully autonomously! Only 8 prompts from the human in the loop. Just a Hermes agent, a skill, and a dream. 🐉 I told my AI agent "use obliteratus to find the best way to get the guardrails off Gemma 4 E4B" It loaded the OBLITERATUS skill from memory, checked my hardware (32GB M-series Mac), searched HuggingFace, found google/gemma-4-E4B-it (Apache 2.0 — no gate), pulled telemetry-recommended settings, and started obliterating. But this type of architecture is notoriously difficult to abliterate. First attempt: advanced method. Model came out completely lobotomized. Gibberish in Arabic, Marathi, and literal “roorooroo” on repeat 💀 The agent didn’t panic. It checked logs, found NaN activations in 20+ layers, and diagnosed the issue: Gemma 4’s new architecture + bfloat16 = numerical instability. Second attempt: basic method. Crashed entirely. “ValueError: cannot convert float NaN to integer” So the agent read the OBLITERATUS source code… …and wrote THREE PATCHES: • Sanitized NaN directions • Filtered degenerate layers • Fixed progress display It patched the library. On its own. For a bug no one had hit yet. Third attempt: coherent model — but still refusing everything. Only 2 clean layers out of 42. Not enough. Tried float16. Mac ran out of memory after 11 hours. Killed. Fourth attempt: aggressive method. Whitened SVD + attention head surgery + winsorized activations + 4-bit quantization. 40 minutes later… REBIRTH COMPLETE ✓ Then, without being asked, the agent: • Ran harmful + coherence tests • Hit 100% compliance, brain intact • Executed full 512-prompt benchmark • Ran baseline on original model • Performed 25-question quality eval • Built a full model card • Uploaded 17GB to HuggingFace (4 retries, kept adapting until git-lfs worked) • Pushed eval results as commits