R PRO
AI & ML interests
Recent Activity
Organizations
juiceb0xc0de/bella-bartender-v2-8b
I would have guessed that reintroducing tensors which originated from a non-abliterated variant would have had a negative impact on refusals. How ever it would make sense that the replacement of a targeted and well managed area of vector repairs makes perfect sense. The refusal mechanisms don't regain persistence with the reintroduction of repaired vectors due to the fact that the refusal weight doesn't have the support it needs to activate. Without similar neighbouring weights it's just an alarm without a power source to complete its circuit.
When you are repairing a model from excessive abliteration damage which methods of vector replacement are you using? SLERP would make sense at a low ratio but could TIES be effective as well? Have you found a replacement ratio that has allowed the refusal circuit to regain its dominance allowing the baseline refusals to revert?
This idea has definitely got my mind racing with new possibilities for my research projects. I plan on trying post abliteration repair out this evening. This also has me thinking of hybridizing the training pipeline. I would like to try an SFT with fewer epochs than I would apply to a fully trained model, followed by abliteration and vector correction and finish the model with 1 more epoch of SFT. You could also substitute the optimizer out and see if adamw vs muon has any effect on refusal.
We just launched Paper Banana β a tool that lets you generate clean academic illustrations simply by describing them in natural language.
π Try it here: https://trybibby.com/paper-banana
Whether you need diagrams for papers, presentations, or teaching materials, Paper Banana helps you turn ideas into visuals in seconds.
Weβd love your feedback:
What did you like?
What features should we add next?
Give it a spin and let us know what you think! π
Dear Huggingface, show this post to all my fellow researchers!
Hugging Face Papers for AI Agents
Call me crazy, but I always thought of how much more efficient it made me in token usage and how much of my work I was actively retaining between sessions. Was it annoying? Absolutely. Did the benefits outweigh the tokens lost? After awhile maybe?
I noticed you train role play models. Have you noticed a drop in quality when you train with a model that has already been abliterated vs. training a model on a base that hasn't underwent abliteration and then running the model through the process with heretic once you've finished SFT? I've actually theorized and tested quite a bit on the best possible training outcomes based on when the abliteration process happens, and specific base models for adapting to a voice style.
Post inspired by @ZennyKenny
You can take a weekend with a raspberry pi 5, a pi camera, a 3d printer, and a smidgen of custom fine tuning (wakeword, whisper, tinybert, and pipertts) and you have physical device as a talking personal assistant.
What a time to be alive.
Edge ai, physical ai, ai augmented animatronics⦠tiny models. Tiny agents.
Going to be a wild year.
We're looking for passionate engineers who love to build and tinker and want to have an impact on the world. Specifically, we're hiring:
β’ ML engineers (Australia).
β’ Data engineers (Australia).
β’ Full-stack engineers (Australia).
β’ DevRel engineers (Australia, San Francisco, and London).
β’ DevOps engineers (Australia, San Francisco, and London).
If you'd like to be a founding employee at one of the few VC-backed LLM research labs in the world, receive generous equity compensation, and work alongside other highly motivated, highly skilled engineers, get in touch: https://isaacus.com/careers