Friday, March 13

News Feed

Category Added in a WPeMatico Campaign

Comparing Precision Knowledge Editing with existing machine unlearning methods
News Feed, Reddit

Comparing Precision Knowledge Editing with existing machine unlearning methods

I've been working on a project called PKE (Precision Knowledge Editing), an open-source method to improve the safety of LLMs by reducing toxic content generation without impacting their general performance. It works by identifying "toxic hotspots" in the model using neuron weight tracking and activation pathway tracing and modifying them through a custom loss function. There's lots of current Machine unlearning techniques that can make LLMs safer right now like: Exact Unlearning: This method involves retraining the model from scratch after removing the undesired data. While it ensures complete removal of the data's influence, it is computationally expensive and time-consuming, especially for large models. Approximate Unlearning: Fine-Tuning: adjusting the model using the remaining data t...
The AI Report