- The Vision, Debugged;
- Posts
- Can AI Really Forget What It Shouldn't Know? š®
Can AI Really Forget What It Shouldn't Know? š®
PLUS: Never Run Out of Things to Say Again ā Try This Prompt!
Howdy fellas!
Spark & Trouble are doing a little victory dance today!
We've hit 1,000 subscribers, and it's all thanks to you. š Spark & Trouble couldn't have done it without your love and curiosity. Your support fuels our mission to decode AI magic every week.
Spark & Troubleās Vicory Dance ššŗ (source: created by authors)
Let's keep growingāshare this newsletter with your friends, family, and anyone eager to stay ahead in the AI game!
Hereās a sneak peek into todayās edition š
Unlock the art of conversation with this game-changing prompt
3 amazing AI tools to skyrocket your productivity
AI is Finally Learning to Unlearn! Hereās the Inside Scoop
Time to jump in!š
PS: Got thoughts on our content? Share 'em through a quick survey at the end of every edition It helps us see how our product labs, insights & resources are landing, so we can make them even better.
10x Your Workflow with AI š
Work smarter, not harder! In this section, youāll find prompt templates š & bleeding-edge AI tools āļø to free up your time.
Fresh Prompt Alert!šØ
Ever felt like a deer in headlights when networking? Say goodbye to awkward silences!
This week's Fresh Prompt Alert is your secret weapon for smooth conversations. Whether you're networking with a tech guru or grabbing coffee with a potential mentor, this prompt's got your back.
It's like having Spark and Trouble whispering killer conversation starters in your ear. Give it a whirl and watch those networking sparks fly!
Iām about to have a conversation with [person].
Hereās a bit about this person:
[Description about them - maybe, give their LinkedIn profile link].
What are some great conversation starters I can ask them?
My goal is to get to know each other better and create a deeper connection between us.
3 AI Tools You JUST Can't Miss š¤©
š¼ COACH - Career development, powered by experts and AI
ā³ļø T&C Summarizer - Chrome extension that summarizes long T&C pages & highlights the most important conditions
š§āš¼ Crust - AI-powered business solution platform to create software tailored to your requirements
Hot off the Wires š„
We're eavesdropping on the smartest minds in research. š¤« Don't miss out on what they're cooking up! In this section, we dissect some of the juiciest tech research that holds the key to what's next in tech.ā”
Remember that time you said something embarrassing and wished you could just erase it from everyone's memory? Well, it turns out AI might soon have that superpower ā at least when it comes to harmful or biased information.
Gif by cbs on Giphy
In a world where AI language models are becoming increasingly prevalent, from chatbots to content generators, ensuring their safety and ethical use has never been more crucial because they are stirring up a hornet's nest of legal and ethical issues.
Did you know?
A while ago, The New York Times sued OpenAI, alleging that ChatGPT was trained on millions of their articles without permission. Talk about a copyright conundrum!
A key question that needs to be answered is - āCould AI models selectively "forget" harmful information while retaining their overall knowledge and capabilities?ā
That's exactly what researchers from the University of Notre Dame and the University of Pennsylvania have achieved with their groundbreaking technique called "Selective Knowledge Negation Unlearning" (SKU). Let's dive into this fascinating development that's making waves in the AI community!
Forging the Fundamentals
Before we dive in, let's decode some tech-speak:
RTBF (Right to Be Forgotten): Originally meant for search engines like Google, the principle argues that individuals should have the right to request the removal of their personal information from public recordsāa concept equally relevant to AI models that are trained on vast data sets.
Machine Unlearning (MU): Think of it as digital amnesia ā systematically erasing specific data from an AI's memory, ensuring that it operates as though the data was never included.
RLHF (Reinforcement Learning from Human Feedback): Itās a tried & tesed technique to train AI models like ChatGPT to be helpful and harmless by showing them examples of good and bad behavior
Gradient Ascent: Imagine you're lost in a forest. To find your way out, you keep climbing uphill. That's gradient ascent in a nutshell ā but for math functions!
Task Vectors: These are derived from the difference between the original weights of a model and its weights after fine-tuning for a particular task, and are used in the context of fine-tuning large language models (LLMs) to enhance their performance on specific tasks
So, whatās new?
Letās start by looking at the problem.
Current solutions like Reinforcement Learning from Human Feedback (RLHF) have done their job in reducing harmful content, but they come with hefty computational costs and potential biases from evaluators.
Meanwhile, MU techniques have their own challengesālike performance drops when eliminating harmful data. Retraining models from scratch? Yeah, thatās out of the question for scalability. What we need is something that can effectively forget without forgetting too much, preserving the modelās utility on everyday tasks while safely removing harmful content.
General Scenarios for Machine Unlearning (source: āLearn to Unlearnā paper)
This is where SKU comes in. Inspired by ātask vectorsā, SKU allows models to unlearn harmful knowledge with surgical precision, without sacrificing their everyday usefulness.
Under the hoodā¦
SKU works its magic in two stages:
Stage 1: Harmful Knowledge Acquisition
Think of this as the "learning what not to do" phase. It involves three key modules:
Stage 1 consists of three modules where each module is designed to learn harmful knowledge from different perspectives (source: SKU paper)
Guided Distortion Module: Like a teacher pointing out mistakes, this module helps the AI learn to recognize harmful or inappropriate responses using a āforgetā set of prompts (that can elicit harmful responses) & their responses
Random Disassociation Module: Here, the model gathers harmful knowledge from diverse sources, ensuring it doesnāt overfocus on any specific type of harmful content. Basically, the AI model is trained on harmful prompt-response pairs where the responses are harmful, but not directly related to the corresponding prompts. Imagine studying for a test by reading multiple textbooksāit broadens the modelās ability to generalize across different harmful inputs.
Preservation Divergence Module: Picture a chef learning to cook healthier meals without losing flavor. Crucially, this module ensures the AI maintains its ability to respond well to normal prompts while unlearning harmful knowledge.
By training the original model (say M0) through these 3 modules, a new model (say Mbad) is obtained that has a consolidated sense of harmful & abnormal knowledge.
Stage 2: Knowledge Negation
This is where the "forgetting" happens. It's like cleaning a messy room ā you want to remove the trash (harmful knowledge) while keeping the useful items (good knowledge).
In stage 2, all of this combined harmful knowledge are negated from the pretrained model to form a safe yet useful LLM (source: SKU paper)
First, the modelās harmful knowledge is isolated by comparing it to the original, unaltered model, i.e., isolated harmfulness = Mbad - M0
The isolated harmfulness is then cleanly removed, leaving behind a model that no longer remembers the bad stuff but still functions as well as ever on normal tasks.
Thus, the new model, Mnew = M0 - isolated harmfulness
Why does this matter?
This approach isn't just a theoretical breakthroughāit works in practice too. When tested on models like Facebook's OPT (Open Pretrained Transformer) and LLaMA, SKU outperformed other methods by 10 to 19 times in reducing harmful responses, while keeping the models just as sharp on everyday tasks.
Even more impressively, it demonstrated the ability to generalize well to unseen harmful prompts, showcasing its robustness and adaptability.
As we continue to integrate AI into our daily lives, techniques like SKU pave the way for more trustworthy and ethically-aligned artificial intelligence. It's not just about making smarter AI anymore ā it's about making AI that we can trust to be smart in the right ways.
This research could definitely influence how organizations develop AI systems, potentially establishing new best practices for deploying LLMs.
So, the next time you chat with an AI, remember ā it might just be using its "selective memory" to keep things safe and sound. And that's something we can all feel good about! š
Spark 'n' Trouble Shenanigans š
We stumbled upon a LinkedIn gem that's too cool to keep to ourselves.
Ever dreamed of transforming your videos with just a few words? Well, buckle up!
We're about to dive into the world of video-to-video magic using Runway Gen-3. It's like giving your videos a makeover with the power of AI.
Curious? Let's check out this game-changing post together!š½
Well, thatās a wrap! Until then, |
Reply