Can AI Really Forget What It Shouldn't Know? šŸ˜®

PLUS: Never Run Out of Things to Say Again ā€“ Try This Prompt!

Howdy fellas!

Spark & Trouble are doing a little victory dance today!

We've hit 1,000 subscribers, and it's all thanks to you. šŸŽ‰ Spark & Trouble couldn't have done it without your love and curiosity. Your support fuels our mission to decode AI magic every week.

Spark & Troubleā€™s Vicory Dance šŸ’ƒšŸ•ŗ (source: created by authors)

Let's keep growingā€”share this newsletter with your friends, family, and anyone eager to stay ahead in the AI game!

Hereā€™s a sneak peek into todayā€™s edition šŸ‘€

  • Unlock the art of conversation with this game-changing prompt

  • 3 amazing AI tools to skyrocket your productivity

  • AI is Finally Learning to Unlearn! Hereā€™s the Inside Scoop

Time to jump in!šŸ˜„

PS: Got thoughts on our content? Share 'em through a quick survey at the end of every edition It helps us see how our product labs, insights & resources are landing, so we can make them even better.

10x Your Workflow with AI šŸ“ˆ

Work smarter, not harder! In this section, youā€™ll find prompt templates šŸ“œ & bleeding-edge AI tools āš™ļø to free up your time.

Fresh Prompt Alert!šŸšØ

Ever felt like a deer in headlights when networking? Say goodbye to awkward silences!

This week's Fresh Prompt Alert is your secret weapon for smooth conversations. Whether you're networking with a tech guru or grabbing coffee with a potential mentor, this prompt's got your back.

It's like having Spark and Trouble whispering killer conversation starters in your ear. Give it a whirl and watch those networking sparks fly!

Iā€™m about to have a conversation with [person].

Hereā€™s a bit about this person:
[Description about them - maybe, give their LinkedIn profile link].

What are some great conversation starters I can ask them?

My goal is to get to know each other better and create a deeper connection between us.

* Replace the content in brackets with your details - this prompt works best with Copilot / Gemini, which have access to the web

3 AI Tools You JUST Can't Miss šŸ¤©

  • šŸ’¼ COACH - Career development, powered by experts and AI

  • āœ³ļø T&C Summarizer - Chrome extension that summarizes long T&C pages & highlights the most important conditions

  • šŸ§‘ā€šŸ’¼ Crust - AI-powered business solution platform to create software tailored to your requirements

Hot off the Wires šŸ”„

We're eavesdropping on the smartest minds in research. šŸ¤« Don't miss out on what they're cooking up! In this section, we dissect some of the juiciest tech research that holds the key to what's next in tech.āš”

Remember that time you said something embarrassing and wished you could just erase it from everyone's memory? Well, it turns out AI might soon have that superpower ā€“ at least when it comes to harmful or biased information.

Mark Harmon Gibbs GIF by CBS

Gif by cbs on Giphy

In a world where AI language models are becoming increasingly prevalent, from chatbots to content generators, ensuring their safety and ethical use has never been more crucial because they are stirring up a hornet's nest of legal and ethical issues.

Did you know?
A while ago, The New York Times sued OpenAI, alleging that ChatGPT was trained on millions of their articles without permission. Talk about a copyright conundrum!

A key question that needs to be answered is - ā€œCould AI models selectively "forget" harmful information while retaining their overall knowledge and capabilities?ā€

That's exactly what researchers from the University of Notre Dame and the University of Pennsylvania have achieved with their groundbreaking technique called "Selective Knowledge Negation Unlearning" (SKU). Let's dive into this fascinating development that's making waves in the AI community!

Forging the Fundamentals

Before we dive in, let's decode some tech-speak:

RTBF (Right to Be Forgotten): Originally meant for search engines like Google, the principle argues that individuals should have the right to request the removal of their personal information from public recordsā€”a concept equally relevant to AI models that are trained on vast data sets.

Machine Unlearning (MU): Think of it as digital amnesia ā€“ systematically erasing specific data from an AI's memory, ensuring that it operates as though the data was never included.

RLHF (Reinforcement Learning from Human Feedback): Itā€™s a tried & tesed technique to train AI models like ChatGPT to be helpful and harmless by showing them examples of good and bad behavior

Gradient Ascent: Imagine you're lost in a forest. To find your way out, you keep climbing uphill. That's gradient ascent in a nutshell ā€“ but for math functions!

Task Vectors: These are derived from the difference between the original weights of a model and its weights after fine-tuning for a particular task, and are used in the context of fine-tuning large language models (LLMs) to enhance their performance on specific tasks

So, whatā€™s new?

Letā€™s start by looking at the problem.

Current solutions like Reinforcement Learning from Human Feedback (RLHF) have done their job in reducing harmful content, but they come with hefty computational costs and potential biases from evaluators.

Meanwhile, MU techniques have their own challengesā€”like performance drops when eliminating harmful data. Retraining models from scratch? Yeah, thatā€™s out of the question for scalability. What we need is something that can effectively forget without forgetting too much, preserving the modelā€™s utility on everyday tasks while safely removing harmful content.

General Scenarios for Machine Unlearning (source: ā€œLearn to Unlearnā€ paper)

This is where SKU comes in. Inspired by ā€œtask vectorsā€, SKU allows models to unlearn harmful knowledge with surgical precision, without sacrificing their everyday usefulness.

Under the hoodā€¦

SKU works its magic in two stages:

Stage 1: Harmful Knowledge Acquisition

Think of this as the "learning what not to do" phase. It involves three key modules:

Stage 1 consists of three modules where each module is designed to learn harmful knowledge from different perspectives (source: SKU paper)

  • Guided Distortion Module: Like a teacher pointing out mistakes, this module helps the AI learn to recognize harmful or inappropriate responses using a ā€œforgetā€ set of prompts (that can elicit harmful responses) & their responses

  • Random Disassociation Module: Here, the model gathers harmful knowledge from diverse sources, ensuring it doesnā€™t overfocus on any specific type of harmful content. Basically, the AI model is trained on harmful prompt-response pairs where the responses are harmful, but not directly related to the corresponding prompts. Imagine studying for a test by reading multiple textbooksā€”it broadens the modelā€™s ability to generalize across different harmful inputs.

  • Preservation Divergence Module: Picture a chef learning to cook healthier meals without losing flavor. Crucially, this module ensures the AI maintains its ability to respond well to normal prompts while unlearning harmful knowledge.

By training the original model (say M0) through these 3 modules, a new model (say Mbad) is obtained that has a consolidated sense of harmful & abnormal knowledge.

Stage 2: Knowledge Negation

This is where the "forgetting" happens. It's like cleaning a messy room ā€“ you want to remove the trash (harmful knowledge) while keeping the useful items (good knowledge).

In stage 2, all of this combined harmful knowledge are negated from the pretrained model to form a safe yet useful LLM (source: SKU paper)

First, the modelā€™s harmful knowledge is isolated by comparing it to the original, unaltered model, i.e., isolated harmfulness = Mbad - M0

The isolated harmfulness is then cleanly removed, leaving behind a model that no longer remembers the bad stuff but still functions as well as ever on normal tasks.

Thus, the new model, Mnew = M0 - isolated harmfulness

Why does this matter?

This approach isn't just a theoretical breakthroughā€”it works in practice too. When tested on models like Facebook's OPT (Open Pretrained Transformer) and LLaMA, SKU outperformed other methods by 10 to 19 times in reducing harmful responses, while keeping the models just as sharp on everyday tasks.

Even more impressively, it demonstrated the ability to generalize well to unseen harmful prompts, showcasing its robustness and adaptability.

As we continue to integrate AI into our daily lives, techniques like SKU pave the way for more trustworthy and ethically-aligned artificial intelligence. It's not just about making smarter AI anymore ā€“ it's about making AI that we can trust to be smart in the right ways.

This research could definitely influence how organizations develop AI systems, potentially establishing new best practices for deploying LLMs.

So, the next time you chat with an AI, remember ā€“ it might just be using its "selective memory" to keep things safe and sound. And that's something we can all feel good about! šŸ˜Š

Spark 'n' Trouble Shenanigans šŸ˜œ

We stumbled upon a LinkedIn gem that's too cool to keep to ourselves.

Ever dreamed of transforming your videos with just a few words? Well, buckle up!

We're about to dive into the world of video-to-video magic using Runway Gen-3. It's like giving your videos a makeover with the power of AI.

Curious? Let's check out this game-changing post together!šŸ”½

Well, thatā€™s a wrap!
Thanks for reading šŸ˜Š

See you next week with more mind-blowing tech insights šŸ’»

Until then,
Stay CuriousšŸ§  Stay AwesomešŸ¤©

PS: Do catch us on LinkedIn - Sandra & Tezan

Reply

or to participate.