We use essential cookies to make our site work. With your consent, we may also use non-essential cookies to improve user experience, personalize content, customize advertisements, and analyze website traffic. For these reasons, we may share your site usage data with our social media, advertising, and analytics partners. By clicking ”Accept,” you agree to our website's cookie use as described in our Cookie Policy. You can change your cookie settings at any time by clicking “Preferences.”

TechDogs-"Harry Potter And The Spell To Erase Memories Of Large Language Models!"

Emerging Technology

Harry Potter And The Spell To Erase Memories Of Large Language Models!

By Lakshana Raichandani

Updated on Wed, Oct 11, 2023

Overall Rating
"Happiness can be found, even in the darkest of times, if one only remembers to turn on the light." – Albus Dumbledore.

This quote has motivated almost every #Potterhead; in fact, from the characters to the plot to the quotes, everything about Harry Potter books (or movies, for that matter) has been iconic. Guess what? Harry Potter has come to the rescue for researchers to make AI (Artificial Intelligence) forget about copyrighted content.

According to a recent report, in the ongoing debate surrounding the utilization of copyrighted material in training large language models (LLMs), such as OpenAI's ChatGPT, Meta's Llama 2 and Anthropic's Claude 2, a crucial question emerges: can these models be modified to erase their knowledge of such works without undergoing extensive retraining or architectural revisions?

In a groundbreaking paper recently published on arXiv.org, Ronen Eldan of Microsoft Research and Mark Russinovich of Microsoft Azure propose an innovative approach. This approach removes specific information from a sample LLM, effectively wiping out all knowledge related to the Harry Potter books, including characters and plots, from Meta's open-source Llama 2-7B.

In the same paper, Microsoft researchers explain, "While the model took over 184K GPU-hours to pre-train, we show that in about 1 GPU hour of fine-tuning, we effectively erase the model's ability to generate or recall Harry Potter-related content."

TechDogs-"A Screengrab Of Ronen Eldan's Tweet On X."
This work represents a significant stride towards adaptable language models. The capability to fine-tune AI models according to evolving organizational requirements is essential for long-term, enterprise-safe deployments.

Let's decode how this AI' Marauder's Map' to forgetting Harry Potter came into the picture!
 
  • "Traditional models of machine learning predominantly focus on adding or reinforcing knowledge through basic fine-tuning but do not provide straightforward mechanisms to 'forget' or 'unlearn' knowledge," asserts the authors. So, how did they overcome this obstacle? The research paper devised a three-pronged technique for approximating unlearning specific information in LLMs.

  • Firstly, they trained a model using the target data (Harry Potter books) to identify tokens most closely associated with it by comparing predictions to a baseline model.

  • Secondly, they substituted unique Harry Potter expressions with generic equivalents and generated alternative predictions to mimic a model that had not been trained on that material.

  • Thirdly, they fine-tuned the baseline model using these alternative predictions, purging the original text from memory when provided context.

  • To gauge the effectiveness of their approach, they examined the model's ability to generate or discuss Harry Potter content using 300 automatically generated prompts and by assessing token probabilities. Eldan and Russinovich declare, "To the best of our knowledge, this is the first paper to present an effective technique for unlearning in generative language models."

  • Their findings revealed that after just an hour of fine-tuning using their technique, the original model could practically "forget" intricate narratives from the Harry Potter series. Its performance on standard benchmarks like ARC, BoolQ and Winogrande "remains almost unaffected."


You might be wondering why this jump from 'Expecto Patronum' to 'Forget-o Patronum' in erasing wizardry from AI. Let's see:
 
  • However, the authors acknowledge that further testing is warranted due to the limitations of their evaluation approach. Their technique may also be more suited for fictional texts than non-fiction, as fictional universes entail many distinctive references.

  • Nevertheless, this proof-of-concept signifies "a foundational step towards creating more responsible, adaptable, and legally compliant LLMs in the future." The authors posit that additional refinement could address concerns related to "ethical guidelines, societal values, or specific user requirements."

  • In conclusion, the authors state, "Our technique offers a promising start, but its applicability across various content types remains to be thoroughly tested. The presented approach offers a foundation, but further research is needed to refine and extend the methodology for broader unlearning tasks in LLMs."


In the future, developing more comprehensive and robust techniques for selective forgetting could ensure that AI systems remain dynamically aligned with evolving priorities, be they business or societal.

Do you this example with “Harry Potter” will emerge as a spell-bounding technique for researcher in AI? Will this help develop LLMs that do not scrape copyrighted works over the web?

Drop your thoughts in the comments section below!

First published on Wed, Oct 11, 2023

Liked what you read? That’s only the tip of the tech iceberg!

Explore our vast collection of tech articles including introductory guides, product reviews, trends and more, stay up to date with the latest news, relish thought-provoking interviews and the hottest AI blogs, and tickle your funny bone with hilarious tech memes!

Plus, get access to branded insights from industry-leading global brands through informative white papers, engaging case studies, in-depth reports, enlightening videos and exciting events and webinars.

Dive into TechDogs' treasure trove today and Know Your World of technology like never before!

Disclaimer - Reference to any specific product, software or entity does not constitute an endorsement or recommendation by TechDogs nor should any data or content published be relied upon. The views expressed by TechDogs' members and guests are their own and their appearance on our site does not imply an endorsement of them or any entity they represent. Views and opinions expressed by TechDogs' Authors are those of the Authors and do not necessarily reflect the view of TechDogs or any of its officials. While we aim to provide valuable and helpful information, some content on TechDogs' site may not have been thoroughly reviewed for every detail or aspect. We encourage users to verify any information independently where necessary.

Join The Discussion

Join Our Newsletter

Get weekly news, engaging articles, and career tips-all free!

By subscribing to our newsletter, you're cool with our terms and conditions and agree to our Privacy Policy.

  • Dark
  • Light