• About
  • Disclaimer
  • Privacy Policy
  • Contact
Saturday, May 24, 2025
Cyber Defense GO
  • Login
  • Home
  • Cyber Security
  • Artificial Intelligence
  • Machine Learning
  • Data Analysis
  • Computer Networking
  • Disaster Restoration
No Result
View All Result
  • Home
  • Cyber Security
  • Artificial Intelligence
  • Machine Learning
  • Data Analysis
  • Computer Networking
  • Disaster Restoration
No Result
View All Result
Cyber Defense Go
No Result
View All Result
Home Artificial Intelligence

Mapping the misuse of generative AI

Md Sazzad Hossain by Md Sazzad Hossain
0
Mapping the misuse of generative AI
585
SHARES
3.2k
VIEWS
Share on FacebookShare on Twitter


Accountability & Security

Printed
2 August 2024
Authors

Nahema Marchal and Rachel Xu

Abstract artwork to depict generative AI and revealing layers of insight and data

New analysis analyzes the misuse of multimodal generative AI at this time, so as to assist construct safer and extra accountable applied sciences

Generative synthetic intelligence (AI) fashions that may produce picture, textual content, audio, video and extra are enabling a brand new period of creativity and industrial alternative. But, as these capabilities develop, so does the potential for his or her misuse, together with manipulation, fraud, bullying or harassment.

As a part of our dedication to develop and use AI responsibly, we printed a new paper, in partnership with Jigsaw and Google.org, analyzing how generative AI applied sciences are being misused at this time. Groups throughout Google are utilizing this and different analysis to develop higher safeguards for our generative AI applied sciences, amongst different security initiatives.

Collectively, we gathered and analyzed practically 200 media reviews capturing public incidents of misuse, printed between January 2023 and March 2024. From these reviews, we outlined and categorized frequent techniques for misusing generative AI and located novel patterns in how these applied sciences are being exploited or compromised.

By clarifying the present threats and techniques used throughout several types of generative AI outputs, our work may also help form AI governance and information firms like Google and others constructing AI applied sciences in growing extra complete security evaluations and mitigation methods.

Highlighting the principle classes of misuse

Whereas generative AI instruments characterize a novel and compelling means to boost creativity, the flexibility to provide bespoke, life like content material has the potential for use in inappropriate methods by malicious actors.

By analyzing media reviews, we recognized two most important classes of generative AI misuse techniques: the exploitation of generative AI capabilities and the compromise of generative AI techniques. Examples of the applied sciences being exploited included creating life like depictions of human likenesses to impersonate public figures; whereas situations of the applied sciences being compromised included ‘jailbreaking’ to take away mannequin safeguards and utilizing adversarial inputs to trigger malfunctions.

Relative frequency generative AI misuse techniques in our dataset. Any given case of misuse reported within the media may contain a number of techniques.

Circumstances of exploitation — involving malicious actors exploiting simply accessible, consumer-level generative AI instruments, usually in ways in which didn’t require superior technical abilities — have been probably the most prevalent in our dataset. For instance, we reviewed a high-profile case from February 2024 the place a world firm reportedly misplaced HK$200 million (approx. US $26M) after an worker was tricked into making a monetary switch throughout a web-based assembly. On this occasion, each different “individual” within the assembly, together with the corporate’s chief monetary officer, was the truth is a convincing, computer-generated imposter.

A few of the most outstanding techniques we noticed, resembling impersonation, scams, and artificial personas, pre-date the invention of generative AI and have lengthy been used to affect the knowledge ecosystem and manipulate others. However wider entry to generative AI instruments could alter the prices and incentives behind info manipulation, giving these age-old techniques new efficiency and potential, particularly to those that beforehand lacked the technical sophistication to include such techniques.

Figuring out methods and mixtures of misuse

Falsifying proof and manipulating human likenesses underlie probably the most prevalent techniques in real-world instances of misuse. Within the time interval we analyzed, most instances of generative AI misuse have been deployed in efforts to affect public opinion, allow scams or fraudulent actions, or to generate revenue.

By observing how dangerous actors mix their generative AI misuse techniques in pursuit of their numerous targets, we recognized particular mixtures of misuse and labeled these mixtures as methods.

Diagram of how the targets of dangerous actors (left) map onto their methods of misuse (proper).

Rising types of generative AI misuse, which aren’t overtly malicious, nonetheless increase moral issues. For instance, new types of political outreach are blurring the strains between authenticity and deception, resembling authorities officers out of the blue talking quite a lot of voter-friendly languages with out clear disclosure that they’re utilizing generative AI, and activists utilizing the AI-generated voices of deceased victims to plead for gun reform.

Whereas the research supplies novel insights on rising types of misuse, it’s value noting that this dataset is a restricted pattern of media reviews. Media reviews could prioritize sensational incidents, which in flip could skew the dataset in direction of explicit varieties of misuse. Detecting or reporting instances of misuse may additionally be more difficult for these concerned as a result of generative AI techniques are so novel. The dataset additionally doesn’t make a direct comparability between misuse of generative AI techniques and conventional content material creation and manipulation techniques, resembling picture modifying or establishing ‘content material farms’ to create massive quantities of textual content, video, gifs, pictures and extra. Thus far, anecdotal proof means that conventional content material manipulation techniques stay extra prevalent.

Staying forward of potential misuses

Our paper highlights alternatives to design initiatives that defend the general public, resembling advancing broad generative AI literacy campaigns, growing higher interventions to guard the general public from dangerous actors, or forewarning individuals and equipping them to identify and refute the manipulative methods utilized in generative AI misuse.

This analysis helps our groups higher safeguard our merchandise by informing our improvement of security initiatives. On YouTube, we now require creators to share when their work is meaningfully altered or synthetically generated, and appears life like. Equally, we up to date our election promoting insurance policies to require advertisers to reveal when their election adverts embody materials that has been digitally altered or generated.

As we proceed to increase our understanding of malicious makes use of of generative AI and make additional technical developments, we all know it’s extra essential than ever to verify our work isn’t occurring in a silo. We just lately joined the Content material for Coalition Provenance and Authenticity (C2PA) as a steering committee member to assist develop the technical customary and drive adoption of Content material Credentials, that are tamper-resistant metadata that exhibits how content material was made and edited over time.

In parallel, we’re additionally conducting analysis that advances present red-teaming efforts, together with enhancing greatest practices for testing the protection of huge language fashions (LLMs), and growing pioneering instruments to make AI-generated content material simpler to determine, resembling SynthID, which is being built-in right into a rising vary of merchandise.

In recent times, Jigsaw has performed analysis with misinformation creators to grasp the instruments and techniques they use, developed prebunking movies to forewarn individuals of makes an attempt to control them, and proven that prebunking campaigns can enhance misinformation resilience at scale. This work varieties a part of Jigsaw’s broader portfolio of knowledge interventions to assist individuals defend themselves on-line.

By proactively addressing potential misuses, we will foster accountable and moral use of generative AI, whereas minimizing its dangers. We hope these insights on the commonest misuse techniques and techniques will assist researchers, policymakers, business belief and security groups construct safer, extra accountable applied sciences and develop higher measures to fight misuse.

Acknowledgements

This analysis was a collective effort by Nahema Marchal, Rachel Xu, Rasmi Elasmar, Iason Gabriel, Beth Goldberg, and William Isaac, with suggestions and advisory contributions from Mikel Rodriguez, Vijay Bolina, Alexios Mantzarlis, Seliem El-Sayed, Mevan Babakar, Matt Botvinick, Canfer Akbulut, Harry Regulation, Sébastien Krier, Ziad Reslan, Boxi Wu, Frankie Garcia, and Jennie Brennan.

You might also like

10 Finest AI Music Video Turbines (Could 2025)

Gemini as a common AI assistant

Studying methods to predict uncommon sorts of failures | MIT Information


Accountability & Security

Printed
2 August 2024
Authors

Nahema Marchal and Rachel Xu

Abstract artwork to depict generative AI and revealing layers of insight and data

New analysis analyzes the misuse of multimodal generative AI at this time, so as to assist construct safer and extra accountable applied sciences

Generative synthetic intelligence (AI) fashions that may produce picture, textual content, audio, video and extra are enabling a brand new period of creativity and industrial alternative. But, as these capabilities develop, so does the potential for his or her misuse, together with manipulation, fraud, bullying or harassment.

As a part of our dedication to develop and use AI responsibly, we printed a new paper, in partnership with Jigsaw and Google.org, analyzing how generative AI applied sciences are being misused at this time. Groups throughout Google are utilizing this and different analysis to develop higher safeguards for our generative AI applied sciences, amongst different security initiatives.

Collectively, we gathered and analyzed practically 200 media reviews capturing public incidents of misuse, printed between January 2023 and March 2024. From these reviews, we outlined and categorized frequent techniques for misusing generative AI and located novel patterns in how these applied sciences are being exploited or compromised.

By clarifying the present threats and techniques used throughout several types of generative AI outputs, our work may also help form AI governance and information firms like Google and others constructing AI applied sciences in growing extra complete security evaluations and mitigation methods.

Highlighting the principle classes of misuse

Whereas generative AI instruments characterize a novel and compelling means to boost creativity, the flexibility to provide bespoke, life like content material has the potential for use in inappropriate methods by malicious actors.

By analyzing media reviews, we recognized two most important classes of generative AI misuse techniques: the exploitation of generative AI capabilities and the compromise of generative AI techniques. Examples of the applied sciences being exploited included creating life like depictions of human likenesses to impersonate public figures; whereas situations of the applied sciences being compromised included ‘jailbreaking’ to take away mannequin safeguards and utilizing adversarial inputs to trigger malfunctions.

Relative frequency generative AI misuse techniques in our dataset. Any given case of misuse reported within the media may contain a number of techniques.

Circumstances of exploitation — involving malicious actors exploiting simply accessible, consumer-level generative AI instruments, usually in ways in which didn’t require superior technical abilities — have been probably the most prevalent in our dataset. For instance, we reviewed a high-profile case from February 2024 the place a world firm reportedly misplaced HK$200 million (approx. US $26M) after an worker was tricked into making a monetary switch throughout a web-based assembly. On this occasion, each different “individual” within the assembly, together with the corporate’s chief monetary officer, was the truth is a convincing, computer-generated imposter.

A few of the most outstanding techniques we noticed, resembling impersonation, scams, and artificial personas, pre-date the invention of generative AI and have lengthy been used to affect the knowledge ecosystem and manipulate others. However wider entry to generative AI instruments could alter the prices and incentives behind info manipulation, giving these age-old techniques new efficiency and potential, particularly to those that beforehand lacked the technical sophistication to include such techniques.

Figuring out methods and mixtures of misuse

Falsifying proof and manipulating human likenesses underlie probably the most prevalent techniques in real-world instances of misuse. Within the time interval we analyzed, most instances of generative AI misuse have been deployed in efforts to affect public opinion, allow scams or fraudulent actions, or to generate revenue.

By observing how dangerous actors mix their generative AI misuse techniques in pursuit of their numerous targets, we recognized particular mixtures of misuse and labeled these mixtures as methods.

Diagram of how the targets of dangerous actors (left) map onto their methods of misuse (proper).

Rising types of generative AI misuse, which aren’t overtly malicious, nonetheless increase moral issues. For instance, new types of political outreach are blurring the strains between authenticity and deception, resembling authorities officers out of the blue talking quite a lot of voter-friendly languages with out clear disclosure that they’re utilizing generative AI, and activists utilizing the AI-generated voices of deceased victims to plead for gun reform.

Whereas the research supplies novel insights on rising types of misuse, it’s value noting that this dataset is a restricted pattern of media reviews. Media reviews could prioritize sensational incidents, which in flip could skew the dataset in direction of explicit varieties of misuse. Detecting or reporting instances of misuse may additionally be more difficult for these concerned as a result of generative AI techniques are so novel. The dataset additionally doesn’t make a direct comparability between misuse of generative AI techniques and conventional content material creation and manipulation techniques, resembling picture modifying or establishing ‘content material farms’ to create massive quantities of textual content, video, gifs, pictures and extra. Thus far, anecdotal proof means that conventional content material manipulation techniques stay extra prevalent.

Staying forward of potential misuses

Our paper highlights alternatives to design initiatives that defend the general public, resembling advancing broad generative AI literacy campaigns, growing higher interventions to guard the general public from dangerous actors, or forewarning individuals and equipping them to identify and refute the manipulative methods utilized in generative AI misuse.

This analysis helps our groups higher safeguard our merchandise by informing our improvement of security initiatives. On YouTube, we now require creators to share when their work is meaningfully altered or synthetically generated, and appears life like. Equally, we up to date our election promoting insurance policies to require advertisers to reveal when their election adverts embody materials that has been digitally altered or generated.

As we proceed to increase our understanding of malicious makes use of of generative AI and make additional technical developments, we all know it’s extra essential than ever to verify our work isn’t occurring in a silo. We just lately joined the Content material for Coalition Provenance and Authenticity (C2PA) as a steering committee member to assist develop the technical customary and drive adoption of Content material Credentials, that are tamper-resistant metadata that exhibits how content material was made and edited over time.

In parallel, we’re additionally conducting analysis that advances present red-teaming efforts, together with enhancing greatest practices for testing the protection of huge language fashions (LLMs), and growing pioneering instruments to make AI-generated content material simpler to determine, resembling SynthID, which is being built-in right into a rising vary of merchandise.

In recent times, Jigsaw has performed analysis with misinformation creators to grasp the instruments and techniques they use, developed prebunking movies to forewarn individuals of makes an attempt to control them, and proven that prebunking campaigns can enhance misinformation resilience at scale. This work varieties a part of Jigsaw’s broader portfolio of knowledge interventions to assist individuals defend themselves on-line.

By proactively addressing potential misuses, we will foster accountable and moral use of generative AI, whereas minimizing its dangers. We hope these insights on the commonest misuse techniques and techniques will assist researchers, policymakers, business belief and security groups construct safer, extra accountable applied sciences and develop higher measures to fight misuse.

Acknowledgements

This analysis was a collective effort by Nahema Marchal, Rachel Xu, Rasmi Elasmar, Iason Gabriel, Beth Goldberg, and William Isaac, with suggestions and advisory contributions from Mikel Rodriguez, Vijay Bolina, Alexios Mantzarlis, Seliem El-Sayed, Mevan Babakar, Matt Botvinick, Canfer Akbulut, Harry Regulation, Sébastien Krier, Ziad Reslan, Boxi Wu, Frankie Garcia, and Jennie Brennan.

Tags: GenerativeMappingmisuse
Previous Post

Greatest VPN Journey Router: High Transportable Router for Safe Connections

Next Post

Google shares its first Well being Affect Report

Md Sazzad Hossain

Md Sazzad Hossain

Related Posts

10 Finest AI Music Video Turbines (Could 2025)
Artificial Intelligence

10 Finest AI Music Video Turbines (Could 2025)

by Md Sazzad Hossain
May 24, 2025
Gemini as a common AI assistant
Artificial Intelligence

Gemini as a common AI assistant

by Md Sazzad Hossain
May 23, 2025
Studying methods to predict uncommon sorts of failures | MIT Information
Artificial Intelligence

Studying methods to predict uncommon sorts of failures | MIT Information

by Md Sazzad Hossain
May 23, 2025
Microsoft AI Introduces Magentic-UI: An Open-Supply Agent Prototype that Works with Folks to Full Complicated Duties that Require Multi-Step Planning and Browser Use
Artificial Intelligence

Microsoft AI Introduces Magentic-UI: An Open-Supply Agent Prototype that Works with Folks to Full Complicated Duties that Require Multi-Step Planning and Browser Use

by Md Sazzad Hossain
May 23, 2025
Katy Perry Didn’t Attend the Met Gala, However AI Made Her the Star of the Evening
Artificial Intelligence

Katy Perry Didn’t Attend the Met Gala, However AI Made Her the Star of the Evening

by Md Sazzad Hossain
May 22, 2025
Next Post
Google shares its first Well being Affect Report

Google shares its first Well being Affect Report

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Recommended

Demis Hassabis & John Jumper awarded Nobel Prize in Chemistry

Demis Hassabis & John Jumper awarded Nobel Prize in Chemistry

February 4, 2025
Phil Tomlinson, SVP, International Choices at TaskUs – Interview Sequence

Phil Tomlinson, SVP, International Choices at TaskUs – Interview Sequence

March 6, 2025

Categories

  • Artificial Intelligence
  • Computer Networking
  • Cyber Security
  • Data Analysis
  • Disaster Restoration
  • Machine Learning

CyberDefenseGo

Welcome to CyberDefenseGo. We are a passionate team of technology enthusiasts, cybersecurity experts, and AI innovators dedicated to delivering high-quality, insightful content that helps individuals and organizations stay ahead of the ever-evolving digital landscape.

Recent

10 Finest AI Music Video Turbines (Could 2025)

10 Finest AI Music Video Turbines (Could 2025)

May 24, 2025
Confirmed Methods to Reshape Your Assault Floor with Fidelis Deception

Confirmed Methods to Reshape Your Assault Floor with Fidelis Deception

May 24, 2025

Search

No Result
View All Result

© 2025 CyberDefenseGo - All Rights Reserved

No Result
View All Result
  • Home
  • Cyber Security
  • Artificial Intelligence
  • Machine Learning
  • Data Analysis
  • Computer Networking
  • Disaster Restoration

© 2025 CyberDefenseGo - All Rights Reserved

Welcome Back!

Login to your account below

Forgotten Password?

Retrieve your password

Please enter your username or email address to reset your password.

Log In