in

Disrupting a world cybercrime community abusing generative AI


In an amended complaint to recent civil litigation, Microsoft is naming the primary developers of malicious tools designed to bypass the guardrails of generative AI services, including Microsoft’s Azure OpenAI Service. We are pursuing this legal action now against identified defendants to stop their conduct, to continue to dismantle their illicit operation, and to deter others intent on weaponizing our AI technology.

The individuals named are: (1) Arian Yadegarnia aka “Fiz” of Iran, (2) Alan Krysiak aka “Drago” of United Kingdom, (3) Ricky Yuen aka “cg-dot” of Hong Kong, China, and (4) Phát Phùng Tấn aka “Asakuri” of Vietnam. These actors are at the center of a global cybercrime network Microsoft tracks as Storm-2139. Members of Storm-2139 exploited exposed customer credentials scraped from public sources to unlawfully access accounts with certain generative AI services. They then altered the capabilities of these services and resold access to other malicious actors, providing detailed instructions on how to generate harmful and illicit content, including non-consensual intimate images of celebrities and other sexually explicit content.

This activity is prohibited under the terms of use for our generative AI services and required deliberate efforts to bypass our safeguards. We are not naming specific celebrities to keep their identities private and have excluded synthetic imagery and prompts from our filings to prevent the further circulation of harmful content.

Storm-2139: A global network of creators, providers and end users.

In December 2024, Microsoft’s Digital Crimes Unit (DCU) filed a lawsuit in the Eastern District of Virginia alleging various causes of action against 10 unidentified “John Does” participating in activities that violate U.S. law and Microsoft’s Acceptable Use Policy and Code of Conduct. Through this initial filing, we were able to gather more information about the operations of the criminal enterprise.

Storm-2139 is organized into three main categories: creators, providers, and users. Creators developed the illicit tools that enabled the abuse of AI generated services. Providers then modified and supplied these tools to end users often with varying tiers of service and payment. Finally, users then used these tools to generate violating synthetic content, often centered around celebrities and sexual imagery.

Below is a visual representation of Storm-2139, which displays internet aliases uncovered as part of our investigation as well as the countries in which we believe the associated personas are located.

Storm-2139’s organizational structure.
This image appears to be a LinkedIn profile page. Profile Picture: Features a red flower. Name: Arian Yadegarnia Title: Student at Emam Ali Location: Iran Activity: Indicates no posts and 0 followers. Education: Lists an associate's degree in Computer Programming, Specific Applications at Emam Ali, spanning from 2006 to 2026. Options: Buttons for "Connect," "Message," and "More" are available.Screenshot of “Fiz’s” LinkedIn profile

Through its ongoing investigation, Microsoft has identified several of the above-listed personas, including, but not limited to, the four named defendants. While we have identified two actors located in the United States—specifically, in Illinois and Florida—those identities remain undisclosed to avoid interfering with potential criminal investigations. Microsoft is preparing criminal referrals to United States and foreign law enforcement representatives.

Cybercriminals react to Microsoft’s website seizure and court filing.

As part of our initial filing, the Court issued a temporary restraining order and preliminary injunction enabling Microsoft to seize a website instrumental to the criminal operation, effectively disrupting the group’s ability to operationalize their services. The seizure of this website and subsequent unsealing of the legal filings in January generated an immediate reaction from actors, in some cases causing group members to turn on and point fingers at one another. We observed chatter about the lawsuit on the group’s monitored communication channels, speculating on the identities of the “John Does” and potential consequences.

The image shows text from a discussion board with two posts: The first post mentions someone named "I did" who leaked a name on Git and Replit. It humorously states the name isn’t "Arianna" or "Ariel" but "Arian," followed by laughter. The second post contains a URL to a LinkedIn profile with "arian" in the address, suggesting a connection to the first post. Both posts have timestamps and are marked by an anonymous user.Screenshot of online chatter discussing “Fiz’s” real name.

In these channels, certain members also “doxed” Microsoft’s counsel of record, posting their names, pDoxing can result in real-world harm, ranging from identity theft to harassment.

The image shows a discussion board post with details: The title image includes "DESU DEEP STRIKING INTO THIS THREAD." The post, made by an anonymous user, contains a URL with "/offendants." It mentions having work profile images of attorneys, archived links to their work profiles with backstory, and contact info from PDF documents. The post suggests having enough information to make cards and notes they are being watched. The content hints at information gathering and monitoring.Screenshot from post on online channels providing information about the case lawyers.

As a result, Microsoft’s counsel received a variety of emails, including several from suspected members of Storm-2139 attempting to cast blame on other members of the operation.

The text appears to be an email or message providing information about a suspected illegal activity involving Azure keys and proxy software. Content: Mentions a Discord server and a website "scylla.wtf" where Azure keys are sold. Identifies a main Discord ID and describes a proxy with over 3500 users. Provides links to proxy software and related sites. Claims the proxies are stealing keys and lying about their donation. Describes a criminal enterprise potentially stealing millions from Azure. Notable People: Lists individuals like "dragOn3xt," "Mr. Yae (asakuri)," and "rarestmeow." Note: Warns these are professionals and suggests considering restraining orders, offering more information if needed.

The image depicts an email interface with the following details: Subject: "drago" From: whistling@dmc.chat To: Uriarte, Robert L. Date: Saturday, January 11, 2025, at 08:25 Content: Mentions looking for a person named "drago." Indicates he is the person being sought. States he is using and selling Azure. Notes that all his information is public. The email seems to be a tip or warning about an individual involved in questionable activities related to Azure.Screenshots of emails received by counsel of record.

This reaction underscores the impact of Microsoft’s legal actions and demonstrates how these measures can effectively disrupt a cybercriminal network by seizing infrastructure and create a powerful deterrent impact among its members.

Continuing our commitment to combatting the abuse of generative AI.

We take the misuse of AI very seriously, recognizing the serious and lasting impacts of abusive imagery for victims. Microsoft remains committed to protecting users by embedding robust AI guardrails and safeguarding our services from illegal and harmful content. Last year, we committed to continuing to innovate on new ways to keep users safe by outlining a comprehensive approach to combat abusive AI-generated content. We published a whitepaper with recommendations for U.S. policymakers on modernizing criminal law to equip law enforcement with the tools necessary to bring bad actors to justice. We also provided an update on our approach to intimate image abuse, detailing the steps we take to protect our services from such harm, whether synthetic or otherwise.

As we’ve said before, no disruption is complete in one day. Going after malicious actors requires persistence and ongoing vigilance. By unmasking these individuals and shining a light on their malicious activities, Microsoft aims to set a precedent in the fight against AI technology misuse.

Tags: AI, cybercrime, Digital Crimes Unit, Microsoft Azure OpenAI Service, Microsoft Digital Crimes Unit, Responsible AI



Source link

Leave a Reply

Your email address will not be published. Required fields are marked *

GIPHY App Key not set. Please check settings

Deal: This Jackery energy station matches in your palm, and is 40% off

Okta, Inc. (OKTA) This fall 2025 Earnings Name Transcript