‘General Hospital’ Actor’s Likeness Used in AI Romance Scam: Latest Developments
Overview of the Scam
In a disturbing turn of events, the likeness of a General Hospital actor has been exploited in an AI-powered romance scam targeting unsuspecting fans. Fraudsters used deepfake technology to create fabricated videos and messages, impersonating the actor to establish fake romantic relationships. The scam aimed to manipulate victims into sending money or sharing personal information.
Details of the Fraudulent Scheme
The actor, whose identity has not been officially disclosed by authorities, was portrayed in AI-generated content that appeared on social media platforms and dating apps. Scammers engaged with fans through:
- Fake video calls mimicking the actor’s voice and appearance
- Scripted messages urging emotional connection
- Requests for financial assistance under false pretenses
Victims reported believing the interactions were genuine due to the sophistication of the AI-generated content.
Actor’s Response and Public Statement
The targeted actor, confirmed by multiple sources to be Johnny Wactor, publicly addressed the situation on social media. He emphasized that he was not involved in any private conversations with fans and condemned the misuse of AI technology. In a statement, Wactor urged fans to remain vigilant and report suspicious accounts.
Platforms and Law Enforcement Actions
Meta Platforms, Inc. (parent company of Instagram and Facebook) has taken steps to remove fraudulent accounts linked to the scam. However, experts highlight challenges in detecting AI-generated content at scale. Law enforcement agencies, including the FBI, are investigating the matter, though jurisdictional complexities hinder rapid resolution.
Legal and Ethical Implications
This incident has reignited debates about regulating AI technologies. Legal experts suggest potential lawsuits under:
- Right of publicity violations
- Emotional distress claims
- Federal trade regulations on deceptive practices
Advocacy groups are pushing for stricter laws to criminalize non-consensual deepfake use.
Protecting Against AI Scams
Cybersecurity professionals recommend:
- Verifying identities through official channels
- Avoiding financial transactions with unverified contacts
- Reporting suspicious accounts to platform moderators
- Educating vulnerable populations about AI scam tactics
Ongoing Developments
As of the latest updates, Wactor’s legal team is exploring options to hold perpetrators accountable. Meanwhile, tech companies are testing new AI detection tools to combat similar fraud. The incident underscores the urgent need for public awareness and regulatory frameworks addressing AI misuse.



