Sora 2, OpenAI’s advanced model for generating realistic, high-quality videos from text or images, is being positioned as a breakthrough in video generation. OpenAISora 2, OpenAI’s advanced model for generating realistic, high-quality videos from text or images, is being positioned as a breakthrough in video generation. OpenAI

Sora 2: Deepfakes Waiting to Happen

Sora 2, OpenAI’s advanced model for generating realistic, high-quality videos from text or images, is being positioned as a breakthrough in video generation. OpenAI touts it as “more physically accurate, realistic, and more controllable than prior systems.” It is capable of synchronized dialogue, sound effects, and inserting real people into generated scenes with remarkable fidelity.  

While OpenAI frames it as a new era of creativity and communication, when viewed through a security lens, Sora 2 introduces a deeply uncomfortable problem: the easier it becomes to create realistic videos of real people, the easier it becomes to make them appear to do and say things they never did.  

By default, the cameos created in Sora 2 are available for anyone to use. If you do not change this setting, anyone could create a realistic video of you doing and saying things that would be out of character or which others might find offensive. 

That single design choice dramatically shifts the threat landscape. What once required sophisticated deepfake tooling and technical expertise is now packaged into a consumer-friendly app, complete with a social feed designed for remixing and reuse. 

Likeness, Metadata, and the Long Tail of Exposure 

When you upload photos and videos to Sora, the default is to retain any location information embedded in the media. Most social platforms strip that data out. However, even if the location information is stripped from the final created content, there’s always a risk of a data breach leading to information disclosure at some point in the future.  

Identity, likeness, voice, and metadata are all being pooled into a system designed for scale. 

OpenAI highlights that you can “drop yourself straight into any Sora scene” after a one-time recording to capture appearance and voice. The capability is very general, as it works for any human, animal, or object.  

From a creative standpoint, that is impressive. From a social engineering standpoint, it is incendiary. 

Romance Scams Get a Powerful New Tool 

Romance scams are an obvious beneficiary. Sora 2 and similar technologies will prove to be a highly effective tool for romance scammers, who currently go to great lengths to convince their victims that they are wealthy and always on the go.   

With Sora 2, a scammer can send videos of themselves sailing on a mega yacht, driving an exotic car, or flying in a private jet with just a few keystrokes. 

Once they develop a cameo of their persona, they can make that cameo do or say just about anything. Video, long treated as higher-trust evidence than text or images, becomes a persuasion engine that is far harder for victims to question. 

Business Scams are Not Here Yet, But They Will Be 

At first glance, the impact on traditional business scams appears limited. We don’t expect Sora 2 to have much of an impact on business scams such as payment diversion, since most payment discussions happen over email or voice, rather than via video.  

Also, Sora 2 videos also aren’t live; they must be pre-generated, which limits their usefulness for real-time impersonation. Yet that limitation is temporary. At the rate the technology is advancing, however, we imagine Sora being able to generate near instantaneous video from text prompts in the relatively near future. 

When that happens, the threat model shifts overnight. We expect to see all manner of abuse, such as your ‘boss’ calling you on Teams to ask you to wire a large sum of money to an offshore account. 

Employees as the New Attack Surface  

Even before real-time video becomes a reality, the business risks go beyond finance. Employees are users who have public profiles, recorded talks, internal presentations, and now, possibly Sora cameos. If those cameos are public, they can be reused (and abused) in ways those users never intended. 

Enterprises should discourage their staff from making their Sora 2 cameos public because they could be used to spread disinformation about the organization. The examples are frightening for a reason. Imagine the danger of a well-known CFO appearing to say that sales are down 50% from the prior quarter, or the impact of your CEO appearing to be drunk and verbally abusive in public. 

Sora 2 is explicitly touted as being good at realism, dialogue, and persistence of world state. OpenAI claims it can follow “intricate instructions spanning multiple shots while accurately persisting world state.”   

That makes fabricated footage more believable, more shareable, and far more damaging.  

Marketing Temptation and the Erosion of Trust  

There is also a quieter internal risk. It might be tempting for your marketing team to generate video content using your subject-matter experts’ likenesses. At scale, that sounds efficient, but without controls, it becomes dangerous. 

For entities that choose this route, the warning is clear: require the subject of the video to sign off on any content before it is published. Otherwise, your expert might be saying things that simply aren’t true on your website, effectively eroding public trust in them. 

Guardrails that Are Easy to Step Around  

Sora 2 claims to have controls around creating characters using life-like human images. In practice, this is easily bypassed by first converting the image to a cartoon or other artistic style using another AI tool, then using the stylized image to create the character.  

The assumption that a stylized image is harmless breaks down once motion, voice, and context reintroduce realism. 

Experiments with public figures show inconsistent enforcement. Sora 2 generated realistic likenesses of certain deceased presidents without issue, blocked others, and applied different policy rationales to different figures. The result is a system that is difficult to predict and easy to probe. 

OpenAI emphasizes that “with characters, you are in control of your likeness end-to-end with Sora.” You can revoke access and remove videos. However, this assumes that users understand the defaults and the risks.  

Once a convincing video has circulated, the damage to reputation may be done, regardless of whether the original asset is later removed. 

The issue with Sora 2 is not that it enables creativity, but that it lowers the bar for misuse. When realistic video can be generated from a cameo, and that cameo can be reused by others, control over how a person appears in public is no longer guaranteed. 

Even if it does not immediately change how traditional business scams operate, it introduces a new and largely unaddressed risk: employees and leaders can be made to appear to say or do things they never did, with real consequences for trust, reputation, and credibility. 

As the technology moves closer to real-time generation, the opportunity for abuse only increases. At that point, video is no longer something organizations can treat as inherently trustworthy; it becomes another channel that must be actively defended. 

Sorumluluk Reddi: Bu sitede yeniden yayınlanan makaleler, halka açık platformlardan alınmıştır ve yalnızca bilgilendirme amaçlıdır. MEXC'nin görüşlerini yansıtmayabilir. Tüm hakları telif sahiplerine aittir. Herhangi bir içeriğin üçüncü taraf haklarını ihlal ettiğini düşünüyorsanız, kaldırılması için lütfen [email protected] ile iletişime geçin. MEXC, içeriğin doğruluğu, eksiksizliği veya güncelliği konusunda hiçbir garanti vermez ve sağlanan bilgilere dayalı olarak alınan herhangi bir eylemden sorumlu değildir. İçerik, finansal, yasal veya diğer profesyonel tavsiye niteliğinde değildir ve MEXC tarafından bir tavsiye veya onay olarak değerlendirilmemelidir.

Ayrıca Şunları da Beğenebilirsiniz

Franklin Templeton CEO Dismisses 50bps Rate Cut Ahead FOMC

Franklin Templeton CEO Dismisses 50bps Rate Cut Ahead FOMC

The post Franklin Templeton CEO Dismisses 50bps Rate Cut Ahead FOMC appeared on BitcoinEthereumNews.com. Franklin Templeton CEO Jenny Johnson has weighed in on whether the Federal Reserve should make a 25 basis points (bps) Fed rate cut or 50 bps cut. This comes ahead of the Fed decision today at today’s FOMC meeting, with the market pricing in a 25 bps cut. Bitcoin and the broader crypto market are currently trading flat ahead of the rate cut decision. Franklin Templeton CEO Weighs In On Potential FOMC Decision In a CNBC interview, Jenny Johnson said that she expects the Fed to make a 25 bps cut today instead of a 50 bps cut. She acknowledged the jobs data, which suggested that the labor market is weakening. However, she noted that this data is backward-looking, indicating that it doesn’t show the current state of the economy. She alluded to the wage growth, which she remarked is an indication of a robust labor market. She added that retail sales are up and that consumers are still spending, despite inflation being sticky at 3%, which makes a case for why the FOMC should opt against a 50-basis-point Fed rate cut. In line with this, the Franklin Templeton CEO said that she would go with a 25 bps rate cut if she were Jerome Powell. She remarked that the Fed still has the October and December FOMC meetings to make further cuts if the incoming data warrants it. Johnson also asserted that the data show a robust economy. However, she noted that there can’t be an argument for no Fed rate cut since Powell already signaled at Jackson Hole that they were likely to lower interest rates at this meeting due to concerns over a weakening labor market. Notably, her comment comes as experts argue for both sides on why the Fed should make a 25 bps cut or…
Paylaş
BitcoinEthereumNews2025/09/18 00:36
What is the 3 5 7 rule in day trading? — A Practical Guide

What is the 3 5 7 rule in day trading? — A Practical Guide

This guide turns money anxiety into practical action. It shows how tracking, a forgiving budget, automatic savings, and small monthly rituals build real financial
Paylaş
Coinstats2026/01/24 00:47
‘Mercy’ Stars Chris Pratt And Rebecca Ferguson On The Dangers Of AI

‘Mercy’ Stars Chris Pratt And Rebecca Ferguson On The Dangers Of AI

The post ‘Mercy’ Stars Chris Pratt And Rebecca Ferguson On The Dangers Of AI appeared on BitcoinEthereumNews.com. Chris Pratt and Rebecca Ferguson attend the UK
Paylaş
BitcoinEthereumNews2026/01/24 01:33