Microsoft chief warns more deepfake threats could be coming soon

When it comes to deepfakes , what we’ve seen so far is just the tip of the iceberg. In the near future, we won’t be certain if the person w...

When it comes to deepfakes, what we’ve seen so far is just the tip of the iceberg. In the near future, we won’t be certain if the person we’re speaking to on a video call is real, or an impostor, and crooks won’t have trouble creating an entire chronology of fake videos to support their claims, or trick people into believing the legitimacy of an offer or campaign.

These harrowing predictions come from Eric Horvitz, Microsoft’s chief science officer, in a new research paper, titled “On the horizon: Interactive and compositional deepfakes”.

Deepfakes are “photoshopped” videos, essentially. By using artificial intelligence (AI) and machine learning (ML), a threat actor is able to create a video of a person saying things that they never said. Now, according to Horvitz, crooks are ready to take it to the next level. Interactive deepfakes are just as you’d expect - real-time videos with which users can interact, which are, in reality, utterly fake.

Synthetic history

Compositional deepfakes, on the other hand, are described as “sets of deepfakes” designed to integrate over time with “observed, expected, and engineered world events to create persuasive synthetic histories.” 

“Synthetic histories can be constructed manually but may one day be guided by adversarial generative explanation (AGE) techniques,” Horvitz adds. 

He also says that in the near future, it will be almost impossible to distinguish fake videos and fake content from authentic ones: “In the absence of mitigations, interactive and compositional deepfakes threaten to move us closer to a post-epistemic world, where fact cannot be distinguished from fiction.”

This absence of mitigations stems from the fact that threat actors can pit artificial intelligence against analysis tools and develop deepfake content that is able to fool even the most advanced detection systems. 

“With this process at the foundation of deepfakes, neither pattern recognition techniques nor humans will be able to reliably recognize deepfakes,” Horvitz notes. 

So, next time a family member calls from abroad to ask for money to pay the rent, make sure it’s not a fraudster impersonating your loved ones. 

Via: VentureBeat



from TechRadar - All the latest technology news https://ift.tt/xhySI2K
via IFTTT

COMMENTS

BLOGGER
Name

Apps,3858,Business,151,Camera,1155,Earn $$$,3,Gadgets,1741,Games,926,GTA,1,Innovations,3,Mobile,1697,Paid Promotions,5,Promotions,5,Sports,1,Technology,8106,Trailers,796,Travel,37,Trending,4,Trendly News,25335,TrendlyNews,183,Video,5,XIAOMI,13,YouTube - 9to5Google,182,
ltr
item
Trendly News | #ListenNow #Everyday #100ShortNews #TopTrendings #PopularNews #Reviews #TrendlyNews: Microsoft chief warns more deepfake threats could be coming soon
Microsoft chief warns more deepfake threats could be coming soon
Trendly News | #ListenNow #Everyday #100ShortNews #TopTrendings #PopularNews #Reviews #TrendlyNews
http://www.trendlynews.in/2022/09/microsoft-chief-warns-more-deepfake.html
http://www.trendlynews.in/
http://www.trendlynews.in/
http://www.trendlynews.in/2022/09/microsoft-chief-warns-more-deepfake.html
true
3372890392287038985
UTF-8
Loaded All Posts Not found any posts VIEW ALL Readmore Reply Cancel reply Delete By Home PAGES POSTS View All RECOMMENDED FOR YOU LABEL ARCHIVE SEARCH ALL POSTS Not found any post match with your request Back Home Sunday Monday Tuesday Wednesday Thursday Friday Saturday Sun Mon Tue Wed Thu Fri Sat January February March April May June July August September October November December Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec just now 1 minute ago $$1$$ minutes ago 1 hour ago $$1$$ hours ago Yesterday $$1$$ days ago $$1$$ weeks ago more than 5 weeks ago Followers Follow THIS PREMIUM CONTENT IS LOCKED STEP 1: Share. STEP 2: Click the link you shared to unlock Copy All Code Select All Code All codes were copied to your clipboard Can not copy the codes / texts, please press [CTRL]+[C] (or CMD+C with Mac) to copy