Synthetic human-like fakes: Difference between revisions

→‎2020's synthetic human-like fakes: + <ref name="The Conversation 2022"> as {{cite web}} + set the access-date=2022-10-05 which was when I saw this first late in the evening/night
(→‎2020's synthetic human-like fakes: working in more links)
(→‎2020's synthetic human-like fakes: + <ref name="The Conversation 2022"> as {{cite web}} + set the access-date=2022-10-05 which was when I saw this first late in the evening/night)
Line 362: Line 362:
[[File:Appearance of Queen Elizabeth II stolen by Channel 4 in Dec 2020 (screenshot at 191s).png|thumb|right|480px|In Dec 2020 Channel 4 aired a Queen-like fake i.e. they had thieved the appearance of Queen Elizabeth II using deepfake methods.]]
[[File:Appearance of Queen Elizabeth II stolen by Channel 4 in Dec 2020 (screenshot at 191s).png|thumb|right|480px|In Dec 2020 Channel 4 aired a Queen-like fake i.e. they had thieved the appearance of Queen Elizabeth II using deepfake methods.]]


* '''2022''' | '''<font color="green">counter-measure</font>''' | The work [https://www.usenix.org/system/files/sec22fall_blue.pdf '''''Who Are You (I Really Wanna Know)? Detecting Audio DeepFakes Through Vocal Tract Reconstruction''''' at usenix.org], [https://www.usenix.org/conference/usenixsecurity22/presentation/blue presentation page], [https://www.usenix.org/system/files/sec22-blue.pdf version included in the proceedings] and [https://www.usenix.org/system/files/sec22_slides-blue.pdf slides] from researchers of the Florida Institute for Cybersecurity Research (FICS) in the [[w:University of Florida]] received funding from the [[w:Office of Naval Research]] and was presented in August 2020 at the [[w:USENIX]] Security Symposium. The scientists wrote an article on their work titled [https://theconversation.com/deepfake-audio-has-a-tell-researchers-use-fluid-dynamics-to-spot-artificial-imposter-voices-189104 '''''Deepfake audio has a tell – researchers use fluid dynamics to spot artificial imposter voices''''' at theconversation.com] that was published Tuesday 2022-09-20.  
* '''2022''' | '''<font color="green">counter-measure</font>''' | The work [https://www.usenix.org/system/files/sec22fall_blue.pdf '''''Who Are You (I Really Wanna Know)? Detecting Audio DeepFakes Through Vocal Tract Reconstruction''''' at usenix.org], [https://www.usenix.org/conference/usenixsecurity22/presentation/blue presentation page], [https://www.usenix.org/system/files/sec22-blue.pdf version included in the proceedings] and [https://www.usenix.org/system/files/sec22_slides-blue.pdf slides] from researchers of the Florida Institute for Cybersecurity Research (FICS) in the [[w:University of Florida]] received funding from the [[w:Office of Naval Research]] and was presented in August 2020 at the [[w:USENIX]] Security Symposium. The scientists wrote an article on their work titled [https://theconversation.com/deepfake-audio-has-a-tell-researchers-use-fluid-dynamics-to-spot-artificial-imposter-voices-189104 '''''Deepfake audio has a tell – researchers use fluid dynamics to spot artificial imposter voices''''' at theconversation.com] that was published Tuesday 2022-09-20.<ref name="The Conversation 2022">
{{cite web
|url=https://theconversation.com/deepfake-audio-has-a-tell-researchers-use-fluid-dynamics-to-spot-artificial-imposter-voices-189104
|title=Deepfake audio has a tell – researchers use fluid dynamics to spot artificial imposter voices
|last=Blue
|first=Logan
|last2=Traynor
|first2=Patrick
|date=2022-09-20
|website=theconversation.com
|publisher=[[w:The Conversation]]
|access-date=2022-10-05
|quote=By estimating the anatomy responsible for creating the observed speech, it’s possible to identify the whether the audio was generated by a person or a computer.}}
 
</ref>


* '''2022''' | '''<font color="green">counter-measure</font>''' | [https://arxiv.org/pdf/2206.12043.pdf '''''Protecting President Zelenskyy against deep fakes''''' a 2022 preprint at arxiv.org] by Matyáš Boháček of Johannes Kepler Gymnasium and [[w:Hany Farid]], the dean and head of of [[w:University of California, Berkeley School of Information|w:Berkeley School of Information at the University of California, Berkeley]]. This brief paper describes their automated digital look-alike detection system and evaluate its efficacy and reliability in comparison to humans with untrained eyes. Their work provides automated evaluation tools to catch so called "deep fakes" and their motivation seems to have been to find automation armor against disinformation warfare against humans and the humanity. Automated digital [[Glossary#Media forensics|media forensics]] is a very good idea explored by many.  Boháček and Farid 2022 detection system works by evaluating both facial mannerisms as well as gestural mannerisms to detect the non-human ones from the ones that are human in origin.  
* '''2022''' | '''<font color="green">counter-measure</font>''' | [https://arxiv.org/pdf/2206.12043.pdf '''''Protecting President Zelenskyy against deep fakes''''' a 2022 preprint at arxiv.org] by Matyáš Boháček of Johannes Kepler Gymnasium and [[w:Hany Farid]], the dean and head of of [[w:University of California, Berkeley School of Information|w:Berkeley School of Information at the University of California, Berkeley]]. This brief paper describes their automated digital look-alike detection system and evaluate its efficacy and reliability in comparison to humans with untrained eyes. Their work provides automated evaluation tools to catch so called "deep fakes" and their motivation seems to have been to find automation armor against disinformation warfare against humans and the humanity. Automated digital [[Glossary#Media forensics|media forensics]] is a very good idea explored by many.  Boháček and Farid 2022 detection system works by evaluating both facial mannerisms as well as gestural mannerisms to detect the non-human ones from the ones that are human in origin.  
Line 375: Line 389:
|last=Rosner
|last=Rosner
|first=Helen
|first=Helen
|author-link=Helen Rosner
|author-link=[[w:Helen Rosner]]
|date=2021-07-15
|date=2021-07-15
|title=A Haunting New Documentary About Anthony Bourdain
|title=A Haunting New Documentary About Anthony Bourdain