Audio deepfake

Audio deepfake technology, also referred to as voice cloning or deepfake audio, is an application of artificial intelligence designed to generate speech that convincingly mimics specific individuals, often synthesizing phrases or sentences they have never spoken.[1][2][3][4] Initially developed with the intent to enhance various aspects of human life, it has practical applications such as generating audiobooks and assisting individuals who have lost their voices due to medical conditions.[5][6] Additionally, it has commercial uses, including the creation of personalized digital assistants, natural-sounding text-to-speech systems, and advanced speech translation services.[7]

Incidents of fraud

Audio deepfakes, referred to as audio manipulations beginning in the early 2020s, are becoming widely accessible using simple mobile devices or personal computers.[8] These tools have also been used to spread misinformation using audio.[3] This has led to cybersecurity concerns among the global public about the side effects of using audio deepfakes, including its possible role in disseminating misinformation and disinformation in audio-based social media platforms.[9] People can use them as a logical access voice spoofing technique,[10] where they can be used to manipulate public opinion for propaganda, defamation, or terrorism. Vast amounts of voice recordings are daily transmitted over the Internet, and spoofing detection is challenging.[11] Audio deepfake attackers have targeted individuals and organizations, including politicians and governments.[12]

In 2019, scammers using AI impersonated the voice of the CEO of a German energy company and directed the CEO of its UK subsidiary to transfer 220,000.[13] In early 2020, the same technique impersonated a company director as part of an elaborate scheme that convinced a branch manager to transfer $35 million.[14]

According to a 2023 global McAfee survey, one person in ten reported having been targeted by an AI voice cloning scam; 77% of these targets reported losing money to the scam.[15][16] Audio deepfakes could also pose a danger to voice ID systems currently used by financial institutions.[17][18] In March 2023, the United States Federal Trade Commission issued a warning to consumers about the use of AI to fake the voice of a family member in distress asking for money.[19]

In October 2023, during the start of the British Labour Party's conference in Liverpool, an audio deepfake of Labour leader Keir Starmer was released that falsely portrayed him verbally abusing his staffers and criticizing Liverpool.[20] That same month, an audio deepfake of Slovak politician Michal Šimečka falsely claimed to capture him discussing ways to rig the upcoming election.[21]

During the campaign for the 2024 New Hampshire Democratic presidential primary, over 20,000 voters received robocalls from an AI-impersonated President Joe Biden urging them not to vote.[22][23] The New Hampshire attorney general said this violated state election laws, and alleged involvement by Life Corporation and Lingo Telecom.[24] In February 2024, the United States Federal Communications Commission banned the use of AI to fake voices in robocalls.[25][26] That same month, political consultant Steve Kramer admitted that he had commissioned the calls for $500. He said that he wanted to call attention to the need for rules governing the use of AI in political campaigns.[27] In May, the FCC said that Kramer had violated federal law by spoofing the number of a local political figure, and proposed a fine of $6 million. Four New Hampshire counties indicted Kramer on felony counts of voter suppression, and impersonating a candidate, a misdemeanor.[28]

Categories

Audio deepfakes can be divided into three different categories:

Replay-based

Replay-based deepfakes are malicious works that aim to reproduce a recording of the interlocutor's voice.[29]

There are two types: far-field detection and cut-and-paste detection. In far-field detection, a microphone recording of the victim is played as a test segment on a hands-free phone.[30] On the other hand, cut-and-paste involves faking the requested sentence from a text-dependent system.[11] Text-dependent speaker verification can be used to defend against replay-based attacks.[29][31] A current technique that detects end-to-end replay attacks is the use of deep convolutional neural networks.[32]

Synthetic-based

A block diagram illustrating the synthetic-based approach for generating audio deepfakes
The Synthetic-based approach diagram

The category based on speech synthesis refers to the artificial production of human speech, using software or hardware system programs. Speech synthesis includes Text-To-Speech, which aims to transform the text into acceptable and natural speech in real-time,[33] making the speech sound in line with the text input, using the rules of linguistic description of the text.

A classical system of this type consists of three modules: a text analysis model, an acoustic model, and a vocoder. The generation usually has to follow two essential steps. It is necessary to collect clean and well-structured raw audio with the transcripted text of the original speech audio sentence. Second, the Text-To-Speech model must be trained using these data to build a synthetic audio generation model.

Specifically, the transcribed text with the target speaker's voice is the input of the generation model. The text analysis module processes the input text and converts it into linguistic features. Then, the acoustic module extracts the parameters of the target speaker from the audio data based on the linguistic features generated by the text analysis module.[8] Finally, the vocoder learns to create vocal waveforms based on the parameters of the acoustic features. The final audio file is generated, including the synthetic simulation audio in a waveform format, creating speech audio in the voice of many speakers, even those not in training.

The first breakthrough in this regard was introduced by WaveNet,[34] a neural network for generating raw audio waveforms capable of emulating the characteristics of many different speakers. This network has been overtaken over the years by other systems[35][36][37][38][39][40] which synthesize highly realistic artificial voices within everyone’s reach.[41]

Text-To-Speech is highly dependent on the quality of the voice corpus used to realize the system, and creating an entire voice corpus is expensive.[citation needed] Another disadvantage is that speech synthesis systems do not recognize periods or special characters. Also, ambiguity problems are persistent, as two words written in the same way can have different meanings.[citation needed]

Imitation-based

A block diagram illustrating the imitation-based approach for generating audio deepfakes
The Imitation-based approach diagram

Audio deepfake based on imitation is a way of transforming an original speech from one speaker - the original - so that it sounds spoken like another speaker - the target one.[42] An imitation-based algorithm takes a spoken signal as input and alters it by changing its style, intonation, or prosody, trying to mimic the target voice without changing the linguistic information.[43] This technique is also known as voice conversion.

This method is often confused with the previous Synthetic-based method, as there is no clear separation between the two approaches regarding the generation process. Indeed, both methods modify acoustic-spectral and style characteristics of the speech audio signal, but the Imitation-based usually keeps the input and output text unaltered. This is obtained by changing how this sentence is spoken to match the target speaker's characteristics.[44]

Voices can be imitated in several ways, such as using humans with similar voices that can mimic the original speaker. In recent years, the most popular approach involves the use of particular neural networks called Generative Adversarial Networks (GAN) due to their flexibility as well as high-quality results.[29][42]

Then, the original audio signal is transformed to say a speech in the target audio using an imitation generation method that generates a new speech, shown in the fake one.

Detection methods

The audio deepfake detection task determines whether the given speech audio is real or fake.

Recently, this has become a hot topic in the forensic research community, trying to keep up with the rapid evolution of counterfeiting techniques.

In general, deepfake detection methods can be divided into two categories based on the aspect they leverage to perform the detection task. The first focuses on low-level aspects, looking for artifacts introduced by the generators at the sample level. The second, instead, focus on higher-level features representing more complex aspects as the semantic content of the speech audio recording.

A diagram illustrating the usual framework used to perform the audio deepfake detection task.
A generic audio deepfake detection framework

Many machine learning and deep learning models have been developed using different strategies to detect fake audio. Most of the time, these algorithms follow a three-steps procedure:

  1. Each speech audio recording must be preprocessed and transformed into appropriate audio features;
  2. The computed features are fed into the detection model, which performs the necessary operations, such as the training process, essential to discriminate between real and fake speech audio;
  3. The output is fed into the final module to produce a prediction probability of the Fake class or the Real one. Following the ASVspoof[45] challenge nomenclature, the Fake audio is indicated with the term "Spoof," the Real instead is called "Bonafide."

Over the years, many researchers have shown that machine learning approaches are more accurate than deep learning methods, regardless of the features used.[8] However, the scalability of machine learning methods is not confirmed due to excessive training and manual feature extraction, especially with many audio files. Instead, when deep learning algorithms are used, specific transformations are required on the audio files to ensure that the algorithms can handle them.

There are several open-source implementations of different detection methods,[46][47][48] and usually many research groups release them on a public hosting service like GitHub.

Open challenges and future research direction

The audio deepfake is a very recent field of research. For this reason, there are many possibilities for development and improvement, as well as possible threats that adopting this technology can bring to our daily lives. The most important ones are listed below.

Deepfake generation

Regarding the generation, the most significant aspect is the credibility of the victim, i.e., the perceptual quality of the audio deepfake.

Several metrics determine the level of accuracy of audio deepfake generation, and the most widely used is the MOS (Mean Opinion Score), which is the arithmetic average of user ratings. Usually, the test to be rated involves perceptual evaluation of sentences made by different speech generation algorithms. This index showed that audio generated by algorithms trained on a single speaker has a higher MOS.[44][34][49][50][39]

The sampling rate also plays an essential role in detecting and generating audio deepfakes. Currently, available datasets have a sampling rate of around 16 kHz, significantly reducing speech quality. An increase in the sampling rate could lead to higher quality generation.[37]

Deepfake detection

Focusing on the detection part, one principal weakness affecting recent models is the adopted language.

Most studies focus on detecting audio deepfake in the English language, not paying much attention to the most spoken languages like Chinese and Spanish,[51] as well as Hindi and Arabic.

It is also essential to consider more factors related to different accents that represent the way of pronunciation strictly associated with a particular individual, location, or nation. In other fields of audio, such as speaker recognition, the accent has been found to influence the performance significantly,[52] so it is expected that this feature could affect the models' performance even in this detection task.

In addition, the excessive preprocessing of the audio data has led to a very high and often unsustainable computational cost. For this reason, many researchers have suggested following a Self-Supervised Learning approach,[53] dealing with unlabeled data to work effectively in detection tasks and improving the model's scalability, and, at the same time, decreasing the computational cost.

Training and testing models with real audio data is still an underdeveloped area. Indeed, using audio with real-world background noises can increase the robustness of the fake audio detection models.

In addition, most of the effort is focused on detecting Synthetic-based audio deepfakes, and few studies are analyzing imitation-based due to their intrinsic difficulty in the generation process.[11]

Defense against deepfakes

Over the years, there has been an increase in techniques aimed at defending against malicious actions that audio deepfake could bring, such as identity theft and manipulation of speeches by the nation's governors.

To prevent deepfakes, some suggest using blockchain and other distributed ledger technologies (DLT) to identify the provenance of data and track information.[8][54][55][56]

Extracting and comparing affective cues corresponding to perceived emotions from digital content has also been proposed to combat deepfakes.[57][58][59]

Another critical aspect concerns the mitigation of this problem. It has been suggested that it would be better to keep some proprietary detection tools only for those who need them, such as fact-checkers for journalists.[29] That way, those who create the generation models, perhaps for nefarious purposes, would not know precisely what features facilitate the detection of a deepfake,[29] discouraging possible attackers.

To improve the detection instead, researchers are trying to generalize the process,[60] looking for preprocessing techniques that improve performance and testing different loss functions used for training.[10][61]

Research programs

Numerous research groups worldwide are working to recognize media manipulations; i.e., audio deepfakes but also image and video deepfake. These projects are usually supported by public or private funding and are in close contact with universities and research institutions.

For this purpose, the Defense Advanced Research Projects Agency (DARPA) runs the Semantic Forensics (SemaFor).[62][63] Leveraging some of the research from the Media Forensics (MediFor)[64][65] program, also from DARPA, these semantic detection algorithms will have to determine whether a media object has been generated or manipulated, to automate the analysis of media provenance and uncover the intent behind the falsification of various content.[66][62]

Another research program is the Preserving Media Trustworthiness in the Artificial Intelligence Era (PREMIER)[67] program, funded by the Italian Ministry of Education, University and Research (MIUR) and run by five Italian universities. PREMIER will pursue novel hybrid approaches to obtain forensic detectors that are more interpretable and secure.[68]

DEEP-VOICE[69] is a publicly available dataset intended for research purposes to develop systems to detect when speech has been generated with neural networks through a process called Retrieval-based Voice Conversion (RVC). Preliminary research showed numerous statistically-significant differences between features found in human speech and that which had been generated by Artificial Intelligence algorithms.

Public challenges

In the last few years, numerous challenges have been organized to push this field of audio deepfake research even further.

The most famous world challenge is the ASVspoof,[45] the Automatic Speaker Verification Spoofing and Countermeasures Challenge. This challenge is a bi-annual community-led initiative that aims to promote the consideration of spoofing and the development of countermeasures.[70]

Another recent challenge is the ADD[71]—Audio Deepfake Detection—which considers fake situations in a more real-life scenario.[72]

Also the Voice Conversion Challenge[73] is a bi-annual challenge, created with the need to compare different voice conversion systems and approaches using the same voice data.

See also

References

  1. ^ Smith, Hannah; Mansted, Katherine (April 1, 2020). Weaponised deep fakes: National security and democracy. Vol. 28. Australian Strategic Policy Institute. pp. 11–13. ISSN 2209-9689.{{cite book}}: CS1 maint: date and year (link)
  2. ^ Lyu, Siwei (2020). "Deepfake Detection: Current Challenges and Next Steps". 2020 IEEE International Conference on Multimedia & Expo Workshops (ICMEW). pp. 1–6. arXiv:2003.09234. doi:10.1109/icmew46912.2020.9105991. ISBN 978-1-7281-1485-9. S2CID 214605906. Retrieved 2022-06-29.
  3. ^ a b Diakopoulos, Nicholas; Johnson, Deborah (June 2020). "Anticipating and addressing the ethical implications of deepfakes in the context of elections". New Media & Society. 23 (7) (published 2020-06-05): 2072–2098. doi:10.1177/1461444820925811. ISSN 1461-4448. S2CID 226196422.
  4. ^ Murphy, Margi (20 February 2024). "Deepfake Audio Boom Exploits One Billion-Dollar Startup's AI". Bloomberg.
  5. ^ Chadha, Anupama; Kumar, Vaibhav; Kashyap, Sonu; Gupta, Mayank (2021), Singh, Pradeep Kumar; Wierzchoń, Sławomir T.; Tanwar, Sudeep; Ganzha, Maria (eds.), "Deepfake: An Overview", Proceedings of Second International Conference on Computing, Communications, and Cyber-Security, Lecture Notes in Networks and Systems, vol. 203, Singapore: Springer Singapore, pp. 557–566, doi:10.1007/978-981-16-0733-2_39, ISBN 978-981-16-0732-5, S2CID 236666289, retrieved 2022-06-29
  6. ^ "AI gave Val Kilmer his voice back. But critics worry the technology could be misused". Washington Post. ISSN 0190-8286. Retrieved 2022-06-29.
  7. ^ Etienne, Vanessa (August 19, 2021). "Val Kilmer Gets His Voice Back After Throat Cancer Battle Using AI Technology: Hear the Results". PEOPLE.com. Retrieved 2022-07-01.
  8. ^ a b c d Almutairi, Zaynab; Elgibreen, Hebah (2022-05-04). "A Review of Modern Audio Deepfake Detection Methods: Challenges and Future Directions". Algorithms. 15 (5): 155. doi:10.3390/a15050155. ISSN 1999-4893.
  9. ^ Caramancion, Kevin Matthe (June 2022). "An Exploration of Mis/Disinformation in Audio Format Disseminated in Podcasts: Case Study of Spotify". 2022 IEEE International IOT, Electronics and Mechatronics Conference (IEMTRONICS). pp. 1–6. doi:10.1109/IEMTRONICS55184.2022.9795760. ISBN 978-1-6654-8684-2. S2CID 249903722.
  10. ^ a b Chen, Tianxiang; Kumar, Avrosh; Nagarsheth, Parav; Sivaraman, Ganesh; Khoury, Elie (2020-11-01). "Generalization of Audio Deepfake Detection". The Speaker and Language Recognition Workshop (Odyssey 2020). ISCA: 132–137. doi:10.21437/Odyssey.2020-19. S2CID 219492826.
  11. ^ a b c Ballesteros, Dora M.; Rodriguez-Ortega, Yohanna; Renza, Diego; Arce, Gonzalo (2021-12-01). "Deep4SNet: deep learning for fake speech classification". Expert Systems with Applications. 184: 115465. doi:10.1016/j.eswa.2021.115465. ISSN 0957-4174. S2CID 237659479.
  12. ^ Suwajanakorn, Supasorn; Seitz, Steven M.; Kemelmacher-Shlizerman, Ira (2017-07-20). "Synthesizing Obama: learning lip sync from audio". ACM Transactions on Graphics. 36 (4): 95:1–95:13. doi:10.1145/3072959.3073640. ISSN 0730-0301. S2CID 207586187.
  13. ^ Stupp, Catherine. "Fraudsters Used AI to Mimic CEO's Voice in Unusual Cybercrime Case". WSJ. Retrieved 2024-05-26.
  14. ^ Brewster, Thomas. "Fraudsters Cloned Company Director's Voice In $35 Million Bank Heist, Police Find". Forbes. Retrieved 2022-06-29.
  15. ^ "Generative AI is making voice scams easier to believe". Axios. 13 June 2023. Retrieved 16 June 2023.
  16. ^ Bunn, Amy (15 May 2023). "Artificial Imposters—Cybercriminals Turn to AI Voice Cloning for a New Breed of Scam". McAfee Blog. Retrieved 16 June 2023.
  17. ^ Cox, Joseph (23 February 2023). "How I Broke Into a Bank Account With an AI-Generated Voice". Vice. Retrieved 16 June 2023.
  18. ^ Evershed, Nick; Taylor, Josh (16 March 2023). "AI can fool voice recognition used to verify identity by Centrelink and Australian tax office". The Guardian. Retrieved 16 June 2023.
  19. ^ "Scammers use AI to enhance their family emergency schemes". Consumer Advice. 2023-03-17. Retrieved 2024-05-26.
  20. ^ "Deepfake audio of Sir Keir Starmer released on first day of Labour conference".
  21. ^ Meaker, Morgan. "Slovakia's Election Deepfakes Show AI is a Danger to Democracy". Wired.
  22. ^ "Political consultant behind fake Biden AI robocall faces charges in New Hampshire".
  23. ^ "Political consultant accused of hiring magician to spam voters with Biden deepfake calls". Law & Crime. 2024-03-15. Retrieved 2024-05-23.
  24. ^ David Wright; Brian Fung; Brian Fung (February 6, 2024). "Fake Biden robocall linked to Texas-based companies, New Hampshire attorney general announces". CNN.
  25. ^ Brian Fung (February 8, 2024). "FCC votes to ban scam robocalls that use AI-generated voices". CNN.
  26. ^ "FCC Makes AI-Generated Voices in Robocalls Illegal | Federal Communications Commission". www.fcc.gov. 2024-02-08. Retrieved 2024-05-26.
  27. ^ Kramer, Marcia (2024-02-26). "Steve Kramer explains why he used AI to impersonate President Biden in New Hampshire - CBS New York". www.cbsnews.com. Retrieved 2024-05-23.
  28. ^ "A political consultant faces charges and fines for Biden deepfake robocalls".
  29. ^ a b c d e Khanjani, Zahra; Watson, Gabrielle; Janeja, Vandana P. (2021-11-28). "How Deep Are the Fakes? Focusing on Audio Deepfake: A Survey". arXiv:2111.14203 [cs.SD].
  30. ^ Pradhan, Swadhin; Sun, Wei; Baig, Ghufran; Qiu, Lili (2019-09-09). "Combating Replay Attacks Against Voice Assistants". Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies. 3 (3): 100:1–100:26. doi:10.1145/3351258. S2CID 202159551.
  31. ^ Villalba, Jesus; Lleida, Eduardo (2011). "Preventing replay attacks on speaker verification systems". 2011 Carnahan Conference on Security Technology. pp. 1–8. doi:10.1109/CCST.2011.6095943. ISBN 978-1-4577-0903-6. S2CID 17048213. Retrieved 2022-06-29.
  32. ^ Tom, Francis; Jain, Mohit; Dey, Prasenjit (2018-09-02). "End-To-End Audio Replay Attack Detection Using Deep Convolutional Networks with Attention". Interspeech 2018. ISCA: 681–685. doi:10.21437/Interspeech.2018-2279. S2CID 52187155.
  33. ^ Tan, Xu; Qin, Tao; Soong, Frank; Liu, Tie-Yan (2021-07-23). "A Survey on Neural Speech Synthesis". arXiv:2106.15561 [eess.AS].
  34. ^ a b Oord, Aaron van den; Dieleman, Sander; Zen, Heiga; Simonyan, Karen; Vinyals, Oriol; Graves, Alex; Kalchbrenner, Nal; Senior, Andrew; Kavukcuoglu, Koray (2016-09-19). "WaveNet: A Generative Model for Raw Audio". arXiv:1609.03499 [cs.SD].
  35. ^ Kuchaiev, Oleksii; Li, Jason; Nguyen, Huyen; Hrinchuk, Oleksii; Leary, Ryan; Ginsburg, Boris; Kriman, Samuel; Beliaev, Stanislav; Lavrukhin, Vitaly; Cook, Jack; Castonguay, Patrice (2019-09-13). "NeMo: a toolkit for building AI applications using Neural Modules". arXiv:1909.09577 [cs.LG].
  36. ^ Wang, Yuxuan; Skerry-Ryan, R. J.; Stanton, Daisy; Wu, Yonghui; Weiss, Ron J.; Jaitly, Navdeep; Yang, Zongheng; Xiao, Ying; Chen, Zhifeng; Bengio, Samy; Le, Quoc (2017-04-06). "Tacotron: Towards End-to-End Speech Synthesis". arXiv:1703.10135 [cs.CL].
  37. ^ a b Prenger, Ryan; Valle, Rafael; Catanzaro, Bryan (2018-10-30). "WaveGlow: A Flow-based Generative Network for Speech Synthesis". arXiv:1811.00002 [cs.SD].
  38. ^ Vasquez, Sean; Lewis, Mike (2019-06-04). "MelNet: A Generative Model for Audio in the Frequency Domain". arXiv:1906.01083 [eess.AS].
  39. ^ a b Ping, Wei; Peng, Kainan; Gibiansky, Andrew; Arik, Sercan O.; Kannan, Ajay; Narang, Sharan; Raiman, Jonathan; Miller, John (2018-02-22). "Deep Voice 3: Scaling Text-to-Speech with Convolutional Sequence Learning". arXiv:1710.07654 [cs.SD].
  40. ^ Ren, Yi; Ruan, Yangjun; Tan, Xu; Qin, Tao; Zhao, Sheng; Zhao, Zhou; Liu, Tie-Yan (2019-11-20). "FastSpeech: Fast, Robust and Controllable Text to Speech". arXiv:1905.09263 [cs.CL].
  41. ^ Ning, Yishuang; He, Sheng; Wu, Zhiyong; Xing, Chunxiao; Zhang, Liang-Jie (January 2019). "A Review of Deep Learning Based Speech Synthesis". Applied Sciences. 9 (19): 4050. doi:10.3390/app9194050. ISSN 2076-3417.
  42. ^ a b Rodríguez-Ortega, Yohanna; Ballesteros, Dora María; Renza, Diego (2020). "A Machine Learning Model to Detect Fake Voice". In Florez, Hector; Misra, Sanjay (eds.). Applied Informatics. Communications in Computer and Information Science. Vol. 1277. Cham: Springer International Publishing. pp. 3–13. doi:10.1007/978-3-030-61702-8_1. ISBN 978-3-030-61702-8. S2CID 226283369.
  43. ^ Zhang, Mingyang; Wang, Xin; Fang, Fuming; Li, Haizhou; Yamagishi, Junichi (2019-04-07). "Joint training framework for text-to-speech and voice conversion using multi-source Tacotron and WaveNet". arXiv:1903.12389 [eess.AS].
  44. ^ a b Sercan, Ö Arık; Jitong, Chen; Kainan, Peng; Wei, Ping; Yanqi, Zhou (2018). "Neural Voice Cloning with a Few Samples". Advances in Neural Information Processing Systems (NeurIPS 2018). 31 (published 12 October 2018): 10040–10050. arXiv:1802.06006.
  45. ^ a b "| ASVspoof". www.asvspoof.org. Retrieved 2022-07-01.
  46. ^ resemble-ai/Resemblyzer, Resemble AI, 2022-06-30, retrieved 2022-07-01
  47. ^ mendaxfz (2022-06-28), Synthetic-Voice-Detection, retrieved 2022-07-01
  48. ^ HUA, Guang (2022-06-29), End-to-End Synthetic Speech Detection, retrieved 2022-07-01
  49. ^ Kong, Jungil; Kim, Jaehyeon; Bae, Jaekyoung (2020-10-23). "HiFi-GAN: Generative Adversarial Networks for Efficient and High Fidelity Speech Synthesis". arXiv:2010.05646 [cs.SD].
  50. ^ Kumar, Kundan; Kumar, Rithesh; de Boissiere, Thibault; Gestin, Lucas; Teoh, Wei Zhen; Sotelo, Jose; de Brebisson, Alexandre; Bengio, Yoshua; Courville, Aaron (2019-12-08). "MelGAN: Generative Adversarial Networks for Conditional Waveform Synthesis". arXiv:1910.06711 [eess.AS].
  51. ^ Babbel.com; GmbH, Lesson Nine. "The 10 Most Spoken Languages In The World". Babbel Magazine. Retrieved 2022-06-30.
  52. ^ Najafian, Maryam; Russell, Martin (September 2020). "Automatic accent identification as an analytical tool for accent robust automatic speech recognition". Speech Communication. 122: 44–55. doi:10.1016/j.specom.2020.05.003. S2CID 225778214.
  53. ^ Liu, Xiao; Zhang, Fanjin; Hou, Zhenyu; Mian, Li; Wang, Zhaoyu; Zhang, Jing; Tang, Jie (2021). "Self-supervised Learning: Generative or Contrastive". IEEE Transactions on Knowledge and Data Engineering. 35 (1): 857–876. arXiv:2006.08218. doi:10.1109/TKDE.2021.3090866. ISSN 1558-2191. S2CID 219687051.
  54. ^ Rashid, Md Mamunur; Lee, Suk-Hwan; Kwon, Ki-Ryong (2021). "Blockchain Technology for Combating Deepfake and Protect Video/Image Integrity". Journal of Korea Multimedia Society. 24 (8): 1044–1058. doi:10.9717/kmms.2021.24.8.1044. ISSN 1229-7771.
  55. ^ Fraga-Lamas, Paula; Fernández-Caramés, Tiago M. (2019-10-20). "Fake News, Disinformation, and Deepfakes: Leveraging Distributed Ledger Technologies and Blockchain to Combat Digital Deception and Counterfeit Reality". IT Professional. 22 (2): 53–59. arXiv:1904.05386. doi:10.1109/MITP.2020.2977589.
  56. ^ Ki Chan, Christopher Chun; Kumar, Vimal; Delaney, Steven; Gochoo, Munkhjargal (September 2020). "Combating Deepfakes: Multi-LSTM and Blockchain as Proof of Authenticity for Digital Media". 2020 IEEE / ITU International Conference on Artificial Intelligence for Good (AI4G). pp. 55–62. doi:10.1109/AI4G50087.2020.9311067. ISBN 978-1-7281-7031-2. S2CID 231618774.
  57. ^ Mittal, Trisha; Bhattacharya, Uttaran; Chandra, Rohan; Bera, Aniket; Manocha, Dinesh (2020-10-12), "Emotions Don't Lie: An Audio-Visual Deepfake Detection Method using Affective Cues", Proceedings of the 28th ACM International Conference on Multimedia, New York, NY, USA: Association for Computing Machinery, pp. 2823–2832, doi:10.1145/3394171.3413570, ISBN 978-1-4503-7988-5, S2CID 220935571, retrieved 2022-06-29
  58. ^ Conti, Emanuele; Salvi, Davide; Borrelli, Clara; Hosler, Brian; Bestagini, Paolo; Antonacci, Fabio; Sarti, Augusto; Stamm, Matthew C.; Tubaro, Stefano (2022-05-23). "Deepfake Speech Detection Through Emotion Recognition: A Semantic Approach". ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). Singapore, Singapore: IEEE. pp. 8962–8966. doi:10.1109/ICASSP43922.2022.9747186. hdl:11311/1220518. ISBN 978-1-6654-0540-9. S2CID 249436701.
  59. ^ Hosler, Brian; Salvi, Davide; Murray, Anthony; Antonacci, Fabio; Bestagini, Paolo; Tubaro, Stefano; Stamm, Matthew C. (June 2021). "Do Deepfakes Feel Emotions? A Semantic Approach to Detecting Deepfakes Via Emotional Inconsistencies". 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW). Nashville, TN, USA: IEEE. pp. 1013–1022. doi:10.1109/CVPRW53098.2021.00112. hdl:11311/1183572. ISBN 978-1-6654-4899-4. S2CID 235679849.
  60. ^ Müller, Nicolas M.; Czempin, Pavel; Dieckmann, Franziska; Froghyar, Adam; Böttinger, Konstantin (2022-04-21). "Does Audio Deepfake Detection Generalize?". arXiv:2203.16263 [cs.SD].
  61. ^ Zhang, You; Jiang, Fei; Duan, Zhiyao (2021). "One-Class Learning Towards Synthetic Voice Spoofing Detection". IEEE Signal Processing Letters. 28: 937–941. arXiv:2010.13995. Bibcode:2021ISPL...28..937Z. doi:10.1109/LSP.2021.3076358. ISSN 1558-2361. S2CID 235077416.
  62. ^ a b "SAM.gov". sam.gov. Retrieved 2022-06-29.
  63. ^ "The SemaFor Program". www.darpa.mil. Retrieved 2022-07-01.
  64. ^ "The DARPA MediFor Program". govtribe.com. Retrieved 2022-06-29.
  65. ^ "The MediFor Program". www.darpa.mil. Retrieved 2022-07-01.
  66. ^ "DARPA Announces Research Teams Selected to Semantic Forensics Program". www.darpa.mil. Retrieved 2022-07-01.
  67. ^ "PREMIER". sites.google.com. Retrieved 2022-07-01.
  68. ^ "PREMIER - Project". sites.google.com. Retrieved 2022-06-29.
  69. ^ Bird, Jordan J.; Lotfi, Ahmad (2023). "Real-time Detection of AI-Generated Speech for DeepFake Voice Conversion". arXiv:2308.12734 [cs.SD].
  70. ^ Yamagishi, Junichi; Wang, Xin; Todisco, Massimiliano; Sahidullah, Md; Patino, Jose; Nautsch, Andreas; Liu, Xuechen; Lee, Kong Aik; Kinnunen, Tomi; Evans, Nicholas; Delgado, Héctor (2021-09-01). "ASVspoof 2021: accelerating progress in spoofed and deepfake speech detection". arXiv:2109.00537 [eess.AS].
  71. ^ "Audio Deepfake Detection: ICASSP 2022". IEEE Signal Processing Society. 2021-12-17. Retrieved 2022-07-01.
  72. ^ Yi, Jiangyan; Fu, Ruibo; Tao, Jianhua; Nie, Shuai; Ma, Haoxin; Wang, Chenglong; Wang, Tao; Tian, Zhengkun; Bai, Ye; Fan, Cunhang; Liang, Shan (2022-02-26). "ADD 2022: the First Audio Deep Synthesis Detection Challenge". arXiv:2202.08433 [cs.SD].
  73. ^ "Joint Workshop for the Blizzard Challenge and Voice Conversion Challenge 2020 - SynSIG". www.synsig.org. Archived from the original on 2022-07-02. Retrieved 2022-07-01.

Read other articles:

Hey, Come On!Sampul Hey, Come On!Album studio karya ShinhwaDirilis28 Juni 2001Direkam2001GenreK-PopBahasaKoreanLabelSM EntertainmentProduserLee Soo ManKronologi Shinhwa Only One (2000)Only One 2000 Hey, Come On!(2001) My Choice(2002)My Choice2002 Hey, Come On! adalah album keempat dari Shinhwa memulai debutnya pada peringkat ke-3. Seperti album sebelumnya, Hey, Come On! diterima dengan baik oleh para penggemar dan judul lagu ini dengan cepat menanjak pada tangga lagu. Daftar lagu Informas...

 

У этого топонима есть и другие значения, см. Катино. ДеревняКатино 57°43′29″ с. ш. 41°00′38″ в. д.HGЯO Страна  Россия Субъект Федерации Костромская область Муниципальный район Костромской Сельское поселение Минское История и география Высота центра 122 м Часовой поя

 

Antonio Jiménez kan verwijzen naar: Antonio Jiménez (atleet) Antoni Jiménez, een Catalaans voetballer Bekijk alle artikelen waarvan de titel begint met Antonio Jiménez of met Antonio Jiménez in de titel. Dit is een doorverwijspagina, bedoeld om de verschillen in betekenis of gebruik van Antonio Jiménez inzichtelijk te maken. Op deze pagina staat een uitleg van de verschillende betekenissen van Antonio Jiménez en verwijzingen daarnaartoe. Bent u hier via een pag...

Sara Stridsberg Sara Stridsberg en 2011Información personalNacimiento 29 de agosto de 1972 (51 años)municipio de Solna (Suecia) Nacionalidad SuecaInformación profesionalOcupación Traductora y escritora Cargos ocupados Seat 13 of the Swedish Academy (2016-2018) Miembro de Academia Sueca (2016-2018) [editar datos en Wikidata] Sara Stridsberg (Solna, 1972) escritora y traductora sueca. Desde 2016 es miembro de la Academia Sueca que otorga anualmente el premio Nobel de Lit...

 

As referências deste artigo necessitam de formatação. Por favor, utilize fontes apropriadas contendo título, autor e data para que o verbete permaneça verificável. (Agosto de 2021) Nesta lista estão as 24 cidades mais populosas do Sudão, todas com mais de cem mil habitantes. Ordenada por população e ainda estimativas das mesmas segundo o site [1] em 2010. Ondurmã, a cidade mais populosa do Sudão Cartum, a 2ª cidade mais populosa do Sudão Nº Cidade População 1 Ondurmã 2 568 5...

 

Cet article est une ébauche concernant le jeu vidéo. Vous pouvez partager vos connaissances en l’améliorant (comment ?) (voir l’aide à la rédaction). Nightmare in the DarkDéveloppeur Eleven / GavakingÉditeur SNKDate de sortie Arcade :Neo-Geo MVSGenre Plates-formesMode de jeu Un à deux joueursPlate-forme Arcade :Neo-Geo MVSmodifier - modifier le code - modifier Wikidata Nightmare in the Dark est un jeu vidéo de plates-formes développé par Eleven et Gavaking, et é...

Các thành phần phản biến của tenxơ ứng suất-năng lượng. Thuyết tương đối rộng G μ ν + Λ g μ ν = 8 π G c 4 T μ ν {\displaystyle G_{\mu \nu }+\Lambda g_{\mu \nu }={8\pi G \over c^{4}}T_{\mu \nu }} Dẫn nhập · Lịch sử · Phát biểu toán họcKiểm chứng Khái niệm cơ sởThuyết tương đối hẹpNguyên lý tương đươngTuyến thế giới · Hình h...

 

2020 Japanese filmCrayon Shin-Chan The Movie: Crash! Graffiti Kingdom and Almost Four HeroesTheatrical Release PosterJapanese nameKanjiクレヨンしんちゃん 激突! ラクガキングダムとほぼ四人の勇者TranscriptionsRevised HepburnKureyon Shinchan Gekitotsu Rakugakingudamu to Hobo Shi-Ri no Yūsha Directed byTakahiko KyogokuBased onCrayon Shin-chanby Yoshito UsuiStarring Yumiko Kobayashi Miki Narahashi Toshiyuki Morikawa Satomi Kōrogi Mari Mashiba Teiyū Ichiryūsai Chie Sat...

 

机场轻轨Airport Rail Link (ARL) 概覽營運地點 泰國曼谷服務類型機場鐵路、通勤鐵路主要車站蘇凡納布國際機場站、目甲訕站、披耶泰站主要路線特快線、城市線技術數據列車長度28.6公里正線數目1車站數目8軌距1,435毫米(標準軌)(標準軌)列車编组西門子Desiro 360/2运营信息開通營運2010年8月23日營運者泰國國家鐵路局 路線圖 圖例  SRT  往蘭實 廊曼  SRT  往

Richard Nikolaus di Coudenhove-Kalergi nel 1926 Il conte Richard Nikolaus di Coudenhove-Kalergi (in tedesco: Richard Nikolaus Eijiro Graf Coudenhove-Kalergi; in giapponese: リヒャルト・ニコラウス・栄次郎・クーデンホーフ=カレルギー Rihyaruto Nikorausu Eijirō Kūdenhōfu-Karerugī) (Tokyo, 17 novembre 1894 – Schruns, 27 luglio 1972) è stato un politico e filosofo austriaco, fondatore dell'Unione Paneuropea e primo uomo politico a proporre un progetto di Europa ...

 

American speed skater Mia KilburgKilburg in 2019Personal informationBirth nameMia ManganelloBorn (1989-10-27) October 27, 1989 (age 34)Fort Walton Beach, Florida[1]SportSportRoad bicycle racingSpeed skatingCycling careerPersonal informationHeight1.73 m (5 ft 8 in)Team informationCurrent teamDNA Pro CyclingDisciplineRoadRoleRiderProfessional teams2016–2017Visit Dallas DNA Pro Cycling2020–DNA Pro Cycling[2] Medal record Women's speed skating Rep...

 

Doge's Palace This is a list of buildings and structures in Venice, Italy. A Ala Napoleonica Arsenal Ateneo Veneto B Biblioteca Marciana C Campanile di San Marco Ca' da Mosto Ca' d'Oro Ca' Farsetti Ca' Foscari Ca' Loredan Ca' Pesaro Ca' Rezzonico Ca' Tron Ca' Vendramin Calergi Campanile di San Marco Campo San Polo Campo San Samuele Campo San Zanipolo Campo Santa Margherita Campo Sant'Angelo Corte del Milion D Dogana da Mar Doge's Palace F Fabbriche Nuove di Rialto (Erberia) Fabbriche Vecchie ...

1982 film by Robert Benton Still of the NightTheatrical release posterDirected byRobert BentonWritten byRobert BentonDavid Newman (story)Produced byArlene DonovanStarringRoy ScheiderMeryl StreepJessica TandyJosef SommerCinematographyNéstor AlmendrosEdited byGerald B. GreenbergBill PankowMusic byJohn KanderProductioncompanyUnited ArtistsDistributed byMetro-Goldwyn-MayerRelease dateNovember 19, 1982 (1982-11-19)Running time93 minutesCountryUnited StatesLanguageEnglishBudget$10 m...

 

Sri Lankan politician Hon.M. Velu KumarMPவேலு குமார் වේලු කුමාර්Member of the Parliament of Sri LankaIncumbentAssumed office August 2020ConstituencyKandy DistrictIn officeAugust 2015 – March 2020ConstituencyKandy DistrictMember of the Central Provincial CouncilIn office2013–2015ConstituencyKandy District Personal detailsBorn (1973-01-16) 16 January 1973 (age 50)Political partyDemocratic People's FrontOther politicalaffiliationsSama...

 

Bhutanese Lama (1594–1651) Not to be confused with Ngawang Namgyal (Rinpungpa). Zhabdrung in a seventeenth-century painting Ngawang Namgyal (later granted the honorific Zhabdrung Rinpoche, approximately at whose feet one submits) (Tibetan: ཞབས་དྲུང་ངག་དབང་རྣམ་རྒྱལ་, Wylie: zhabs drung ngag dbang rnam rgyal; alternate spellings include Zhabdrung Ngawang Namgyel; 1594–1651) and known colloquially as The Bearded Lama, was a Tibetan Buddhist l...

División Mayor del Fútbol Colombiano Acrónimo DimayorTipo Organización deportivaFundación 26 de junio de 1948 (75 años)Sede central Bogotá (Colombia)Deporte FútbolPresidente Fernando Jaramillo GiraldoSecretaria María del Pilar Avella[1]​Gerente Deportivo Ricardo 'Gato' Pérez[2]​Sitio web Dimayor.com.co[editar datos en Wikidata] La División Mayor del Fútbol Colombiano (Dimayor) es una entidad deportiva dependiente de la Federación Colombiana de Fútbol q...

 

For the song by Saul Williams, see Amethyst Rock Star. This article uses bare URLs, which are uninformative and vulnerable to link rot. Please consider converting them to full citations to ensure the article remains verifiable and maintains a consistent citation style. Several templates and tools are available to assist in formatting, such as reFill (documentation) and Citation bot (documentation). (September 2022) (Learn how and when to remove this template message) Cover of the first editio...

 

France national team. Group H was one of eight groups of the preliminary round of the 2023 FIBA Basketball World Cup. It took place from 25 to 29 August 2023 and consisted of Canada, Latvia, Lebanon, and France.[1][2] Each team played each other once, for a total of three games per team, with all games played at the Indonesia Arena, Jakarta, Indonesia. The top two teams advanced to the second round and the bottom two teams qualified for the classification rounds.[3] Te...

2000 video game This article needs additional citations for verification. Please help improve this article by adding citations to reliable sources. Unsourced material may be challenged and removed.Find sources: Dark Reign 2 – news · newspapers · books · scholar · JSTOR (August 2012) (Learn how and when to remove this template message) 2000 video gameDark Reign 2Developer(s)Pandemic StudiosPublisher(s)ActivisionDirector(s)Greg BorrudDesigner(s)Christoph...

 

This article needs additional citations for verification. Please help improve this article by adding citations to reliable sources. Unsourced material may be challenged and removed.Find sources: Dogpatch, San Francisco – news · newspapers · books · scholar · JSTOR (December 2009) (Learn how and when to remove this template message) 37°45′38″N 122°23′28″W / 37.76060°N 122.39107°W / 37.76060; -122.39107 Neighborhood in...

 

Strategi Solo vs Squad di Free Fire: Cara Menang Mudah!