The World of Generative AI: Deepfakes and Large Language Models (2024)

The World of Generative AI: Deepfakes and Large Language Models (1) Alakananda Mitra
Nebraska Water Center
University of Nebraska-Lincoln, USA
amitra6@unl.edu
\AndThe World of Generative AI: Deepfakes and Large Language Models (2) Saraju P. Mohanty
Dept. of Computer Science and Engineering
University of North Texas, USA
saraju.mohanty@unt.edu
\AndThe World of Generative AI: Deepfakes and Large Language Models (3) Elias Kougianos
Dept. of Electrical Engineering
University of North Texas, USA
elias.kougianos@unt.edu

Abstract

We live in the era of Generative Artificial Intelligence (GenAI). Deepfakes and Large Language Models (LLMs) are two examples of GenAI. Deepfakes, in particular, pose an alarming threat to society as they are capable of spreading misinformation and changing the truth. LLMs are powerful language models that generate general-purpose language. However due to its generative aspect, it can also be a risk for people if used with ill intentions. The ethical use of these technologies is a big concern. This short article tries to find out the interrelationship between them.

Keywords Deepfake, Generative AI, Large Language Models (LLMs)

1 Introduction

The latest development in artificial intelligence (AI), chatbots, the product of generative AI, has captivated the public in the last two years. But it similarly poses an unprecedented challenge and can have potentially unwanted effects on our lives. OpenAI released the chatbot ChatGPT on November 30, 2022. The overwhelming response of the public towards ChatGPT usage pushed Google to release Bard, ChatGPT’s rival, and Microsoft to release AI-powered Bing. But the recent GPT-4 topped the list as it has more capabilities than any other existing chatbot. Being LLM-based, these chatbots create synthetic media with the intention of creating better content, enhanced quality, or professional voices. The capabilities of such chatbots raise questions on the ethical use of AI.

In the meantime, deepfakes, which are high-quality AI-generated fake videos, have been circulating online. Synthetically generated deepfake videos have exceeded acceptable limits in terms of reality distortion. This disruptive technological development significantly impacts the truth [1]. Due to the easy accessibility of generative AI models, the probability of misuse of this technology aggressively increases [2]. Though deepfake technology came into this game in 2017, large language models like ChatGPT, Bing, Bard, GPT-4, etc. add a whole new dimension to this scenario. In particular, political figures are targeted in deepfake videos, contributing to the erosion of media credibility.

2024 is going to be an eventful year. A total of 78 countries are having major elections in 2024, according to the nonprofit organization for improving social media, the Integrity Institute [3]. Some examples are the U.S. presidential election, and national elections in South Africa, and the biggest democracy in the world, India. The extraordinary advent of generative AI, especially deepfake, has the potential of making these elections highly controversial [3].

Deepfake videos are also showing up in Hollywood. Deceased actors can talk and even act in full-length feature films from beyond the grave, raising concerns over who holds the intellectual rights to a deceased person’s semblance [4]. The most disturbing part is that anybody with basic computer skills can generate deepfakes to a certain extent. However, generating very high-quality deepfakes needs advanced technology like generative adversarial networks (GANs) [5].

This article discusses how generative AI chatbots powered by these large language models can impact deepfake technology (Fig. 1). In the succeeding sections, we described what a deepfake is, how it is created, what a large language model is, and finally how the AI chatbots can change deepfake technology.

The World of Generative AI: Deepfakes and Large Language Models (4)

The rest of this article is organized in the following manner:The concept of a deepfake is introduced in Section 2. Details of the creation of deepfake videos are presented in Section 3. Section 4 introduces Large Language Models (LLMs). The roles of ChatGPT for the creation of deepfakes is the scope of Section 5.Section 6 outlines some thoughts on efforts to reduce deepfakes.Conclusions are discussed in Section 7.

2 What are Deepfakes?

The term deepfake originated in 2017 by a Reddit user who used deep learning technologies (“deep”) to create fake videos (“fake”) [6]. Deepfake technology allows people to swap faces in videos and images, change voices, and alter texts in documents. Fig. 2 shows different types of deepfakes. They quickly proliferated and have now reached a “mature” state. They take digital forgery to a new level. Though graphic-based and content-changing alterations are still common in multimedia forgeries, deep learning technologies have made the forgery much faster, simpler, and more realistic. Earlier generations of deepfakes used to have unnatural movements, artificial facial reactions, sudden shifts in audio quality, mismatches in color and lighting, etc. However, the new generation of deepfakes is sophisticated, and nobody can recognize them with bare eyes. Special deep learning-based methods are required for that.

The World of Generative AI: Deepfakes and Large Language Models (5)

This technology has spread disinformation, rumors, and propaganda, provoked political strife, blackmailed individuals, and threatened democracy. The capabilities of this technology to manipulate reality have gone too far. This paradigm-shifting technological revolution is altering the truth and creates a zero-trust society insidiously.

3 How are Deepfake Videos Created?

Deepfakes are generated using deep learning models and supervised learning. Different types of deep learning models and different training data are used for the four types of deepfakes. Two key generative neural network architectures for creating deepfake videos are variational autoencoders [7] and Generative Adversarial Networks (GANs) [8]. Facial recognition commonly employs a variational autoencoder to learn latent representations. Autoencoders compress and encode input data (image) to a lower-dimensional latent space, then reconstruct it to provide output data (image). The latent representation comprises all these essential facts that the autoencoder uses to create a more adaptable model that permits “face swap” using common features. Figs. 2(a) and 2(b) show the training of the autoencoder and the generation of deepfakes using the autoencoder.

The World of Generative AI: Deepfakes and Large Language Models (6)
The World of Generative AI: Deepfakes and Large Language Models (7)
The World of Generative AI: Deepfakes and Large Language Models (8)

Deepfakes use GANs for realism. A “generator” creates new images using the latent representation of the source image, and a “discriminator” assesses their quality (Fig. 2(c)). If the generator’s image fails the discriminator, it is ordered to create more until one “fools”the discriminator.

Although it can be hard to make quality deepfake content on a standard personal computer, a plethora of online tools are available that can aid users in this endeavor, e.g., FaceApp, TalkingFaces, DeepBrain, Reface, FaceSwap, MorphMe, DeepSwap, DeepFaceLab, etc. ChatGPT integrates with DeepBrain to create real-time conversational deepfakes with custom virtual avatars.

4 What are Large Language Models?

Large language models (LLMs) are a type of deep learning model that can perform a variety of natural language processing tasks. In 1967, the world’s first chatbot, Eliza, introduced the initial concept of LLMs. Almost three decades later, Long Short Term Memory networks (LSTMs) [9] were introduced, which can generate short phrases.

In 2017, the launch of Transfomer (the large language model, “Attention Is All You Need”) by Google [10] revolutionized the generative text industry. In 2019, LLMs were able to generate contextual responses and tasks with the launch of Google’s Bidirectional Encoder Representations from Transformer (BERT).

In a similar time frame, OpenAI was also working on their LLMs. OpenAI’s ChatGPT model began its journey with the release of the GPT-1 model in 2018, which was not made available to the public. An improved version, GPT-2, was published in November 2019. However, GPT-3 (launched in November 2022) [11] was the “pivotal moment” when people were able to interact with the power of generative AI with the launch of OpenAI’s chatbot, ChatGPT, or Chat Generative Pre-trained Transformer (GPT).

ChatGPT is a chatbot based on a large-language model. It has prompted a lot of questions due to its technological capabilities which resemble those of humans. ChatGPT is an application of generative AI that uses user input to generate responses in images, videos, and texts. It is even capable of using voice to engage in a lucid conversation. People use it for a variety of purposes, such as:

  • Generating software codes and checking them for bugs.

  • Drafting emails.

  • Summarizing texts.

  • Writing social network posts.

  • Scripting conversations.

  • In general, creating any type of textual content.

The original ChatGPT used the GPT-3 model. The recent ChatGPT uses an upgraded version, GPT-3.5. Up to this generation, the LLM accepts only textual input. However, ChatGPT Plus (paid version) and the enterprise version, which use the new GPT-4 LLM [12], can accept both text and images as input.

On the other hand, Google’s Bard now uses its own model, PaLM-2 [13]; earlier, it used to have LaMDA [14]. The specialty of PaLM-2 is that it can summarize a page from a URL. Bing uses the Prometheus model, an ensemble of GPT-4, Bin Orchestrator, and the original Bing search.

Another strong contender in this field is the new LLaMa-2 model [15] by Meta. The size of the LLaMa models (LLaMA-2 and LLaMA) is much smaller compared to the other LLMs. As a result, training LLaMa is much easier and more cost-effective as it requires less computing power. There are other alternatives for other text generators on the market, such as AI-Writer, Deep Write, Jasper, CopySmith, etc., and other coding alternatives like AlphaCode, Amazon CodeWhisper, CodeWp, Seek, Tabine, etc.

Apple, Inc. is not behind in this game. In October 2023, Apple and Cornell University quietly announced their open-source multi-modal LLM (MLLM), Ferret [16]. It uses parts of images as the input. Ferret analyzes a region of a picture, detects the elements that potentially answer a question, and creates a bounding box around them. It can then utilize the specified elements to query and respond traditionally.If a user highlights an animal in a larger image and asks the LLM what it is, it will identify the species. It can then employ other graphic elements to provide further reactions or context for the animal’s actions.Fig. 4 shows different LLMs. Fig. 3(a) shows the basic Transformer architecture. Bert is built with the encoder part of the Transformer while GPT 4 is built with the decoder part of it. Fig. 3(b) shows the architecture of Llama 2. and Fig. 3(c) that of Ferret.

The World of Generative AI: Deepfakes and Large Language Models (9)
The World of Generative AI: Deepfakes and Large Language Models (10)
The World of Generative AI: Deepfakes and Large Language Models (11)

However, LLMs can create a new level of misinformation, e.g., academic dishonesty, wrong information creation, assisting deepfakes, or creating something from nothing [2].

5 Role of ChatGPT in Deepfake Creation

Previously, if you wanted to make a convincing dialogue for a deepfake video, you had to compose the lines yourself. These days, it’s straightforward to let AI handle everything. You only need to write an outline of the content to obtain a credible dialogue from ChatGPT, Microsoft’s Bing chatbot, or any other text generator. They save time and effort by writing the dialogue. Anyone can create dialogue in any language and voice. So, it is not a barrier for anybody to create a deepfake video in another language. It spreads the boundaries of a crime, it can create more engaging dialogue and is cost-effective, as there is no need for professional help.

With OpenAI’s new text-writing system, users can now generate lifelike talking heads by simply inputting a prompt and selecting from dozens of avatars and accents created by actors. Various startups like Hour One, Synthesia, and Uneeq are integrating ChatGPT, large language models (LLMs), generative AIs, and other AI with deep learning tools into their platforms to create high-quality synthetic videos without any videographers or high-end video cameras. So, the AI tools create videos that do not exist in real life. In other words, these tools are allowing us to create anything from something else.

Meta is emphasizing open, safe, and responsible generative AI. Purple Llama is the “umbrella” project for this effort. They released a set of “cyber security safety evaluation benchmarks” CyberSec Eval [17] and a “safety classifier for input and output filtering” Llama Guard [18]. “Purple” suggests a collaborative approach for attack and defensive postures.The discussion will not be complete without mentioning Google’s multi-modal model, Gemini [19]. It can transform any type or mode of input into any type of output. Google claims that it outperforms human experts in Massive Multitask Language Understanding (MMLU) and surpasses all its predecessors from any company. During the short gap between writing and publication of the article, a few more LLMs might emerge.

6 Harnessing Efforts

Political parties, as well as the global community, are keenly cognizant of the tremendous influence that social media exert over the public. World leaders have already made numerous attempts to harness their power. Presently, in light of the emergence of generative AI as a formidable novel instrument, technology behemoths are honing their efforts to thwart its potential propaganda capabilities and voter manipulation.Alphabet, the parent company of Google, has confirmed that it will restrict the ways in which its chatbot Bard and search-generative experience respond to queries concerning the next election [20], adding to an already delicate situation. Meanwhile, Meta, the parent company of Facebook, has prohibited campaigns and other regulated industries from utilizing its generative AI tools to create AI-generated political advertisem*nts. Furthermore, Meta has implemented a policy mandating the disclosure of artificial intelligence (AI) or digital modifications made to political, social, and election-related advertisem*nts on Instagram and Facebook. Meta has taken this action to increase transparency and combat misinformation.Interestingly, Elon Musk’s X reversed its global political ad restriction [21]. U.S. candidates and parties can advertise on the site. Governments worldwide are restricting AI-powered political ads. The EU mandates identifying such commercials, including sponsors, costs, and targeted elections.

These efforts are mostly focused on political elections; however, celebrities and common people are also victims of fraud, scams, and blackmail. Deepfake p*rnographic content targets mostly women to harass or defame them [22]. To fight against those situations, academic researchers across the globe are also proposing various methods to address this problem. Unfortunately, these methods alone are insufficient to effectively combat the threat of deepfakes. These methods are mostly based on supervised learning, so when a specific kind of deepfake is detected, the detection method comes in.

7 Conclusions

Deepfake and AI chatbots are both the products of generative AI. Both individually threaten the cybersecurity industry, and together they worsen the situation significantly. Both technologies create something that does not exist in reality. Deepfakes are AI-generated fake images or videos. Chatbots make the audio of those videos fluent and lucid. Because of such chatbots, deepfakes can progress to the next stage of perfection.

Although both technologies have several advantages that can help us advance in many sectors, it is unfortunate that, because of their human-imitating power, they are drawing the attention of the global hacker community. As the 2024 U.S. election approaches, the chances of such AI-generated cyber crimes are increasing significantly.

Technology is always one step ahead, and then rules and regulations come. Therefore, it is necessary to establish more stringent laws and regulations related to data privacy and security. Recently, a bill (HF1370) was passed in the Minnesota House stating that nonconsensual sharing of deepfake p*rnography and political misinformation is a criminal offense. The accused can be charged up to five years of imprisonment and $10,000currency-dollar10000\$10,000$ 10 , 000 in fines. In an example of how politicians do not understand technology, the Senate stalled the bill. India’s current prime minister, Mr. Narendra Modi, expressed serious concerns about this technology recently. The IT ministry went one step forward and called this technology a “new threat to democracy.” The Indian government has announced a new watchdog website to limit the spread of deepfakes. People can complain about the person or group of people who are making deepfakes. The Indian government is expecting 100%percent100100\%100 % compliance from the social media companies in fighting deepfakes. Sony is developing an in-camera authenticity technology to tackle deepfakes by introducing a digital signature when a picture is taken through Sony cameras, thereby identifying real images. However, smartphone cameras capture most of the images today instead of professional cameras. So, these efforts are not enough. We need more research on deepfake prevention technology and methods that can detect deepfakes in real time and on mobile devices. We believe that more extensive use of digital watermarking, the use of distributed ledgers, and letting people access the metadata of any published videos can stop the spreading of rumors and lies.

References

  • [1]A.Mitra, S.P. Mohanty, P.Corcoran, and E.Kougianos, “A machine learningbased approach for deepfake detection in social media through key video frameextraction,” SN Computer Science, vol.2, pp. 1–18, 2021.
  • [2]M.R. Shoaib, Z.Wang, M.T. Ahvanooey, and J.Zhao, “Deepfakes,misinformation, and disinformation in the era of frontier ai, generative ai,and large ai models,” in 2023 International Conference on Computer andApplications (ICCA), 2023, pp. 1–7. [Online]. Available:https://api.semanticscholar.org/CorpusID:265499108
  • [3]E.Strickland, “Content credentials will fight deepfakes in the 2024elections,” IEEE Spectrum, 2023. [Online]. Available:https://spectrum.ieee.org/deepfakes-election
  • [4]E.Harrison, “Itv’s deep fake neighbour wars trailer is the most disturbingthing you’ll watch today,” Independent, November 16, 2022. [Online].Available:https://www.independent.co.uk/arts-entertainment/tv/news/deep-fake-neighbour-wars-itv-b2226293.html
  • [5]A.Mitra, S.P. Mohanty, P.Corcoran, and E.Kougianos, “Easydeep: An iotfriendly robust detection method for gan generated deepfake images in socialmedia,” in Proc. of the 4th IFIP International Internet of ThingsConference.Springer, 2021, pp.217–236.
  • [6]A.Mitra, S.P. Mohanty, P.Corcoran, and E.Kougianos, “Anovel machine learning based method for deepfake video detection in socialmedia,” in Proc. of IEEE International Symposium on Smart ElectronicSystems (iSES)(Formerly iNiS).IEEE,2020, pp. 91–96.
  • [7]D.P. Kingma, M.Welling etal., “An introduction to variationalautoencoders,” Foundations and Trends® in MachineLearning, vol.12, no.4, pp. 307–392, 2019.
  • [8]I.Goodfellow, J.Pouget-Abadie, M.Mirza, B.Xu, D.Warde-Farley, S.Ozair,A.Courville, and Y.Bengio, “Generative adversarial networks,”Communications of the ACM, vol.63, no.11, pp. 139–144, 2020.
  • [9]S.Hochreiter and J.Schmidhuber, “Long short-term memory,” NeuralComputation, vol.9, no.8, pp. 1735–1780, 1997.
  • [10]A.Vaswani, N.Shazeer, N.Parmar, J.Uszkoreit, L.Jones, A.N. Gomez,Ł.Kaiser, and I.Polosukhin, “Attention is all you need,”Advances in neural information processing systems, vol.30, 2017.
  • [11]T.Brown, B.Mann, N.Ryder, M.Subbiah, J.D. Kaplan, P.Dhariwal,A.Neelakantan, P.Shyam, G.Sastry, A.Askell etal., “Languagemodels are few-shot learners,” Advances in neural informationprocessing systems, vol.33, pp. 1877–1901, 2020.
  • [12]J.Achiam, S.Adler, S.Agarwal, L.Ahmad, I.Akkaya, F.L. Aleman, D.Almeida,J.Altenschmidt, S.Altman, S.Anadkat etal., “Gpt-4 technicalreport,” arXiv preprint arXiv:2303.08774, 2023.
  • [13]R.Anil, A.M. Dai, O.Firat, M.Johnson, D.Lepikhin, A.Passos, S.Shakeri,E.Taropa, P.Bailey, Z.Chen etal., “Palm 2 technical report,”arXiv preprint arXiv:2305.10403, 2023.
  • [14]R.Thoppilan, D.DeFreitas, J.Hall, N.Shazeer, A.Kulshreshtha, H.-T. Cheng,A.Jin, T.Bos, L.Baker, Y.Du etal., “Lamda: Language models fordialog applications,” arXiv preprint arXiv:2201.08239, 2022.
  • [15]H.Touvron and etal., “Llama 2: Open foundation and fine-tuned chat models,”arXiv, 2307.09288, 2023.
  • [16]H.You, H.Zhang, Z.Gan, X.Du, B.Zhang, Z.Wang, L.Cao, S.-F. Chang, andY.Yang, “Ferret: Refer and ground anything anywhere at any granularity,”arXiv, 2310.07704, 2023.
  • [17]M.Bhatt, S.Chennabasappa, C.Nikolaidis, S.Wan, I.Evtimov, D.Gabi,D.Song, F.Ahmad, C.Aschermann, L.Fontana etal., “Purple llamacyberseceval: A secure coding benchmark for language models,” arXivpreprint arXiv:2312.04724, 2023.
  • [18]H.Inan, K.Upasani, J.Chi, R.Rungta, K.Iyer, Y.Mao, M.Tontchev, Q.Hu,B.Fuller, D.Testuggine etal., “Llama guard: Llm-based input-outputsafeguard for human-ai conversations,” arXiv preprintarXiv:2312.06674, 2023.
  • [19]G.Team, R.Anil, S.Borgeaud, Y.Wu, J.-B. Alayrac, J.Yu, R.Soricut,J.Schalkwyk, A.M. Dai, A.Hauth etal., “Gemini: a family of highlycapable multimodal models,” arXiv preprint arXiv:2312.11805, 2023.
  • [20]Reuters, “Alphabet to limit election queries bard and ai-based search cananswer,”https://www.reuters.com/technology/alphabet-limit-election-queries-bard-ai-based-search-can-answer-2023-12-19/,December 19, 2023.
  • [21]E.Woollacott, “X lifts ban on political ads,”https://www.forbes.com/sites/emmawoollacott/2023/08/30/x-lifts-ban-on-political-ads/?sh=1549fbe2222c,August 30, 2023.
  • [22]K.Hao, “Deepfake p*rn is ruining women’s lives. now the law may finally banit,” Technology Review, vol.12, p. 2021, 2021.

Authors

The World of Generative AI: Deepfakes and Large Language Models (12)

Alakananda Mitra is a Research Assistant Professor at the Nebraska Water Center at the Institute of Agriculture and natural Resources, University of Nebraska-Lincoln, Lincoln, NE, USA. She started working as a Visiting Computer Scientist at the USDA-ARS Adaptive Cropping Systems Laboratory at the Beltsville Agricultural Research Center, Beltsville, MD in March 2023. She earned her Ph.D. degree in computer science and engineering from University of North Texas, Denton, TX, USA in 2022. She earned her Bachelor of Science degree (Hons.) in physics from the Presidency College, University of Calcutta, in 2001, and her B. Tech. and M.Tech. degrees in radiophysics and electronics from the Institute ofRadiophysics and Electronics, University of Calcutta, in 2004 and 2006,respectively. Her research interests include application-specific AI/ML/deep learning technologies, computer vision, and edge AI, especially in smart agriculture and multi-media forensics. Currently, she is working on AI-based crop models, tinyMLdevices for plant disease detection, and application of federated learning in smart agriculture. She is also working on a project for developing crop and soil simulation models, graphical user interfaces, databases, and other suitable agro-climatology modeling tools. Dr. Mitra received numerous academic awards, honors, and travel grants throughout her career. During her doctoral research, she received the Outstanding Early-Stage Doctoral Student Award. She also received several Best Paper awards and has one US patent (pending) and one US provisional patent. She is a member of IEEE and American Geophysical Union.

The World of Generative AI: Deepfakes and Large Language Models (13)

Saraju P. Mohanty received the bachelor’s degree (Honors) in electrical engineering from the Orissa University of Agriculture and Technology, Bhubaneswar, in 1995, the master’s degree in Systems Science and Automation from the Indian Institute of Science, Bengaluru, in 1999, and the Ph.D. degree in Computer Science and Engineering from the University of South Florida, Tampa, in 2003. He is a Professor with the University of North Texas. His research is in “Smart Electronic Systems” which has been funded by National Science Foundations (NSF), Semiconductor Research Corporation (SRC), U.S. Air Force, IUSSTF, and Mission Innovation. He has authored 500 research articles, 5 books, and 10 granted and pending patents. His Google Scholar h-index is 57 and i10-index is 242 with 13,000 citations. He is regarded as a visionary researcher on Smart Cities technology in which his research deals with security and energy aware, and AI/ML-integrated smart components. He introduced the Secure Digital Camera (SDC) in 2004 with built-in security features designed using Hardware Assisted Security (HAS) or Security by Design (SbD) principle. He is widely credited as the designer for the first digital watermarking chip in 2004 and first the low-power digital watermarking chip in 2006. He is a recipient of 18 best paper awards, Fulbright Specialist Award in 2021, IEEE Consumer Electronics Society Outstanding Service Award in 2020, the IEEE-CS-TCVLSI Distinguished Leadership Award in 2018, and the PROSE Award for Best Textbook in Physical Sciences and Mathematics category in 2016. He has delivered 24 keynotes and served on 14 panels at various International Conferences. He has been serving on the editorial board of several peer-reviewed international transactions/journals, including IEEE Transactions on Big Data (TBD), IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems (TCAD), IEEE Transactions on Consumer Electronics (TCE), and ACM Journal on Emerging Technologies in Computing Systems (JETC). He has been the Editor-in-Chief (EiC) of the IEEE Consumer Electronics Magazine (MCE) during 2016-2021. He served as the Chair of Technical Committee on Very Large Scale Integration (TCVLSI), IEEE Computer Society (IEEE-CS) during 2014-2018 and on the Board of Governors of the IEEE Consumer Electronics Society during 2019-2021. He serves on the steering, organizing, and program committees of several international conferences. He is the steering committee chair/vice-chair for the IEEE International Symposium on Smart Electronic Systems (IEEE-iSES), the IEEE-CS Symposium on VLSI (ISVLSI), and the OITS International Conference on Information Technology (OCIT). He has mentored 3 post-doctoral researchers, and supervised 15 Ph.D. dissertations, 26 M.S. theses, and 21 undergraduate projects.

The World of Generative AI: Deepfakes and Large Language Models (14)

Elias Kougianos received a BSEE from the University of Patras, Greece in 1985 and an MSEE in 1987, an MS in Physics in 1988 and a Ph.D. in EE in 1997, all from Louisiana State University. From 1988 through 1998 he was with Texas Instruments, Inc., in Houston and Dallas, TX. In 1998 he joined Avant! Corp. (now Synopsys) in Phoenix, AZ as a Senior Applications engineer and in 2000 he joined Cadence Design Systems, Inc., in Dallas, TX as a Senior Architect in Analog/Mixed-Signal Custom IC design. He has been at UNT since 2004. He is a Professor in the Department of Electrical Engineering, at the University of North Texas (UNT), Denton, TX. His research interests are in the area of Analog/Mixed-Signal/RF IC design and simulation and in the development of VLSI architectures for multimedia applications. He is an author of over 200 peer-reviewed journal and conference publications.

The World of Generative AI: Deepfakes and Large Language Models (2024)

FAQs

What is generative AI and large language models? ›

A large language model can decipher the nuances of language, while generative AI can create accurate translations and localized versions of the content.

Are deepfakes considered generative AI? ›

This poses a significant threat to international security, as generative AI can also be used to create deepfakes of world leaders giving speeches that are not genuine, or fabricated satellite images that include objects that do not really exist (Defense One 2019).

What is the main goal of generative AI AI DL ML GenAI? ›

#### Final Answer The main goal of Generative AI is to generate new, realistic data that resembles the training data, used for various applications such as image generation, text creation, and data augmentation.

What is the difference between AI and generative AI? ›

While machine learning excels at analyzing data and making predictions, generative AI pushes the boundaries of creativity by generating new and innovative content.

What are generative AI examples? ›

Generative AI or generative artificial intelligence refers to the use of AI to create new content, like text, images, music, audio, and videos. Generative AI is powered by foundation models (large AI models) that can multi-task and perform out-of-the-box tasks, including summarization, Q&A, classification, and more.

Is ChatGPT NLP or generative AI? ›

ChatGPT is a form of generative AI that helps with content creation and information retrieval. In other words, generative AI is a broad field of artificial intelligence, while ChatGPT is a specific implementation of it. Working with experts can allow you to unlock the potential of generative AI tools.

Is Deep fake AI illegal? ›

Deepfake's legal standing is tricky and changing. This harmful content not intrinsically outlawed but can breach the law. This is especially true if they infringe on privacy, intellectual property, or involve defamation, harassment, or fraud. The tricky part is that current laws weren't made with deepfakes in mind.

Which are the five types of deepfakes? ›

​​5 Types Of Deep Fakes You Should Be Aware Of​
  • ​​Textual Deep Fakes​ ...
  • ​​Deep Fake Video​ ...
  • ​​Deep fake Audio​ ...
  • ​​Deep Fakes on Social Media​ ...
  • ​​Real-time or Live Deepfakes​ ...
  • ​​Deep Fake Frauds in India​ ...
  • ​​How To Avoid Deep Fakes​
Jan 13, 2024

How can deepfakes be stopped? ›

The liveness detection step must use advanced AI and machine learning models that have been trained on a variety of real-world data. This step is absolutely crucial for stopping fraudsters who can now easily create very lifelike deepfakes that easily fool older identity verification technology.

What is an example of a hallucination when using generative AI? ›

AI hallucinations can take many different forms. Some common examples include: Incorrect predictions: An AI model may predict that an event will occur when it is unlikely to happen. For example, an AI model that is used to predict the weather may predict that it will rain tomorrow when there is no rain in the forecast.

What is one thing current generative AI applications cannot do? ›

Generative AI can't generate new ideas or solutions

One of the key limitations of AI is its inability to generate new ideas or solutions.

Which industry is likely to benefit the most from generative AI? ›

The healthcare industry stands to benefit greatly from generative AI. One of the key areas where generative AI can make a significant impact is in medical imaging.

What is the downside of generative AI? ›

Data privacy and security

One of the foremost challenges related to generative AI is the handling of sensitive data. As generative models rely on data to generate new content, there is a risk of this data including sensitive or proprietary information.

Which is the best generative AI tool? ›

Among the best generative AI tools for images, DALL-E 2 is OpenAI's recent version for image and art generation. DALL-E 2 generates better and more photorealistic images when compared to DALL-E. DALL-E 2 appropriately goes by user requests.

Are large language models the same as generative AI? ›

Generative AI broadly refers to technologies capable of creating new, diverse outputs such as images, music, and synthetic data. For example, a computer designing artwork or simulating complex medical scenarios. Large Language Models (LLMs), a subset of generative AI, focus on producing text that mirrors human writing.

What is a large language model in AI? ›

A large language model (LLM) is a type of artificial intelligence (AI) program that can recognize and generate text, among other tasks. LLMs are trained on huge sets of data — hence the name "large." LLMs are built on machine learning: specifically, a type of neural network called a transformer model.

What is the difference between AI ML vs DL vs generative AI? ›

AI serves as the broad, encompassing concept, while ML learns patterns from data, DL leverages deep neural networks for intricate pattern recognition, and Generative AI creates new content.

Is ChatGPT a large language model? ›

OpenAI's large language models, including the models that power ChatGPT, are developed using three primary sources of information: (1) information that is publicly available on the internet, (2) information that we license from third parties, and (3) information that our users or our human trainers provide.

What is the difference between GPT and LLM? ›

While GPT is a particular application of LLM, the " LLM " concept encompasses a wider selection of models designed for different tasks in natural language processing. Training Objectives: The objectives for training of GPT and LLM can differ based on the particular model and the purpose for which it is intended.

Top Articles
Latest Posts
Article information

Author: Moshe Kshlerin

Last Updated:

Views: 6630

Rating: 4.7 / 5 (77 voted)

Reviews: 84% of readers found this page helpful

Author information

Name: Moshe Kshlerin

Birthday: 1994-01-25

Address: Suite 609 315 Lupita Unions, Ronnieburgh, MI 62697

Phone: +2424755286529

Job: District Education Designer

Hobby: Yoga, Gunsmithing, Singing, 3D printing, Nordic skating, Soapmaking, Juggling

Introduction: My name is Moshe Kshlerin, I am a gleaming, attractive, outstanding, pleasant, delightful, outstanding, famous person who loves writing and wants to share my knowledge and understanding with you.