--- pinned: true title: "History of the Generative AI" description: "The rise in power of the generative AI is a revolution based on Deep Learning, transforming the tech world into a new one." authors: ["glegoux"] time_reading_minutes: 9 category: "Machine Learning" --- **Generative AI** is the state of the art of machine learning and has excelled recently with the apparition of **AI ChatBots with high performance**. {% include content/image.html src="https://miro.medium.com/v2/resize:fit:700/1*UenOkL0Tn-XuP-hOho9b3Q.png" abs_url=true source_author=true title="Generative AI for all types of content" %} Letโ€™s discover its history, from simple textual machine learning models specialized to **universal** **large language models** at a vast scale: - ๐Ÿ’ฌ AI ChatBots - ๐Ÿ”ฅ From NLP to Multimodal GAI ChatBots based on LLM - ๐ŸงชLLMs Timeline: Closed-source vs. Open-source - ๐ŸŽฏ From Specialized Model to Generalized ML Model - ๐Ÿ“ˆ Scaling Laws & Performance - ๐Ÿ Technologic and Economic Race - ๐Ÿš€ Go Further with Limited & Augmented LLMs # ๐Ÿ’ฌ AI ChatBots Building generalist **AI chatbots** is the most complex challenge for a technology that simulates human intelligence. It is the **base of the artificial human assistant**. They are experiencing dazzling success with the **release of ChatGPT**, a new step that has been overcome after Google Home and Alexa by AWS. The technical component behind these products has disrupted and improved the ecosystem considerably, **although not perfect yet** (perhaps it will never be). The interest in chatBots followed the one for deep learning from 2010. Before, it was reserved for a restrained audience. Still, it exploded in **November 2022** with the release of **chatGPT** open to the general public with a ramp-up of **100M Monthly Active Users (MAUs) after two months**, a worldwide record (4.5x faster than TikTok) and **1.6B MAUs now**. At the same time, machine learning and deep learning trends have continued growing in the background for ten years. {% include content/image.html src="https://miro.medium.com/v2/resize:fit:961/1*imwYfo_SLbjeVzeZ420QhQ.png" abs_url=true title="Google Trends from 2010" source="https://trends.google.com/trends/explore?date=2010-01-01%202023-04-01&geo=FR&q=Machine%20Learning,Deep%20Learning,Chatbot&hl=en" %} {% include content/image.html src="https://miro.medium.com/v2/resize:fit:853/1*bdBhXytZGBsTrI3t-zxd4A.png" abs_url=true title="Similar web for chat.openai.com" source="https://www.similarweb.com/website/chat.openai.com/#ranking" %} To have an order of magnitude, the human brain has, on average, 100 billion neurons and 100 trillion synapses. The estimated resources used for **chatGPT+** (paying version of chatGPT) use deep learning models based on GPT-4 with an estimated model size close to these values (digits kept private by OpenAI). But a **human neuron/synapse** is much more powerful than a **deep learning neuron/synapse.** The gap is still important, above on reasoning tasks on unknown problems requiring many resources. {% include content/image.html src="https://miro.medium.com/v2/resize:fit:382/1*Ty6-E7apO67lSr02i-6kvQ.png" abs_url=true source_adapted="https://www.researchgate.net/figure/Visual-comparison-between-human-nerve-cell-and-neural-network-architecture14_fig3_360640306" title="Human brain vs. Deep learning" %} # ๐Ÿ”ฅFrom NLP to Multimodal GAI ChatBots based on LLM Thanks to **Large Language Models (LLMs)** working with Generative AI, chatBots continue getting closer to the objective of simulating human intelligence**.** The **NLP AI models** have progressively migrated from simple parsing to c**omplex processing on the language structure**, from shallow learning to **deep learning**, from a single-language vocabulary of words to a **vocabulary of tokens with multi-language embeddings**, from an RNN (LSTM) to the **Transformer with multi-head attention layers** approach. {% include content/image.html src="https://miro.medium.com/v2/resize:fit:700/1*4N5CPyYSVFzfEapqGM5gug.png" abs_url=true title="NLP timeline with the Deep Learning" source="https://arxiv.org/pdf/2202.05924.pdf" %} Finally, by mixing other types of data (like images, audio, โ€ฆ), NLP models became **multi-modal generative models** where varied instructions (in input) give varied results (in output). {% include article/read-more.md src="https://medium.com/@glegoux/history-of-the-generative-ai-aa1aa7c63f3c" %}