{"id":76396,"date":"2024-07-05T18:37:03","date_gmt":"2024-07-05T16:37:03","guid":{"rendered":"https:\/\/intellias.com\/?post_type=blog&p=76396"},"modified":"2025-12-23T14:39:15","modified_gmt":"2025-12-23T12:39:15","slug":"from-chatbots-to-ai-agents-a-leap-in-knowledge-productivity","status":"publish","type":"blog","link":"https:\/\/intellias.com\/from-chatbots-to-ai-agents\/","title":{"rendered":"From Chatbots to AI Agents: A Leap in Knowledge Productivity"},"content":{"rendered":"
When it comes to improving operational efficiency, the topic of automation often comes up. For about three decades already, even the most complex production tasks \u2014 like surface-mount technology (SMT) assembly in electronics or composite layup in aerospace manufacturing \u2014 have been performed by machines, often autonomously.<\/p>\n
However, until recently, far fewer successful advances had been made in streamlining knowledge work: tasks that require looking up, analyzing, and synthesizing information.<\/p>\n
Computers capable of generating human-like responses have existed since the 1960s, and business interest in chatbots peaked in the mid-2010s. Yet, early chatbots lacked the ability to hold meaningful conversations and provide value beyond answering simple questions.<\/p>\n
The launch of ChatGPT in late 2022 once again brought conversational interfaces into the limelight, as users were amazed by the application\u2019s cognitive powers. The subsequent hype is hard to escape, but it also somewhat conceals the actual state of adoption.<\/p>\n
A 2024 global study<\/a> by MIT Technology Review Insights (MITTR) found that although 75% of businesses have experimented with generative AI, only 9% have widely adopted the technology. Many leaders aren\u2019t fully convinced about AI\u2019s utility \u2014 especially when some of the biggest chatbot projects, like IBM Watson<\/a> and Alexa from Amazon<\/a>, have failed to live up to high expectations.<\/p>\n To understand the value of new generative virtual assistants and their actual capabilities, let\u2019s take a closer look at the underlying technology.<\/p>\n AI-Powered Digital Assistant Platform<\/p>\n The idea of a \u201cconversational computer\u201d obsessed scientists (and sci-fi authors) for decades. Joseph Weizenbaum, a professor at MIT, was the pioneer. In 1966, he revealed ELIZA \u2014 the first application that<\/a> \u201cmade certain kinds of natural language conversation between man and computer possible<\/em>\u201d and passed the Turing Test.<\/p>\n The term \u201cchatbot\u201d was coined in 1991 to describe TINYMUD, a real-time multiplayer game that chats with players. Both ELIZA and TINYMUD used a limited number of rules and keywords to detect and manipulate input queries, but they lacked proper contextual understanding. Users liked both (and sometimes mistook the software for real humans), but the agents\u2019 conversational capabilities were limited.<\/p>\n In 1995, Richard Wallace developed an Artificial Intelligence Markup Language (AIML) to power A.L.I.C.E. \u2014 a smarter version of ELIZA. A.L.I.C.E. had about 41,000 templates and related patterns<\/a> in its knowledge base, compared to ELIZA\u2019s 200 keywords and rules. However, A.L.I.C.E. relied on pattern matching without perceiving the whole conversation, as was evident in its replies.<\/p>\n As the internet grew in the early 2000s, chat apps became popular, sparking renewed interest in chatbots. This time around, researchers were focused on training bots to perform simple information retrieval tasks. Launched in 2001, SmarterChild<\/a> became the first chatbot capable of searching databases for information to provide weather reports, check stock prices, or get the latest sports scores.<\/p>\n Chatting technology improved with new techniques like probabilistic context-free grammars (PCFGs)<\/a> and dependency parsing algorithms<\/a>. Meanwhile, machine learning techniques like support vector machines and Na\u00efve Bayes classifiers were adapted to text classification tasks.<\/p>\n In the quest to find useful applications of natural language processing (NLP) technologies, researchers switched their focus to goal-oriented dialogues, where a machine helps a human accomplish a precise task.<\/p>\n <\/div> \n <\/div>\n Cleverbot<\/a>, released in 2006, was more eloquent than SmarterChild, handling free-form conversations and improving through user interactions. Commercially, businesses started adding chatbots as a marketing gimmick (you could converse with the Michelin Man<\/a> on 10,000 different topics in the early 2010s!) and to streamline customer support tasks (as Alaska Airlines\u2019 virtual assistant Jenn<\/a> started doing in 2008).<\/p>\n In 2011, IBM unveiled Watson, the most advanced natural language processing system. Although IBM Watson understood human language well enough to win the popular Jeopardy! quiz show, it didn\u2019t revolutionize<\/a> healthcare or general business operations.<\/p>\n Other types of chatbots also sprung up in the 2010s: voice assistants including Siri, Cortana, and Google Assistant. These had a better understanding of natural human language and could deliver information by predicting user requirements.<\/p>\n Social media chatbots emerged on the Facebook (now Meta) Messenger platform, and later on another messaging app. In 2016, 54% of developers<\/a> worked on chatbot projects for the first time. By the end of the year, over 34,000 chatbots<\/a> had been launched for customer support, marketing, entertainment, and educational tasks.<\/p>\n Chatbots took over tasks like providing product information or answering support questions, but user sentiment was lukewarm. A 2019 CGS survey<\/a> found that 86% of respondents preferred a human agent over a bot, and only 30% believed chatbots made issue resolution easier.<\/p>\n In 2021, Gartner moved \u201cchatbots\u201d to the trough of disillusionment stage of its Hype Cycle for Artificial Intelligence<\/a>.<\/p>\n Source: Gartner<\/a><\/em><\/p>\n But while the markets cooled for natural language processing and machine learning chatbots, important research was happening in the background.<\/p>\n Although large language models (LLMs) and generative AI may seem new, the technology took over five years to develop.<\/p>\n <\/div> \n <\/div>\n Since the mid-2010s, computer scientists have been experimenting with techniques like generative adversarial networks (GANs), variational autoencoders (VAE), and transformers \u2014 a neural network architecture that transforms an input sequence into an output sequence by learning context and tracking relationships between sequence components.<\/p>\n In 2018, OpenAI published a paper<\/a> showing how they had improved language understanding by combining transformers and unsupervised pre-training techniques. They called their new model the Generative Pre-trained Transformer (GPT).<\/p>\n GPT-1 was trained on 117 million parameters and over 4.5 GB<\/a> of text from 7,000 unpublished books. Despite being computing heavy, the GPT model achieved 80% higher accuracy<\/a> than typical baselines at that time.<\/p>\n GPT-2 1.0, released in 2019, addressed performance and stability issues with modified normalization and was trained on 1.5 billion parameters and a larger text corpus.<\/p>\n In May 2020, the OpenAI team published a paper<\/a> about the GPT-3 model, \u201can autoregressive language model with 175 billion parameters, 10x more than any previous non-sparse language model<\/em>\u201d. GPT-3 outperformed other NLP models but wasn\u2019t widely known to the public. Its capacity was 10 times larger<\/a> than Microsoft\u2019s Turing NLG<\/a>, the next largest NLP model known at the time.<\/p>\n In parallel, Google\u2019s research team was working on their own language processing architecture: Bidirectional Encoder Representations from Transformers<\/a> (BERT). Introduced in 2018, BERT aimed to improve the user search experience by deciphering ambiguous queries and providing more relevant answers.<\/p>\n BERT and GPT, two leading large language models (LLMs), employ distinct methodologies for processing language:<\/p>\n Both models build upon foundational NLP research in word representations. They leverage earlier technologies such as Stanford\u2019s GloVe<\/a> (Global Vectors for Word Representation) and Google\u2019s Word2vec<\/a>.<\/p>\n GloVe, introduced by Stanford in 2014, and Word2vec, developed by Google in 2013, are techniques that map words into a meaningful space where the distance between words reflects their semantic similarities. These embeddings provide a base knowledge of language from which the models can learn how to complete more complex tasks.<\/p>\n In November 2022, OpenAI released ChatGPT \u2014 a conversational interface for its GPT-3 model \u2014 drawing the attention of businesses to generative AI.<\/strong><\/p>\n Source: OpenAI<\/a><\/em><\/p>\n Large language models from other big tech companies soon followed (Llama from Meta<\/a>, Gemini from Google<\/a>, Ernie from Baidu<\/a>), as did new GenAI startups (Anthropic<\/a>, Mistral<\/a>, Stability AI<\/a>).<\/p>\n The latest LLM chatbots<\/a> have higher cognitive and conversational fluency than machine learning-based chatbots. The models can recognize sentiment, generate coherent replies, adapt to user prompts, and access hyper-relevant knowledge with retrieval-augmented generation (RAG).<\/p>\n Proposed by Meta AI<\/a>, RAG is a framework for retrieving data from an external knowledge base to supply the virtual agent with the latest knowledge. Instead of retraining the model, which is a resource-intensive task, you can fine-tune it to use a specific corpus of knowledge when generating responses, such as customer support or internal policy documents.<\/p>\n Thanks to RAG, companies can adapt general-purpose LLMs like GPT-4 or Mixtral 8<\/p>\n to specific business tasks. For instance, Morgan Stanley<\/strong> fine-tuned OpenAI\u2019s GPT-4 model<\/a> to provide contextual answers to its workforce using data from 100,000 corporate documents. Intellias<\/strong> developed a GPT-enabled assistant for employee skill assessments<\/a> that is capable of scanning SharePoint and LinkedIn to assemble employees\u2019 skill profiles.<\/p>\n Improved language interpretation and contextual search abilities has ushered in a new era of knowledge productivity. With virtual assistants being integrated into browsers and business apps, workers can streamline a host of document management, analytics, content generation, and data management tasks.<\/p>\n The first chatbots only provided textual prompts and summary responses. GenAI copilots can perform tasks side by side with humans, augmenting their knowledge and productivity.<\/p>\nEvolution of chatbots: From rule-based text generation to human-like cognitive capabilities<\/h2>\n
<\/p>\n
<\/p>\n\n
<\/p>\nChatbots vs GenAI virtual agents: Capabilities compared<\/h3>\n