India's largest platform and marketplace for GCCs, AI & Analytics leaders & professionals

Sign in

⁠India’s largest platform & marketplace for AI & GCC leaders & professionals

3AI Digital Library

ChatGPT and Large Language Models: A prioritization Framework for Leaders

3AI April 21, 2023

Featured Article:

Author: Robin John, Artificial Intelligence Strategist, APAC, Kimberly-Clark

Ever since ChatGPT was released last year, a flurry of LLM(Large Language Model) – related product announcements from tech giants have taken the world by storm. The key driving factor behind the success of ChatGPT is its human-like interface into AI. A technique called Reinforcement Learning with Human Feedback (RLHF) helped ChatGPT provide more “human-like” answers compared to earlier LLMs. Gradual improvements in intelligence combined with rapid improvements in conversationality enabled ChatGPT to cross a certain threshold of usability and the results are for all of us to see.

Being a conversational interface, ChatGPT is increasingly becoming a window for the world to access the powers of AI. Many people might be voluntarily interacting with an AI for the first time through ChatGPT. While it is beneficial to explore ChatGPT for a multitude of personal use-cases, it is important to have a curated approach while exploring potential enterprise use-cases.

In order to judge which use-cases would “click” for ChatGPT and LLMs in an enterprise, it is important to develop a framework to help analyze potential use-cases and also apply operational considerations before implementing them.

The framework

There are two key dimensions along which ChatGPT use-cases should be thought through. The first is the level of conversationality required for the use-case. Thanks to ChatGPT’s improved tuning to provide human-like answers, use-cases such as chatbots that require conversationality are hot favourites for implementation and rank higher on this axis.

The second dimension is the cost of error – i.e. how much would a wrong chat output cost an enterprise. For example, a bot providing medical advice should have a much lesser tolerance for factually incorrect responses compared to a ChatGPT-powered code generator. We should also take into account the ability to correct response errors by having a human in the loop before AI output reaches the end-consumer. For example, we have the opportunity to correct errors in a marketing copy generated by AI before publishing it but not in the case of a chatbot responding to a user query.

There is also an inverse relation between cost of error and the degree of creativity required. For example, there is very less creativity that should be possible in a medical advice generator whereas a content generation use-case could be all about creativity.

Figure 1: A framework for evaluating potential ChatGPT/LLM use-cases

Figure 1 includes these 2 dimensions to create a framework which we could leverage to identify the best-fit ChatGPT use-cases for an enterprise.

It is also worth noting that a lot of traditional machine learning use-cases leveraging tabular data will likely fall into the low-potential quadrant. There might be just limited  potential to make predictions and forecasts by making ChatGPT read tabular data. Traditional ML methods are likely to perform much better in these cases. Also, wherever conversationality required is less, it might be practical to explore some of the older LLMs that are likely to be simpler and more cost-efficient to implement.

Operational considerations

Once we prioritize ChatGPT-like capabilities, it is important for a decision-maker to be aware of certain key considerations that would impact delivery, especially if the path being pursued is to “build” and not “buy”.

  1. Training the model: Enterprise use-cases mostly would require “tweaking” ChatGPT to the enterprise context leveraging internal data. This could be achieved using prompt-engineering or embeddings. Each of these approaches have their unique advantages and disadvantages. Implementation teams should have conversations internally and with stakeholders before zeroing in on an approach. Practical implementations are likely to require some amount of “chaining” and “templating” to achieve right outcomes. Having a clear vision regarding training process would enable a faster path to production.
  • Platform selection: Microsoft recently announced availability of ChatGPT as a service on Azure. There are other LLM platforms too but each of these LLM platforms are at different stages of maturity. Platform maturity, combined with their availability on an enterprise’s stack would decide how soon and how well can a use-case be implemented.
  • Data privacy/Security : ChatGPT was a first of its kind and was clearly the tipping point in the LLM maturity timeline but there are other models too that could be looked at. BigScience’s BLOOM is worth looking at from a data privacy perspective. Databricks recently announced Dolly, with a “ChatGPT-like instruction following ability” but with a much simpler model that can be “directly owned” by enterprises. Enterprises should choose models basis their requirements around privacy and security. Aligning to a model that suits the enterprise priorities should help accelerate value delivery.

The current consensus is that the ChatGPT wave is comparable to the internet gold-rush of the late nineties. The ensuing bubble had triggered a lot of ideas and start-ups but soon there was a phase of consolidation as well. Similarly, it could be that a lot of potential ChatGPT ideas right now may not live up to the expectations. It is important for leaders to tread this wave with cautious optimism and spend investment dollars in use-cases with a clear path to value.

Disclaimer: Opinions expressed are solely my own and do not express the views or opinions of my employer

    3AI Trending Articles

  • Using OpenAI GPT-3 to write E-Commerce Product Descriptions

    gpt

    Featured Article: Author: Raj Bhatt, CEO, Knowledge Foundry The ProblemWriting e-commerce product descriptions has traditionally been a manual process. This activity necessitates a substantial amount of creative effort given the hundreds of new products produced by each brand/marketplace each year. While large companies use copywriters to create these product descriptions, many smaller companies opt for […]

  • Combining the power of Generative & Predictive AI

    Featured Article: Author: Shuvajit Basu GenAI has taken the world by storm. I certainly don’t have to outline the numerous possibilities it brings to bear. Both GenAI and Predictive AI have their share of strengths andbopportunities. However, if we can combine the power of these Generative models with Predictive AI/ML models, we might be able […]

  • India’s answer to Dropbox and Google Drive: NITI Aayog’s DigiBoxx

    Homegrown data storage and management platform DigiBoxx has been launched by NITI Aayog and it will provide affordable services to both regular users as well as businesses. India now has its own cloud storage platform, DigiBoxx. A potential alternative to Google Drive, Dropbox, and Microsoft OneDrive, the indigenous cloud service is seen as a step that furthers the initiative […]

  • 2020 marked a “turning point” for Cybersecurity in Australia

    Australians are on high alert about the threat of cyber attacks following Prime Minister Scott Morrison’s warning in June that Australia was targeted by a sophisticated “state-based” cyber-attack. It was, according to the Australian Cyber Security Centre (ACSC), the most significant and coordinated cyber-targeting against Australian institutions to date. The Prime Minister said while such […]