Skip to main content

How to make efficient use of generative AI

Generative AI is here to stay.

It is not only a hype that probably gets worse before it gets better. And we clearly still are in a hype, as the following chart showing the search interest for ChatGPT between October 1, 2022 and April 12, 2023 from Google Trends shows.

Similarly, the Gartner Group sees Generative AI technology approaching the peak of inflated expectations in its 2022 hype cycle for artificial intelligence.

To be sure, we see only the tip of the iceberg when looking at voice, text or image based services that we all know and use. The Gartner Group also foresees many industrial use cases reaching from drug and chip design to the design of parts to overall solutions. 

You think that these scenarios lie far in the future? Read this Nature article from 2021 and think again.

And in contrast to some of the other hypes that we have seen in the past few years, there are actual use cases that support the technology’s survival of the trough of disillusionment. As there are viable use cases, unlike “Metaverse”, Blockchain or NFTs have shown, generative AI is not a solution in search of its problem.

Apart from OpenAI’s GPT and Dalle-E models that surely caught everybody’s attention in the past weeks and months, there are a good number of large language models that are just less known. A brief research that I recently conducted, unearthed more than 50 models that got published over the past few years. For their paper A Survey of Large Language Models that focuses on “review[ing] the recent advances of LLMs by introducing the background, key findings, and mainstream techniques”, a group of AI researchers identified a similarly impressive number of large language models that got developed in the past few years.

Fig. 1: A timeline of large language models; source: A Survey of Large Language Models

This increased competition, along with research into how the resource consumption of large models can be reduced, e.g. through sparsification, will lead to improved pricing.

Consequently, businesses need to find a way to leverage this technology without harassing their own data security and/or exposing their intellectual property (IP).

That this is easily possible, has been already learned the hard way by companies as diverse as Amazon or Samsung, to name just two better known cases. And then, there are ongoing use cases around the unauthorized use of IP-protected data for training and even many questions around the ownership of generated (derivative?) work products of generative AI, not even speaking in unwanted bias. 

The way to be walked involves three major steps:

Education

It is necessary to educate the teams not only on the benefits of using generative AI but also on how to check the output and, importantly, when to not use generative AI. Every employee has a role in keeping the business protected from avoidable mistakes. And no employee wants to make one of those mistakes. Sources that are used for any content that is generated should to be attributed and/or verifiably not be copyrighted. Create some sensitivity about “when in doubt, don’t use it”. Enable your people. For the time being, this training should be frequently reviewed, as the whole area is evolving fast.

Second, it is important to enable personnel to fine tune and then train large language models. This is not necessarily a trivial task that requires some knowledge about machine learning.

Governance

As with most technologies, there is a need for governance to augment the training. Generative AI is a very powerful tool, so it needs to be used responsibly and ethically. To help the employees with their own decision making about when and how it is permissible to use generative AI services in the course of their work, and according to the corporate values, it is necessary to develop, publish and implement a policy that outlines and explains the relevant guidelines to follow.

Points that should be covered by the policy include;

  • Who is allowed to use generative AI services. Depending on numerous factors like corporate culture, sensitivity of information dealt with, etc. it may be useful to explicitly limit the use of generative AI.

  • Proprietary or confidential information must not be revealed or disclosed.

  • Similar to other policies (e.g. use of the Internet in general), there may be a limitation to business use or the permission to also use it for private purposes.

  • The services should not be used for tasks that are violating any law, regulation, or the company’s code of conduct. This involves discrimination, offensive or hate speech that should be prohibited. This also includes content that is intended to deceive or defraud others.

  • Whatever the generative AI produces needs to be reviewed. 

  • The use of generative AI to create a work product should be disclosed in order to maintain transparency.

The policy should also include a reference to the offered training that personnel should undertake. Last, but not least, and as bad as it sounds, the consequences of violating the policy must be made clear as well. Here, too, I would concentrate on explanation and rationalizing instead of just wielding a stick.

It might be useful to review this policy frequently, as we are still in a learning period.

Following my own suggestion here: I have used You.com and Google Bard to give me some points that the policy should cover.

Execution

In parallel to this effort, it is time to look for use cases that can meaningfully be implemented. Collect and assess them. Meaningful means that the use cases are important enough to add value and isolated enough to still be able to manage any risks. These use cases can include improving existing capabilities or adding additional ones. What is important is that for all use cases there need to be KPIs that shall get improved and a cost-benefit analysis.

Remember: Not every problem needs to be solved with the help of a LLM! For some problems it is more efficient to use more “traditional” technologies.

At the same time, it is important to be aware of the level of LLM-knowledge that is available in house. This level is, especially in smaller organizations, not too high. The degree of available knowledge is a boundary factor for project execution. So, it is best to start off with a problem that requires only minor fine-tuning of an existing LLM that is powerful enough to to support multiple of the collected use cases. This way, it is possible to learn fast while also getting results fast. Additionally, fine tuning a model comes at far lower cost than fully training a model, as the number of required training cycles is smaller and the amount of necessary training data is far lower. Out of the many available models (see above), one or more will be capable of supporting most of the selected use cases after fine tuning.

The way to success is thinking big while acting small and to walk before starting to run.


 

Comments

Last Year's Top 5 Popular Posts

SugarCRM explains how the third wave of CRM adds value

The news On October 4 and 5, 2023, SugarCRM held its Connected event followed by an analyst summit in London. The first day – Connected – was targeted mostly at customers while the second day focused on analysts.  The event started off with an intense speech by Katherine Grainger, DBE , a British rowing champion. Her core messages were about team bonding, the importance of communication, continuous improvement, and perseverance (well, at least that’s my take). This was followed by information about what is new in the software and, more importantly, a customer panel.  The main sponsor, Mobileforce , placed some words about the partnership. In addition, the analysts had 1:1s with customers, partners, and Sugar executives. The second day was filled with information targeted at analysts. CEO Craig Charlton and his executive team shared about financial status, strategy and more in-depth product news. Sugar being a privately held, VC backed company, the financials are of course under NDA, s

Relevance, reliability, responsibility are key for AI – the SAP way

The News A lot is going on in the SAPverse during October and the early days of November 2023. First, SAP conducted its CXLive event with CX-related announcements, then the company reported good Q3/2023 figures, a new version of its CX software that includes new generative AI capabilities got released and lastly, it executed its SAP TechEd event with a good number of AI-, BTP-, and ERP related announcements. As this is quite a lot, I covered the CX world in a previous post and will cover the TechEd related news in this post.  So, what is new at SAP TechEd ? For one, it is enough to fill a 17-page pre-event news guide that SAP sent out. SAP certainly is able to stack up the news for major events. I took the liberty to ask ChatGPT for a summary of the document, which I slightly edited afterwards. Here we are: AI and Development Environments: ·       SAP introduces SAP Build Code with generative AI, improving application development and testing, while new AI capabilities are integrate

How to play the long game Zoho style

The news On February 7 and 8 2024, Zoho held its annual ZohoDay conference, along with a pre-conference get together and an optional visit to SpacX’s not-too-far-away Starbase. Our guide, who went by Chief, and is probably best described as a SpaceX-paparazzi was full of facts and anecdotes, which made the visit very interesting although we couldn’t enter Starbase itself. The event was jam-packed with 125 analysts, 17 customer speakers, and of course Zoho staff for us analysts to talk to. This was a chance we took up eagerly. This time, the event took place in MacAllen, TX, instead of Austin, TX. The reason behind this is once more Zoho’s ruralization strategy, transnational localism.  Which gives also one of the main themes of the event. It was more about understanding Zoho than about individual products, although Zoho disclosed some roadmaps. More about understanding Zoho in a second.  The second main theme was customer success and testimonials. Instead of bombarding us with presenta

Don't mess with Zoho - A Zohoday 2022 recap

After spending two days in Austin, TX, attending the ZohoDay 2022, it is time for a little recap of this interesting event.  We were 99 analysts and 24 customers and plenty of knowledgeable Zoho personnel. The incredible Sandra Lo and her team organized the event around open and transparent communication. So, there was plenty of access for us to customers and the Zoho team.  Which was very important, as already the keynote session by founder and CEO Sridhar Vembu was quite hardcore. Vembu talked about how strategy and culture need to be one, how culture needs to be the root of strategy, and how Zoho implements this. The Zoho strategy lies on three main pillars ·       Transnational localism, a unique concept that in its essence is about embedding a company into a local community by not only selling into it but also by investing into it. This investment is e.g., by offering high paying jobs in areas where these are scarce, by fostering local education, but also by own local sourcing in

The Generative AI Game of Thrones - Is OpenAI toast?

The News This has been an exciting weekend for the generative AI industry. On Friday November 17, OpenAI announced that the company fired its figurehead CEO Sam Altmann and appointed Chief Technology Officer Mira Murati as interims CEO in a surprise move. The press release states that Altmann “ was not consistently candid in his communications with the board .” Surprised was apparently not only Sam Altmann, but also the till then chairman of the board Greg Brockman who first stepped down from this position and subsequently quit OpenAI. Investors, notably Microsoft, found themselves blindsided, too – or flat footed depending on the individual point of view. Satya Nadella was compelled to state that Microsoft stays committed to the partnership with OpenAI in a blog post that got updated on November 19, 11:55 pm. All hell broke loose. Microsoft shares took a significant hit. A number of additional senior OpenAI personnel quit. Both, Altman and Brockman, voiced the idea of founding anoth