Stronger, scarier: A guide to AI in 2024

Sam Altman, CEO of OpenAI, which developed ChatGPT. The chatbot heralded the rise of GenAI.  (AFP)
Sam Altman, CEO of OpenAI, which developed ChatGPT. The chatbot heralded the rise of GenAI. (AFP)

Summary

  • While GenAI has captured everyone’s imagination, there’s a need for stronger global guard rails

Bengaluru: Last month, The New York Times filed a suit against OpenAI and Microsoft, alleging that their generative artificial intelligence (AI) tools—ChatGPT and Bing Chat (now rebranded as Co-pilot)—were built by copying and “using millions of The Times’s copyrighted news articles, in-depth investigations, opinion pieces, reviews, how-to guides, and more".

While GenAI has captured everyone’s imagination and will continue to mature in the years ahead, the suit underlines the need for stronger global guard rails. Apart from copyright violations, there are fears of machines acquiring human biases and GenAI vapourizing millions of white-collar jobs.

Here’s a glimpse of how individuals and enterprises are embracing these technologies as of today and what we must expect this year.

Classical AI to Generative AI

AI adoption is poised to add $500 billion to India’s GDP by 2025, according to Nasscom, an industry body.
View Full Image
AI adoption is poised to add $500 billion to India’s GDP by 2025, according to Nasscom, an industry body. (istockphoto)

By the time OpenAI’s large language model (LLM)-powered chatbot, ChatGPT, released on 30 November 2022, AI was already powering smartphones, digital wallets, search engines, online shopping recommendations, banking apps, healthcare products and services and weather forecasts, among other things. But ChatGPT heralded the rise of GenAI.

Other GenAI systems include Google’s Gemini 1.0, Meta’s LLaMA 2, Anthropic’s Claude 2, OpenAI’s Dall-E 2 and 3 and MidJourney.

Unlike traditional machine learning that can analyse data patterns to make predictions, GenAI systems can instantly create new articles, images, videos, presentations and even write simple code with the help of ‘prompts’ in natural languages like English and Hindi.

Large language models have more than 100 billion parameters. The second half of 2023 saw numerous small language models (SLMs) emerge. They have fewer parameters, consume less energy, are affordable and can perform domain-specific tasks relatively well after fine-tuning on fewer samples. Examples include LlaMA 2 13B (B stands for billion, and refers to parameters), Mistral 7B, Falcon-7B, IBM Granite-13B, and Microsoft Phi-2 2.7B.

So, where do we go from here?

Research firm Gartner believes that GenAI will impact the pharmaceutical, manufacturing, media, architecture, interior design, engineering, automotive, aerospace, defence, medical, electronics and energy sectors by augmenting core processes with AI models. It adds that 30% of outbound marketing messages from large organizations will be synthetically generated, up from less than 2% in 2022, even as more than 30% of new drugs and materials will be systematically discovered using GenAI techniques by 2025, up from zero today.

AI adoption is poised to add $500 billion to India’s GDP by 2025, according to Nasscom, an industry body.

That said, enterprises are treading this ground with a lot of caution since GenAI tools can be misused to create ‘deepfakes’ and generate fake copies of products. Further, since LLMs are trained on large amounts of publicly-available data, they can end up violating General Data Protection Regulation (GDPR) and intellectual property (IP). Enterprises also need to account for hallucinations (generating fake content) and privacy violations by such tools. Besides, GenAI models hog a lot of electricity, which could compromise an organization’s sustainability goals.

Ring-fencing AI

Masayoshi Son, the CEO of SoftBank. According to Son, artificial super intelligence would be realized in 20 years and surpass human intelligence by a factor of 10,000.
View Full Image
Masayoshi Son, the CEO of SoftBank. According to Son, artificial super intelligence would be realized in 20 years and surpass human intelligence by a factor of 10,000. (Mint)

Some believe that the phenomenal growth of GenAI models indicates that machines will soon think and act like humans, a trend referred to as artificial general intelligence, or AGI. Some also call it artificial super intelligence, or ASI. Further, these models can plagiarize, be biased, and potentially replace thousands of routine jobs while also posing security and privacy risks. The fear is that if we are unable to fully understand the workings of these unsupervised networks, they could automatically evolve into Syknet-like machines we see in the Terminator movie franchise.

The cohort of people with this fear includes Elon Musk, co-founder of Tesla and SpaceX, and owner of X, formerly Twitter. It also has Masayoshi Son, CEO of SoftBank, and AI experts such as Geoffery Hinton and Yoshua Bengio. According to Son, ASI would be realized in 20 years and surpass human intelligence by a factor of 10,000.

The other cohort, which includes AI experts such as Yann LeCun, Fei-Fei Li, Andrew Ng, believe that AI is nowhere close to becoming sentient. They underscore that AI’s benefits—such as powering smartphones, driverless vehicles, low-cost satellites, chatbots, flood forecasts and warnings—far outweigh its perceived risks.

Despite their differences, all of them agree that AI and GenAI need to be reined in, and governments around the world are doing so. On 8 December, for instance, the European Union (EU) presented a comprehensive risk-based draft AI Act and is likely to publish the final text in early 2024. The Act seeks to ban biometric categorization systems that use certain sensitive characteristics, and untargeted scraping of facial images from the internet or CCTV footage to create facial recognition, among other things.

Earlier, the Group of Seven (G7), comprising Canada, France, Germany, Italy, Japan, the UK and the US, introduced guiding principles and a code of conduct for AI, even as all these countries have draft regulations in place. India is already a founding member of the Global Partnership on Artificial Intelligence (GPAI), an international forum that “aims to bridge the gap between theory and practice on AI by supporting cutting-edge research and applied activities on AI-related priorities".

India, which has a Digital Personal Data Protection Bill, is soon expected to introduce the Digital India Act (DIA)—the new law is expected to regulate AI and intermediaries, even as the Indian government seeks to strike a fine balance between acknowledging the good in AI and penalizing its misuse.

The Indian LLMs

Akash Ambani, chairman of Reliance Jio Infocomm. He announced a partnership with IIT Bombay to build ‘Bharat GPT’, an LLM tailored to suit the country’s needs.
View Full Image
Akash Ambani, chairman of Reliance Jio Infocomm. He announced a partnership with IIT Bombay to build ‘Bharat GPT’, an LLM tailored to suit the country’s needs. (PTI)

In December, Akash Ambani, chairman of India’s largest telecom service provider, Reliance Jio Infocomm, announced a partnership with the Indian Institute of Technology (IIT) Bombay to build ‘Bharat GPT’, an LLM tailored to suit the country’s needs. That very month, Ola co-founder Bhavish Aggarwal’s new company, Krutrim, too, joined the growing band of companies focusing on building India-specific LLMs. It is certainly the need of the hour but is easier said than done, given high computing costs and paucity of good Indian data sets.

The move makes sense since India is home to more than 400 languages, making it one of the most linguistically diverse countries in the world, and yet, most foundation models and LLMs are trained primarily using internet data, which is predominantly English. As per Statista, English was the most popular language for web content, representing nearly 59% of websites as of January this year. Russian ranked second with 5.3% of web content, followed by Spanish with 4.3%.

That said, many of the 22 official Indian languages do not have digital data, which makes it challenging to build and train an AI model with local data sets. Bhashini, a unit of the National Language Translation Mission, has so far spent $6-7 million to collect data from different sources, according to its CEO, Amitabh Nag. Bhashini has also employed more than 200 people to collect data (text as well as speech) and feed it into the system, following which the data is curated, annotated, and labelled.

Likewise, the ‘Nilekani Center at AI4Bharat’ at IIT Madras is building open-source language AI for Indian languages, including data sets, models, and applications. Sarvam AI, a generative AI startup founded by Vivek Raghavan and Pratyush Kumar (both co-founders of AI4Bharat), is developing LLMs specifically for India—the OpenHathi Series. Meanwhile, Bengaluru-based AI and Robotics Technology Park and the Indian Institute of Science are partnering with Google India to launch a LLM called Project Vaani.

CoRover, a GenAI company, has launched its own indigenous LLM called BharatGPT, which is available in more than 12 Indian languages in partnership with Bhashini. Last, but not the least, the Mahindra Group aims to construct an indigenous LLM, specifically designed to converse in a multitude of Indic languages. In the first phase, the Indus Project, as it is called, targets the inclusion of a remarkable 40 Hindi dialects.

Quantum is coming of age

IBM Quantum scientist Maika Takita in a lab. Some companies and research institutes are racing to develop ‘fault-tolerant’ quantum computers.
View Full Image
IBM Quantum scientist Maika Takita in a lab. Some companies and research institutes are racing to develop ‘fault-tolerant’ quantum computers. (Photo: IBM)

The computers we use in our homes and offices today process information with bits (ones and zeroes), while quantum computers have qubits that can process the ones and zeroes simultaneously due to a property known as superposition. This means that two bits in your normal computer can be in four possible states (00, 01, 10, or 11) but can represent only one of these states at any given time.

A quantum computer, on the other hand, allows two qubits to represent the exact same four states at the same time, akin to four computers running simultaneously. Further, as you add more qubits, the power of your quantum computer grows exponentially. And a property called quantum entanglement allows qubits to communicate with each other even if they are miles (or even millions of miles) apart.

But building a quantum computer is very expensive and researchers have to address the unintended interactions between qubits and the environment, which are referred to as ‘noise’. Further, qubits are coherent—a property that keeps the qubits stable—only for a small amount of time. Adding qubits increases the decoherence, making the task even more challenging.

Nevertheless, companies including IBM, Google, Microsoft, Amazon, and Baidu, along with top research institutes, are racing to develop a ‘fault-tolerant’ quantum computer that can be used for real-world applications. And IBM is already working to integrate GenAI to help automate the development of quantum code for Qiskit, an open-source software development kit (SDK) for working with quantum computers.

Quantum computers can simulate molecular structures and interactions more accurately than classical computers, accelerating the discovery of new drugs, catalysts, and materials by predicting their properties and behaviour. For instance, scientists in China say their latest quantum computer, the JiuZhang 3, has solved a mathematical problem within a millionth of a second – more than 20 billion years quicker than the world’s fastest supercomputer could achieve the same task, according to a paper published by the peer-reviewed journal Physical Review Letters.

Further, quantum computers could also improve certain machine learning algorithms and optimization tasks by handling large data sets more efficiently or finding optimal solutions for complex models.

AI, the energy-guzzler

Global data centre electricity consumption has grown by 20-40% annually in recent years, researchers have pointed out.
View Full Image
Global data centre electricity consumption has grown by 20-40% annually in recent years, researchers have pointed out. (AFP)

Flooding is estimated to impact more than 250 million people globally each year and cause around $10 billion in economic damages. AI and data analytics can help design better systems. As an example, Tamil Nadu, a state that gets flooded annually, has an integrated coastal flood warning system called CFLOWS, and a Real-Time Flood Forecasting and Spatial Decision Support System (RTFF & SDSS) for the Chennai basin. Even IIT-Madras has a high-resolution regional weather forecasting model, whose output is being integrated with the RTFF & SDSS on a pilot basis. Currently, though, the efficacy of the system is reportedly being hampered by lack of adequate field equipment.

That said, a good flood forecasting model can help researchers suggest better designs for stormwater drainage and flood management systems, among other flood interventions. Google Flood Hub’s AI, as an example, uses publicly-available data sources such as weather forecasts and satellite imagery. It’s being used in 80 countries, including India and Bangladesh. The system combines two models: the hydrologic model, which forecasts the amount of water flowing in a river, and the inundation model, which predicts what areas are going to be affected and how deep the water will be. Google is also working to expand flood forecasting alerts in Search and Maps.

Governments, meanwhile, could also use smart robots and drones for disaster management. AI-powered robotic flood control systems use autonomous robots that are equipped with advanced sensors and AI algorithms. It helps detect potential weak points in flood defenses—like cracks or erosions, and perform the necessary repairs.

While the use cases are promising, AI and generative AI are also energy-guzzlers. The most energy-and carbon-intensive tasks are those that generate new content: text generation, summarization, image captioning, and image generation. Tasks involving images are more energy-and carbon-intensive compared to those involving text alone. For instance, using a text-to-image model to craft an AI-generated image can require almost the same amount of power required to charge a smartphone, according to a paper released on arXiv by a team of researchers from Hugging Face and Carnegie Mellon University.

The researchers pointed out that global data centre electricity consumption has grown by 20-40% annually in recent years. However, the contribution of the AI sector specifically towards these figures is yet to be ascertained. As tech companies continue to build GenAI models, they will consume more energy, necessitating climate-friendly data centres and IT infrastructure, among other things.

Catch all the Business News, Market News, Breaking News Events and Latest News Updates on Live Mint. Download The Mint News App to get Daily Market Updates.
more

topics

MINT SPECIALS