startups

5 Steps to Ensure Startups Successfully Deploy LLMs | GVS – Global Village space


blank

The challenges of training and deploying an LLM

ChatGPT’s launch ushered in the age of large language models. In addition to OpenAI’s offerings, other LLMs include Google’s LaMDA family of LLMs (including Bard), the BLOOM project (a collaboration between groups at Microsoft, Nvidia, and other organizations), Meta’s LLaMA, and Anthropic’s Claude. More will no doubt be created. In fact, an April 2023 Arize survey found that 53% of respondents planned to deploy LLMs within the next year or sooner. One approach to doing this is to create a “vertical” LLM that starts with an existing LLM and carefully retrains it on knowledge specific to a particular domain. This tactic can work for life sciences, pharmaceuticals, insurance, finance, and other business sectors.

LLMs have already led to newsworthy issues, such as their tendency to “hallucinate” incorrect information. That’s a severe problem, and it can distract leadership from essential concerns with the processes that generate those outputs, which can be similarly problematic.

One issue with using LLMs is their tremendous operating expense because the computational demand to train and run them is so intense (they’re not called large language models for nothing). First, the hardware to run the models on is costly. The H100 GPU from Nvidia, a popular choice for LLMs, has been selling on the secondary market for about $40,000 per chip. One source estimated it would take roughly 6,000 chips to train an LLM comparable to ChatGPT-3.5. That’s roughly $240 million on GPUs alone.

Another significant expense is powering those chips. Merely training a model is estimated to require about 10 gigawatt-hours (GWh) of power, equivalent to 1,000 U.S. homes’ yearly electrical use. Once the model is trained, its electricity cost will vary but can get exorbitant. That source estimated that the power consumption to run ChatGPT-3.5 is about 1 GWh a day, or the combined daily energy usage of 33,000 households.

Read More   US-headquartered Techstars stands out as most active startup ... - PitchBook News & Analysis

Power consumption can also be a potential pitfall for user experience when running LLMs on portable devices. That’s because heavy use on a device could drain its battery very quickly, which would be a significant barrier to consumer adoption.

LLMs are exciting, but developing and adopting them requires overcoming several feasibility hurdles.

Deploying an LLM can provide a powerful competitive advantage — but only if it’s done well.

LLMs are exciting, but developing and adopting them requires overcoming several feasibility hurdles. The challenges of training and deploying an LLM are significant. The hardware required to run LLMs is costly, with the H100 GPU from Nvidia selling for about $40,000 per chip. It would take approximately 6,000 chips to train an LLM comparable to ChatGPT-3.5, resulting in a cost of around $240 million for GPUs alone.

In addition to the high hardware costs, powering the chips used to train LLMs is also expensive. Training a model can consume about 10 gigawatt-hours (GWh) of power, equivalent to the yearly electrical use of 1,000 U.S. homes. Once trained, the electricity cost to run an LLM like ChatGPT-3.5 can be exorbitant, with an estimated daily power consumption of 1 GWh, equivalent to the combined energy usage of 33,000 households.

Furthermore, power consumption can be a potential issue when running LLMs on portable devices. Heavy usage of LLMs on these devices can quickly drain their batteries, posing a significant barrier to consumer adoption.

Despite these challenges, deploying an LLM can provide a powerful competitive advantage if done well. However, it is crucial to address the feasibility hurdles associated with LLM development and adoption.

Read More   Ensuring Business Continuity Amidst Power Disruptions: Info-Tech Research Group Publishes Guide for Uninterrupted ... - PR Newswire



READ SOURCE

This website uses cookies. By continuing to use this site, you accept our use of cookies.