小程序
传感搜
传感圈

Nvidia launches generative AI and supercomputing cloud services as part of ‘new business model’

2023-02-26
关注

GPU maker Nvidia is launching a new service which will allow customers to access its DGX AI supercomputer in the cloud. The company also revealed a new suite of cloud-based tools for building generative AI systems. Such models have experienced a boom in popularity since the launch of OpenAI’s ChatGPT chatbot, and Nvidia has been cashing in.

Nvidia CEO Jensen Huang has announced new cloud AI products. (Photo by MANDEL NGAN / AFP)

CEO Jensen Huang revealed details of the new products on the company’s earnings call late on Wednesday, which saw it report revenue of $6.05bn, slightly up on the expected $6.01bn, but down 21% year-on-year.

Though the global economic slowdown is hitting the company’s bottom line, it expects to bring in $6.5bn in the current quarter, up on the previous estimate of $6.33bn, thanks largely to the generative AI craze fuelling demand for its chips.

Nvidia brings GDX supercomputer to the cloud

Huang revealed details of the new services, which will be officially launched at the company’s GTC Developer Conference in March, on his call with investors.

He said the company’s DGX supercomputer hardware for building AI systems would now be available virtually in the cloud. “Nvidia DGX Cloud, the fastest and easiest way to have your own DGX AI supercomputer, [you can] just open your browser,” Huang said. “Nvidia DGX Cloud is already available through Oracle Cloud Infrastructure and Microsoft Azure, Google Cloud Platform and others are on the way.”

The Nvidia CEO also used the call to talk about the company’s new AI cloud services, which he said would be offered directly by Nvidia to customers, and hosted on major cloud platforms such as Azure and Google Cloud.

Huang said the move would “offer enterprises easy access to the world’s most advanced AI platform, while remaining close to the storage, networking, security and cloud services offered by the world’s most advanced clouds”. He explained: “Customers can engage Nvidia AI cloud services at the AI supercomputer, acceleration library software, or pre-trained AI model layers.”

Customers, he said, would be able to use the platform for training and deploying large language models or other AI workloads. Nvidia will also offer a pre-trained generative AI model called NeMo and BioNeMo. Huang described these as “customisable AI models, to enterprise customers who want to build proprietary generative AI models and services for their businesses”.

Content from our partners

The role of modern ERP in transforming the distribution and logistics sector

The role of modern ERP in transforming the distribution and logistics sector

How designers are leveraging tech to apply the brakes to fast fashion

How designers are leveraging tech to apply the brakes to fast fashion

Why the tech sector must embrace faster, smarter talent recruitment

Why the tech sector must embrace faster, smarter talent recruitment

He added: “With our new business model, customers can engage Nvidia’s full scale of AI computing across their private to any public cloud.”

View all newsletters Sign up to our newsletters Data, insights and analysis delivered to you By The Tech Monitor team

Interest in generative AI goes ‘through the roof’ – Nvidia CEO

Nvidia’s announcement may not go down well with some of its biggest customers such as Google and Microsoft, both of which buy Nvidia chips to power their own AI products and services.

The company currently dominates the market for GPUs, which are used to accelerate AI workloads, so it is no surprise that the success of ChatGPT and the emergence of other generative AI models such as Google Bard has given a big boost to the company’s finances as more and more businesses seek to buy its chips.

Huang said the demand for GPUs and other processors to power generative AI systems is a major driver behind the company’s improved financial outlook. Describing the onset of such systems as a “new era of computing”, he said: “This type of computer is utterly revolutionary in its application because it’s democratised programming to so many people.” Huang said.

Referring to the ability of such models to write accurate code, he added: “This is an AI model that can write a programme for any programme. For this reason, everybody who develops software is either alerted or shocked into alert or actively working on something that is like ChatGPT to be integrated into their application or integrated into their service.”

He added: “The activity around the AI infrastructure that we build on Hopper [the company’s GPU microarchitecture] and the activity around inferencing using Hopper and [another Nvidia architecture] Ampere to inference large language models, has just gone through the roof in the last 60 days. And so there’s no question that whatever our views are of this year as we enter the year has been fairly, dramatically changed as a result of the last 60-90 days.”

Read more: China’s generative AI revolution is only just beginning

参考译文
英伟达推出生成式人工智能和超级计算云服务,作为“新商业模式”的一部分
GPU制造商英伟达推出了一项新服务,允许客户在云端访问其DGX AI超级计算机。该公司还发布了一套新的基于云的工具,用于构建生成式人工智能系统。自OpenAI的ChatGPT聊天机器人推出以来,这类模型受到了广泛欢迎,而英伟达一直在从中获利。该公司首席执行官黄仁勋在周三晚些时候的收益电话会议上透露了新产品的细节,该公司报告营收为60.5亿美元,略高于预期的60.1亿美元,但同比下降21%。尽管全球经济放缓正在影响该公司的利润,但该公司预计本季度将实现65亿美元的收入,高于此前63.3亿美元的预期,这在很大程度上得益于人工智能热潮推动了对其芯片的需求。黄仁勋在与投资者的电话会议上透露了新服务的细节,该服务将于3月在该公司的GTC开发者大会上正式推出。他表示,该公司用于构建人工智能系统的DGX超级计算机硬件现在可以在云端虚拟使用。黄说:“英伟达DGX云是拥有自己的DGX AI超级计算机的最快、最简单的方式,(你)只需要打开浏览器。”“英伟达DGX云已经可以通过甲骨文云基础设施和微软Azure使用,谷歌云平台和其他云平台正在路上。”这位英伟达首席执行官还在电话中谈到了该公司新的人工智能云服务,他表示,该服务将由英伟达直接向客户提供,并托管在Azure和谷歌cloud等主要云平台上。黄仁勋表示,此举将“为企业提供世界上最先进的人工智能平台,同时保持接近世界上最先进的云提供的存储、网络、安全和云服务”。他解释说:“客户可以在AI超级计算机、加速库软件或预先训练的AI模型层上使用英伟达的AI云服务。”他说,客户将能够使用该平台来训练和部署大型语言模型或其他人工智能工作负载。英伟达还将提供预训练生成式人工智能模型NeMo和BioNeMo。黄仁勋将其描述为“面向希望为其业务构建专有生成式AI模型和服务的企业客户的可定制AI模型”。他补充说:“通过我们的新业务模式,客户可以在其私有云到任何公共云之间使用英伟达的全面人工智能计算。”英伟达的声明可能不会受到其一些最大客户的欢迎,如谷歌和微软,这两家公司都购买英伟达的芯片来支持他们自己的人工智能产品和服务。该公司目前主导着用于加速AI工作负载的gpu市场,因此随着越来越多的企业寻求购买其芯片,ChatGPT的成功以及谷歌Bard等其他生成式AI模型的出现大大提振了公司的财务状况也就不足为奇了。黄仁勋表示,为生成式人工智能系统提供动力的gpu和其他处理器的需求是该公司改善财务前景的主要驱动力。他将这种系统的出现描述为“计算的新时代”,他说:“这种类型的计算机在应用上是完全革命性的,因为它为如此多的人提供了民主化的编程。”黄说。在谈到这类模型编写准确代码的能力时,他补充说:“这是一个可以为任何程序编写程序的人工智能模型。出于这个原因,每个开发软件的人要么警觉起来,要么震惊起来,要么积极地开发类似ChatGPT的东西,将其集成到他们的应用程序或服务中。”
您觉得本篇内容如何
评分

评论

您需要登录才可以回复|注册

提交评论

提取码
复制提取码
点击跳转至百度网盘