国产一区二区美女诱惑_国产精品免费播放_91精品国产综合久久香蕉麻豆 _久久精品30_久久综合88_国产精品亚洲成人_黑人极品videos精品欧美裸_亚洲色图欧美激情

原創生活

國內 商業 滾動

基金 金融 股票

期貨金融

科技 行業 房產

銀行 公司 消費

生活滾動

保險 海外 觀察

財經 生活 期貨

當前位置:科技 >

Large Models: an Exclusive Party for the Upper Crust

文章來源:鈦媒體APP  發布時間: 2023-07-11 18:26:42  責任編輯:cfenews.com
+|-

Credit:Visual China

By Zhang Shuai and Shaw Wan

BEIJING, July 11 (TiPost)— “As long as you have a graphics processing unit (GPU), I can buy the whole server,” said a buyer. Zhang Yang, head of a cloud computing service provider, witnessed customers going on a shopping spree since March. “At that time, customers were in such a rush that all they cared about was to hoard the devices. They didn’t have any requirements for the products, nor did they mention cloud networking or data storage. They didn’t even know how to put them into good use,” said Zhang.


(資料圖片)

It wasn’t until April that some of these buyers started to figure out what kind of devices they truly needed. They went on a wrong track of hoarding a bunch of GPUs, when training a large model demands massive distributed computing power that often pairs with a full set of services.

The computing power industry involves various parts, such as artificial intelligence (AI) chips, servers, optical modules, data centers and cloud computing platforms, forming the driving force for the digital economy. Because of such scale and complexity, only a small number of enterprises are able to afford to join the race. Since training a large model is the start point of the large model ecosystem basic, enough computing power is the admission ticket to the industry.

The birth of the AI chatbot ChatGPT in last November has proven the partnership between Microsoft and Open AI a success. Training models in the cloud turned out to make the cut. For companies, cloud service providers can provide scalable computing resources, including hardware and software, which saves businesses from building their own infrastructure. They can also serve talents in the industry, such as R&D engineers, algorithm engineers and individual developers. Since they are often backed up by tech giants, they are rich in financial support, talents and data, enabling them to crack on with the big model training.

The frenzied ChatGPT wave

The companies that are qualified for the race are mostly established giants. For example, the super computer behind ChatGPT is Azure, a Microsoft-backed cloud computing platform that was initially released 14 years ago . At the current stage, to catch up with ChatGPT or to get ahead in the race, companies are competing against each other with their strategies and technologies that were developed in the past years.

“The large model training is obviously hyped. The industry should be more rational and avoid taking advantage of the concept to lure investment. I think the companies that want to scramble for a piece of the pie shouldn’t start from scratch. There are possible chances for them, but there are also great challenges,” said a person in charge of large-scale model products of a tech giant.

From the perspective of academia, OpenAI does not embody revolutionary innovation. It is more like "engineered innovation" of the artificial general intelligence (AGI) products. The engineered production process involves different phases of big model training, including the research, engineering, products and organization.

“It is also hard to pull the engineered production off. However, it at least proved that having more computing power and more data would work,” said Han Kai, principal engineering manager at Microsoft.

Although engineered production has proven to be a success, it was hard for many other companies to choose the path in the first place because the huge input didn’t promise a bright future. Chinese enterprises tend to follow others’ steps, which is also why ChatGPT wasn’t born in China.

Challenges in big model training

There are at least three major challenges to achieve the engineered production of cloud computing.

The first is computing power. “For example, GPT-3, which has 175 billion parameters, required 314 zettaFLOPS of computing power in order for it to be trained. However, as one GPU only delivers 312 teraFLOPS of deep learning performance, it takes 32 years to train one large model. Therefore, it is necessary to introduce the distributed training, where we use multiple devices and multiple GPUs to train large models,” said Chen Xi, an expert at Chinatelecom Cloud, a cloud computing platform backed by China Telecom.

Data storage can also be a problem. The video memory of a single GPU can no longer load a model with hundreds of billions of parameters. It takes about a few terabytes to fully load so many parameters. That number can grow even bigger, when intermediate results generated during the training process, such as gradients and optimizer states, are also taken into consideration. Thus, hundreds of GPUs are needed.

Therefore, companies generally adopt a way of pipeline parallelism, and run different layers of the model in GPUs of different nodes. In this way, a group of nodes only need to load a limited number of parameters, reducing the pressure on the memory.

As the big model training task is broken into a sequence of processing stages, there will be a large amount of communication between clusters, resulting in high requirements on the bus and its bandwidth. The amount of data transferred can reach up to hundreds of gigabytes.

Apart from these three major challenges, the fast growth of large model parameters and the slow development of chip technology is also hindering the industry. In recent years, with the introduction of the transformer, the number of model parameters has increased by 15 times every two years. However, the development of chip technology lags behind. The computing power of a single GPU grew by less than 4 times, with the chip process decreased from 7 nanometers to 4 nanometers.

Large model training requires not only computing power, but also storage, security, and the training framework. A complete set of platforms or services are needed to provide support. “We feel like there are not many service providers who can satisfy the needs of large model training. And the overall supply of high-performance computing power is relatively tight,” said Chen.

Opportunities for Chinese chip makers

As the Chinese companies are trying to jump on the bandwagon of large model training, the demand for chips soar. Although Chinese chip makers are trying to catch up with the top chip designers, they are not the first choice of many computing power platforms.

“At present, when everyone is working on large model training, time is of the essence. What the industry needs is high-end products, so that they can avoid stability or maturity problems. That’s why the Chinese chips are left out,” said Zhang Yalin, the chief operating officer of Enflame, a Shanghai-based AI start-up developing cloud-based deep learning chips for AI training platforms.

The American chip maker Nvidia is the dominant supplier for the reasoning and training of large models in China. Chinese tech giant Baidu once purchased tens of thousands of Nvidia A800s within just half a year.

According to Nvidia’s financial results for first quarter fiscal 2024, the revenue of Nvidia"s data center business was 4.28 billion dollars, a record high with an increase of 14 percent compared with the same period last year. In May , shares of the company soared, taking the company"s valuation above one trillion dollars.

However, in terms of reasoning, there are still business opportunities for Chinese chips. “I think Chinese chip makers should take a different path, by starting with reasoning and fine-tuning. They can later cooperate with research institutes from universities and national laboratories to move on to large model training,” Zhang said.

The development of AI chips is faster than Moore"s Law, which could also lead to a decline in growth in some cases, according to Xie Guangjun, the vice president of Baidu. The temporary shortage of computing power is because the computing power cannot keep up with the demand. The overall imbalanced supply chain is also part of the cause.

As of now, no Chinese chips can replace the high-end chips produced by Nvidia, such as the A100. Several Chinese chip makers planned to release similar products later this year. As the shortage of Nvidia chips persists, Chinese chips are likely to grab a slice of the cake after next year, once they are able to meet the requirements.

“Although Internet companies care more about the price-performance ratio, they need to pay more attention to the total cost of ownership when it comes to computing power. For example, the GPU cluster of my company can support 1,000 GPUs, whose performance is similar to that of Nvidia’s cluster of 600 GPUs. But our products can also be competitive, as long as we can provide more cost-effective and customized services,” said Zhang.

If Enflame’s products were to be favored by the Internet clients, they needed to have 1.5 times the performance of Nvidia products and twice the price-performance ratio in the desired scenarios and businesses, Zhang added.

As early as June 2021, Baidu AI Cloud began to plan the construction of a new GPU cluster of high-performance. Together with NVIDIA, it completed the design of the InfiniBand network architecture, which can be equipped with over 10,000 GPUs to provide EFLOPS-level computing power. Thanks to this cluster, Baidu released its ChatGPT-style AI bot called Wenxin Yiyan, or ERNIE Bot, in March.

關鍵詞:

專題首頁|財金網首頁

投資
探索

精彩
互動

獨家
觀察

京ICP備2021034106號-38   營業執照公示信息  聯系我們:55 16 53 8 @qq.com 關于我們 財金網  版權所有  cfenews.com
日韩一卡二卡三卡四卡| 在线视频国内自拍亚洲视频| 日韩毛片一二三区| 一本久久综合亚洲鲁鲁五月天| 欧美日韩一区小说| 日韩激情视频在线| 黄色av免费在线观看| 国产黄a三级三级三级av在线看 | 少妇高潮一区二区三区99| 丝袜亚洲另类欧美| 老司机一区二区| 丁香婷婷综合色啪| 中文字幕一区二区不卡| 欧美曰成人黄网| 成年人视频在线网站| 日本福利专区在线观看| 日韩一级二级| 成人羞羞视频播放网站| 日韩成人精品在线| 日本一二三不卡| 欧美日韩久久不卡| 午夜免费啪视频观看视频| 影音先锋在线视频| а√中文在线天堂精品| 激情成人综合| 久久日韩精品一区二区五区| 亚洲超碰精品一区二区| 日韩福利视频在线观看| 黄色片免费在线观看| 精品国产欧美| 久久精选视频| 亚洲欧美一区二区久久| 亚洲精品在线观看网站| 国产三区在线观看| 天海翼精品一区二区三区| 免费黄网站欧美| 性做久久久久久| 在线免费观看色| 婷婷丁香久久| 久久亚洲影院| 亚洲va天堂va国产va久| 992tv在线观看| 精品久久亚洲| 久久精品久久99精品久久| 亚洲成人一区二区| 一二三在线视频社区| 亚洲日本va中文字幕| 久久97超碰国产精品超碰| 精品人伦一区二区三区蜜桃免费| 伊人网在线视频| 国产福利一区二区精品秒拍| 国产精品亚洲第一| 在线中文字幕一区| 调教视频免费在线观看| 精品理论电影在线| 2023国产精品| 亚洲人成五月天| 91国拍精品国产粉嫩亚洲一区 | 1234区在线观看视频免费| 日日夜夜天天综合入口| 欧美一区在线看| 洋洋av久久久久久久一区| 在线观看视频你懂的| 人人精品亚洲| 欧美国产日产图区| 一二三区高清| 国产免费久久| 综合在线观看色| 免费a在线观看| 欧美超碰在线| 一区二区三区四区蜜桃| 国产三级视频在线看| 国产韩国精品一区二区三区| 亚洲高清在线精品| 欧美激情视频在线播放| 一区二区激情| 欧美伦理视频网站| 日本一区免费网站| 丁香另类激情小说| 亚洲片国产一区一级在线观看| 日韩在线视频一区二区三区| 久久精品欧美日韩精品| 日韩一二三四| 亚洲先锋成人| 91精品国产一区二区| 四虎永久精品在线| 国产午夜亚洲精品不卡| 欧美理论在线观看| 在线免费观看欧美| 欧美一区二区三区不卡| 国产精一区二区| 国产亚洲精品超碰| 黄色片在线免费看| 亚洲毛片一区| 精品久久久久久久久久久久包黑料 | 超碰精品在线观看| 国产亚洲人成网站| 三区四区在线视频| 秋霞电影网一区二区| 亚洲国模精品一区| 美女精品一区最新中文字幕一区二区三区| 国产精品久久久久精k8 | 日韩电影在线一区| 日韩av综合中文字幕| 国产日产一区 | 欧美成人黑人| 91丨九色丨黑人外教| 久久伊伊香蕉| 日韩av中文字幕一区二区三区| 亚洲欧美www| 91九色精品国产一区二区| 欧美日韩国产另类一区| 一区二区三区国产好| 亚洲午夜精品网| 成人在线视频免费看| 一色屋精品亚洲香蕉网站| 成人日韩欧美| av午夜精品一区二区三区| av在线资源站| 国产成人三级在线观看| 国产精品毛片一区二区三区四区| 日本中文字幕一区二区有限公司| 992tv在线| 美女网站在线免费欧美精品| 91短视频在线| 美女一区二区三区在线观看| 日本v片在线免费观看| 蜜桃av一区二区| 久草在现在线| 国产精品自在欧美一区| av在线免费播放网站| 国产伦精品一区二区三区在线观看| 色视频在线看| 国产成人精品影院| caopen在线视频| 久久先锋资源网| 欧美少妇网站| 亚洲精品国产a| 国产精品免费精品自在线观看| 岛国av午夜精品| 一本色道久久综合亚洲精品酒店 | 欧美一级大片在线视频| 色综合久久久久综合体| 天天躁日日躁狠狠躁欧美| 日韩免费一区二区三区在线播放| 91综合久久一区二区| 亚洲欧美日韩综合| 日韩激情中文字幕| 成人三级黄色免费网站| 久久久美女毛片| **国产精品| 欧美一区二区在线观看| 亚洲美女91| 欧美性天天影视| 综合av第一页| 亚洲理论电影| 亚洲色图25p| 国产一区二区三区国产| 蜜桃麻豆av在线| 色综合天天狠狠| 亚洲精品久久| 色视频免费在线观看| 97久久超碰精品国产| 国产成人免费av一区二区午夜 | 精品免费视频.| 青青草97国产精品免费观看| av毛片在线免费看| 亚洲第一在线综合网站| 欧美大片aaaa| 日韩偷拍自拍| 亚洲欧美日韩一区二区| 国产在视频线精品视频www666| 米奇.777.com| 久久久精品免费免费| avtt综合网| 日本免费视频www| 91免费小视频| 欧美wwwsss9999| av福利导福航大全在线播放| 91啦中文在线观看| gogo人体一区| 黄动漫视频高清在线| 国产日产欧美一区二区三区| 女仆av观看一区| 尤物视频免费在线观看| 国产精品美女www爽爽爽| 精品一区av| 欧美日韩视频在线播放| 欧美日韩在线免费观看| 99精品热视频只有精品10| 丰乳肥臀在线| 欧美xfplay| gogo大胆日本视频一区| 欧美网色网址| avav免费在线观看| 欧美午夜精品久久久久久孕妇| 另类人妖一区二区av| 精品国产一区二| 瑟瑟在线观看| 在线国产电影不卡|