Article count:1933 Read by:13229860

Account Entry

Intel + Baidu Smart Cloud work together to accelerate the intelligent transition in the AI ​​era

Latest update time:2024-09-27
    Reads:

At the 2024 Baidu Cloud Intelligence Conference held on September 25-26, 2024, as the co-organizer of the conference, Intel brought a full-stack AI hardware and software solution, and explained in depth how to efficiently deploy and run large language models based on Intel® Xeon® processors and a new generation of cost-effective accelerator cards. It also demonstrated a series of optimization strategies and an open architecture AI software stack (OPEA) for enterprise customers to help enterprises accelerate the implementation of AI and maximize the computing performance of the entire AI infrastructure.


For a long time, Baidu Smart Cloud and Intel have been cooperating deeply in products and technologies, and have achieved remarkable results in cloud computing, big data, PaddlePaddle deep learning framework optimization, vehicle-road collaboration, edge computing and other fields.


Now, follow the editor's steps and take a look at the wonderful highlights of the event!


Intel made a grand appearance at the main forum of the conference, expounding on the business insights and technology prospects of "AI-driven industry core power"



Main Forum Speech

AI drives industry "core power"



"To embrace the new productivity represented by AI, Intel and Baidu have worked together to build modern infrastructure for cloud data centers, continuously optimize software, and deepen their efforts in areas such as green and sustainable development. Based on years of cooperation, the two parties will continue to create diversified solutions through continuous technological innovation in the future to provide customers from all walks of life with better services and experience."


—— Liang Yali

Vice President of Intel Marketing Group

General Manager of Cloud and Industry Solutions and Data Center Sales, China



After the main forum, Intel's special session focused on how to inject "core power" into the AI ​​industry. From advanced technologies and products to cooperative solutions, it comprehensively introduced Intel's AI product hardware and software combination, as well as the latest cooperation results with Baidu Smart Cloud Qianfan Big Model Platform.



Intel Special



  • Opening Remarks

    Hu Kai, Internet Industry Director, Intel Cloud and Industry Solutions Group

    Different industries have diverse demands for AI computing power, and their requirements for processor frequency, bandwidth, etc. are different. From cloud to edge to end, Intel continues to provide a comprehensive product technology portfolio for the AI ​​industry, and cooperates with a wide range of partners to provide end customers with more economical and diversified computing power support.


  • CPU-based large language model reasoning - Baidu Smart Cloud Qianfan large model platform implementation

    Chen Xiaoyu Baidu Intelligent Cloud Senior R&D Engineer

    The widespread application of big models in various industries has driven a new round of industrial revolution, and also posed severe challenges to AI computing power. As a leading domestic artificial intelligence cloud service platform, Baidu Smart Cloud Qianfan Big Model Platform provides developers with a rich selection of big models and supporting software tools for model and application development, which can help users build various intelligent applications.

    In order to improve the CPU-based LLM reasoning performance, Baidu Smart Cloud introduced the Intel® Xeon® Scalable Processor and used the built-in Intel® Advanced Matrix Extensions Technology, combined with the large model reasoning acceleration solution xFasterTransformer (xFT), to help users achieve CPU-based LLM reasoning acceleration on the Qianfan Large Model Platform . At present, the Qianfan Large Model Platform has output excellent application cases in multiple scenarios such as education, office, and medical care .


  • xFT unlocks Xeon® computing power and unleashes AI potential

    Miao Jincheng, Senior Software Engineer at Intel


    Intel is committed to delivering complete hardware and software solutions in the field of AI. The Xeon® Scalable Processor equipped with the Intel® AI Engine comprehensively improves the performance of AI applications out of the box. The fourth and fifth generation Intel® Xeon® Scalable Processor has an AI accelerator built into it, the Intel® Advanced Matrix Extensions Engine, which consists of a 2D register file (TILE) and TMUL, supports two data types , INT8 and BF16, and can effectively accelerate deep learning training and reasoning workloads.

    As a fully optimized open source LLM reasoning framework, xFT supports multi-machine cross-node distributed operation and supports two API interfaces, C++ and Python, making it easier for users to use and integrate it into their own business framework. At the same time, xFT supports a variety of low-precision data types and a variety of mainstream large models on the market.



  • Full-stack implementation of the enterprise GenAI open platform based on retrieval-enhanced generation

    Guo Bin, Senior Cloud Computing Software Architect at Intel

    Although the RAG technology based on retrieval enhancement is not new, its application potential in enterprises is becoming increasingly prominent as the capabilities of large language models are enhanced. Enterprises face both challenges and opportunities in the implementation of AI. Especially after the generative AI technology exploded the market, the focus of the industry has shifted from model pre-training to specific application implementation.

    OPEA (Open Platform for Enterprise AI) is an open source project initiated by Intel and donated to the Linux Foundation. It aims to build an open AI software ecosystem, use generative AI to help enterprises tap into the value of data, effectively help their own business development, and reduce the complexity of the ecosystem to achieve solution scale.

  • Cost-effective solution for large models

    Mu Yanfeng, Software Technology Manager, Intel


    The development of large models is in full swing. With the rapid growth of model parameter scale, enterprises are increasingly in need of improving computing power. To effectively explore the value of large models, enterprises need to improve system scalability while avoiding single GPU lock-in, and effectively control costs while improving business efficiency.

    Intel® Gaudi ® 2 AI accelerator is designed for generative AI and large models, and aims to provide high-performance and efficient generative AI computing capabilities. As a fully programmable high-performance Al accelerator, it integrates many technological innovations, has high memory bandwidth/capacity and horizontal expansion capabilities based on standard Ethernet technology. At the same time, Intel provides an end-to-end AI software stack with the Intel® Gaudi ® Software Suite as the core, which puts the development and deployment of AI models on the fast track.


In addition to the Intel session, Intel technical experts also shared valuable insights on big models and cloud native topics.



Special Forum on Large Model Platform Technology Practice



  • Unlocking the potential of Xeon® —a new choice for large model reasoning

    Guo Bin, Senior Cloud Computing Software Architect at Intel


    The computing power demand caused by generative AI continues to increase. Faced with problems such as GPU shortages and high prices, companies need to find new options that can continue to provide sufficient computing power for generative AI and have both scalability and availability .

    Facing the AI ​​era, the Intel® Xeon® platform continues to update and iterate, adding AI acceleration engines in addition to the continuous improvement of computing power. Using the Intel® Advanced Matrix Extensions provided by Intel® Xeon® Scalable Processors, users can more fully tap the potential of the CPU and more easily obtain the performance required for AI workloads.



Cloud Native Forum



  • Data security in the era of big models: the key role of Xeon® processors and confidential computing

    Song Chuan, Chief Engineer of Intel Data Center and Artificial Intelligence Division


    In the era of big models, computing power demand is critical, but data security and circulation cannot be ignored. With the continuous expansion of big model applications in various industries, the amount of data is growing explosively. It is crucial to achieve orderly data circulation and strengthen data security protection to promote the continuous advancement of big model technology.

    Intel® Xeon® Scalable processors are equipped with a variety of security engines that provide hardware -based confidential computing solutions and better protect data privacy while making full use of data. Among them, Intel® SGX can provide application-level security isolation, while Intel® TDX can provide virtual machine-level security isolation .


Finally, follow the editor’s footsteps to visit the Intel exhibition area “online” and appreciate the charm of the site!


Intel exhibited a wide range of products and solutions around the three hot topics and products of large model optimization and deployment, the latest Intel® Xeon® 6 processor , and AI PC .


<
Slide to view the next picture
>


Large model optimization and deployment

  • Unlock the potential of Xeon® , a new choice for Qianfan reasoning

  • Secure and Trusted LLM Reasoning Service: Application of Intel® TDX in Heterogeneous Computing

  • LLM Reasoning Solution Based on Arc TM GPU


Intel® Xeon® 6

  • Intel® Xeon® 6 processors provide extreme performance


AI PC

  • OpenVINOTM Toolkit and Intel AI PC Accelerate AIGC Technology and Applications

  • AI PC to play Black Myth: Wukong and various AI applications


Relying on more than ten years of technical cooperation, Intel will continue to work with Baidu Smart Cloud to jointly promote generative artificial intelligence (GenAI) technology innovation and industry implementation through full-stack computing power, powerful AI acceleration capabilities, and a complete end-to-end cooperation ecosystem, turning technical capabilities into productivity and serving the intelligent construction of thousands of industries!


For more information about Intel® Xeon® solutions for AI acceleration, please click " Read original article " to learn more



© Intel Corporation. Intel, the Intel logo, and other Intel trademarks are trademarks of Intel Corporation or its subsidiaries in the U.S. and/or other countries. Other names and brands may be claimed as the property of others.

*Other names and trademarks may be claimed as the property of their respective owners


Want to see more "core" information

Tell us with your Likes and Watching ~




Intel and Baidu Smart Cloud continue to optimize LLM reasoning algorithms and implementations

Intel and Baidu join forces to advance the future of generative artificial intelligence

Intel Debuts at Baidu Cloud Intelligence Conference to Discuss the Era of Artificial Intelligence



Latest articles about

 
EEWorld WeChat Subscription

 
EEWorld WeChat Service Number

 
AutoDevelopers

About Us Customer Service Contact Information Datasheet Sitemap LatestNews

Room 1530, Zhongguancun MOOC Times Building,Block B, 18 Zhongguancun Street, Haidian District,Beijing, China Tel:(010)82350740 Postcode:100190

Copyright © 2005-2024 EEWORLD.com.cn, Inc. All rights reserved 京ICP证060456号 京ICP备10001474号-1 电信业务审批[2006]字第258号函 京公网安备 11010802033920号