輸入您購票時的 Email:

唐文力 - Neural-network Processing Unit Hardware-software Co-design IP - 2023 Taiwan AI Academy Conf

VIP / Speakers

« Back to List

Neural-network Processing Unit Hardware-software Co-design IP

Time / Place:

⏱️ 09/16 (Sat.) 11:30-12:00 at R1 - 1st Conference Room

Abstract:

Large language models (LLMs) have become an increasingly prominent feature in various computing platforms, such as data centers, smartphones, and microcontrollers. As such, their potential applications in both human-to-machine interfaces (HMIs) and machine-to-machine interfaces (M2MI) have been extensively studied and discussed. However, implementing hyper-scale LLMs on heterogeneous multicore System-on-Chips (SoCs) poses significant challenges. In this talk, we will discuss the key challenges, including issues related to hardware/software interfaces, verification, and SoC architecture exploration. At the end of this talk, we will propose a practical workflow and corresponding solutions we’ve applied in many cases. And see how our standard tool - ONNC - is used in the workflow.

大型語言模型(LLM)已成為各種計算平台(例如資料中心、智慧型手機和微控制器)中日益突出的功能。因此,它們在人機介面(HMI)和機器對機器介面(M2MI)中的潛在應用已被廣泛研究和討論。然而,在異質多核晶片上系統(SoC)上實現大型語言模型(LLM)面臨著重大挑戰。在本次演講中,我們將討論關鍵挑戰,包括與硬體 / 軟體介面、驗證和 SoC 架構探索相關的問題。在本次演講的最後,我們將提出一個實用的架構以及我們在許多案例中應用的解決方案。並了解我們的標準工具 - ONNC - 如何在架構中使用。

😊 Share this page to friends:

Biography:

唐文力
  • 唐文力 Luba Tang
  • Skymizer / CEO
  • Luba Tang is the founder and CEO of Skymizer Taiwan Inc., which is in the business of providing system software to IC design teams. Skymizer’s system software solutions enable AI-on-Chip design houses to automate AI application development, improve system performance, and optimize inference accuracy. Luba Tang’s research interests include electronic system level (ESL) design, system software, and neural networks. He had focused on iterative compilers, ahead-of-time compilers, link-time optimization, neural network compilation, and neural network optimization. His most recent work focuses on exploiting various types of parallelism from different accelerators in a hyper-scale system-on-chip.

😊 Share this page to friends:

😊 Share this page to friends: