Tutorials

Main Tutorials (Wednesday, October 6, 2021)
[Main Tutorial 1] 13:00~14:30
On-Device Artificial Intelligence (AI) Accelerators: Trends & Applications

Dr. Ashutosh Mishra
Yonsei Institute of Convergence Technology, Yonsei University, Korea

Biography Abstract

Biography
Dr. Ashutosh Mishra had received his Bachelor of Technology in Electronics & Communication Engineering in 2008 from Uttar Pradesh Technical University, India. He had completed his Master of Technology in Microelectronics & VLSI design from the National Institute of Technology Allahabad, India, In 2011. Dr. Mishra has awarded a Ph.D. in Electronics Engineering at the Indian Institute of Technology (BHU) Varanasi, India, in 2018. He has worked as an Assistant professor in Electronics & Communication Engineering at the National Institute of Technology Raipur, India. Dr. Mishra has received the Korea Research Fellowship (KRF)-2019 provided by the National Research Foundation of Korea through the Ministry of Science & ICT, South Korea. Currently, he is working as a Research faculty & Brain Pool Fellow in the Seamless Transportation Lab at the Yonsei Institute of Convergence Technology, Yonsei University, South Korea. His research interests include Smart sensors, Intelligent systems, Autonomous vehicles, and Artificial Intelligence.
Abstract
Intelligence is moving towards edge devices. Increased computing power, sensor data, and improved artificial intelligence (AI) algorithms are the driving trends towards the on-device AI. A few examples are smartphones, automobiles, robots, drones. Therefore, on-device AI has been an emerging system on chip (SoC) technology. The on-device AI has remarkably implemented the connected devices, including automobiles, HD cameras, smartphones, wearables, and other IoT devices. In addition, the on-device inference has lower latency and increased privacy as compared to the cloud-based paradigm. However, such intelligent devices require efficient hardware. Because such compute-intensive tasks on the smaller devices require optimization of the computation power and energy consumption. This tutorial will focus on the overview, challenges, and multifaceted applications of the AI accelerators in the SoC paradigm. In tutorial session #1, we will describe the training and inference requirements of the on-device AI accelerators. In tutorial session #2, we will provide insight into the AI accelerators for future autonomous vehicles (FAVs).
On-Device AI Accelerators for Autonomous Vehicles

Prof. Shiho Kim
School of Integrated Technology, Yonsei University, Korea

Biography Abstract

Biography
Prof. Kim is a full-Professor in the School of Integrated Technology, Yonsei University, South Korea. He received his B.S. degree in Electronic Engineering from Yonsei University, South Korea in 1986. He received his M.S. and Ph.D. degrees in Electrical Engineering in 1988 and 1995, respectively from KAIST, South Korea. His main research interest includes the development of software and hardware technologies for intelligent vehicles and reinforcement learning for autonomous vehicles. He is a member of the editorial board and reviewer for various Journals and International conferences. So far, he has organized two International conferences as Technical Chair/General Chair. He is a member of IEIE (Institute of Electronics and Information Engineers of Korea), KSAE (Korean Society of Automotive Engineers), vice president of KINGC (Korean Institute of Next Generation Computing), and a senior member of IEEE. He is the co-author for over 100 papers and holding more than 50 patents in the field of artificial intelligence, autonomous vehicles and Hardware accelerators.
Abstract
In tutorial session #1-2 will focus on emerging trends and technologies in AI acceleration for Future Autonomous Vehicles. Here, we will describe the training and inference requirements of the on-board hardware (HW) accelerators as well as challenges, requirements, and emerging trends in AI accelerators for future autonomous vehicles (FAVs).
[Main Tutorial 2] 14:30~16:00
Statistical modeling-based simulator for analog neural networks

Chaeun Lee
Clova OCR, NAVER, Korea

Biography Abstract

Biography
Chaeun Lee received the B.S. and M.S. degrees in Electrical and Computer Engineering from Seoul National University (SNU) in 2018 and 2020. From 2020 to 2021, he worked as researcher at Inter-university Semiconductor Research Center (ISRC) and Pohang University of Science and Technology (POSTECH). His research topics was spread from modeling lightweight neural networks including Binarized Neural Networks (BNN) which are robust to noise and variations to co-optimizing neural networks for edge devices with emerging memory devices. He is currently working at NAVER Clova as a machine learning engineer. His recent research interests include overcoming issues in training neural networks with imbalanced data and out of sample detection.
Abstract
Growing interests in designing neural networks with analog memory devices and circuits has been ignited, recently. Accurate simulations to verify the performance of the embedded neural networks are required before fabricating a chip. However, deep neural networks are composed of a ton of parameters such as weights and biases, and it leads to simulating a whole system with classical circuit simulators impossible. To cope with the difficulty, functional modeling-based simulators using high-level programming languages have been proposed. However, there exist unexpected errors in approximated simulation results compared to those of classical circuit simulators. In addition, they can only support a limited range of memory devices and circuits. Motivated from these limitations, statical modeling-based simulators has been suggested. They can emulate the unexpected dynamics of the system as they have the ability to represent non-analytic dynamics of memory devices and circuits. In this talk, we introduce various types of existing simulation methods and challenges for designing statistical modeling-based methods in relation to promising deep learning models such as physics-informed neural networks and Neural ODE.
[Main Tutorial 3] 16:00~17:30
Design a DLL Easily Using Only Standard Cells for Clock Synchronization in A Heterogeneous Multi-Die IC

Prof. Shi-Yu Huang
Electrical Engineering, National Tsing Hua University, Taiwan

Biography Abstract

Biography
Shi-Yu Huang received his Ph.D. degree in Electrical and Computer Engineering from University of California, Santa Barbara, in 1997. Since 1999, he has joined National Tsing Hua University, Taiwan until now. His recent research is concentrated on all-digital timing circuit designs, such as all-digital phase-locked loop (PLL), all-digital delay-locked loop (DLL), time-to-digital converter (TDC), and their applications to parametric fault testing and reliability enhancement for 3D-ICs. He has published more than 160 technical papers (including 45 IEEE journal papers). Dr. Huang ever co-founded a company in 2007-2012, TinnoTek Inc., specializing a cell-based PLL compiler and system-level power estimation tools. He has served as a tutorial speaker in several IEEE conferences and received the best-presentation award or best-paper award for 5 times, (e.g., VLSI-DAT’2006, VLSI-DAT’2013, ATS’2014, WRTLT’2017, ISOCC’2018). He is now a senior member of IEEE.
Abstract
When we design an SoC or a multi-die IC consisting of 3rd-party IPs, heterogeneous components, or functional dice, synchronization of the clock signals across all of them could be a headache. Fortunately, Delay-Locked Loop (DLL) comes to the rescue. However, a DLL is traditionally built with some analog circuitry inside and thus making the design process complicated if not mysterious for system integrators. The emergence of cell-based DLL design style over the past two decades has alleviated this problem greatly. A cell-based DLL design is not only small and easy-to-design, but also robust to the process and temperature variation. Also, it could lend itself to automation as a DLL compiler and so one can generate a DLL instance on the push of a button. In this tutorial, we will take on a step-by-step journey to show you how to make your own DLL easily using only standard cells. Specific topics such as phase detector, tunable delay line, phase-locking procedure, coping strategy for process and temperature variation, and duty-cycle correction will be covered in detail.
Short Tutorials (Friday, October 8, 2021)
[Short Tutorial 1] 13:00~13:40
Posit Arithmetic: An Efficient Computing Paradigm for Deep Learning Acceleration

Prof. Seok-Bum Ko
Department of Electrical and Computer Engineering, University of Saskatchewan, Canada

Prof. Hao Zhang
Faculty of Information Science and Engineering, Ocean University of China, China

Biography Abstract

Biography
Dr. Seok-Bum Ko is currently a Professor at the Department of Electrical and Computer Engineering and the Division of Biomedical Engineering, University of Saskatchewan, Canada. He got his PhD degree from the University of Rhode Island, USA in 2002. His research interests include computer architecture, computer arithmetic, efficient hardware implementation of compute-intensive applications, deep learning processor architecture and biomedical engineering. He is a senior member of IEEE Circuits and Systems Society, a member of IEEE VLSI Systems and Application Technical Committee and associate editors of IEEE Transactions on Circuits and Systems I and IEEE Access.

Dr. Hao Zhang is currently working as an Associate Professor in the Faculty of Information Science and Engineering at the Ocean University of China, Qingdao, China. He received his PhD degree from the University of Saskatchewan, Canada in 2019. He did his postdoctoral research experience at the University of Saskatchewan in Canada. His research interests include computer arithmetic, computer architecture, reconfigurable computing, AI hardware processor, and deep learning applications. He had published multiple research papers on Posit arithmetic units and their applications in deep learning. He is a member of IEEE Circuits and Systems Society and a member of IEEE VLSI Systems and Application Technical Committee.

Abstract
Posit is designed as an alternative to IEEE 754 floating-point format for many applications. With the same bit-width, it can provide a much larger dynamic range than IEEE floating-point format. In addition, posit has non-uniformed number distribution. This distribution fits well with the data distribution of deep learning and signal processing applications. Due to these advantages, posit format is nowadays popular in these applications. Many posit based arithmetic units and posit based deep learning accelerator have been proposed in the literature. However, due to the dynamic component bit-width, the cost of posit computation is still higher than the floating-point counterpart. There are many optimizations that can potentially be investigated to reduce its hardware cost. In this tutorial, the posit format and posit based arithmetic units will be first discussed. Then, several state-of-the-art posit based deep learning accelerator designs will be presented. And finally, the challenges and future research directions will be discussed to motivate more related research works.
[Short Tutorial 2] 13:00~13:40
Neuromorphic Circuits and Systems for Applications in Edge Intelligence

Prof. Jong-Hyeok Yoon
Information and Communication Engineering, DGIST, Korea

Biography Abstract

Biography
Jong-Hyeok Yoon received B.S., M.S., and Ph.D. degrees in Electrical Engineering from the Korea Advanced Institute of Science and Technology (KAIST), Daejeon, Republic of Korea, in 2012, 2014, and 2018, respectively. From 2018 to 2020, He was a postdoctoral fellow at the Georgia Institute of Technology, Atlanta, GA, USA. In 2021, he joined the Daegu Gyeongbuk Institute of Science and Technology (DGIST), Daegu, Republic of Korea, where he is currently an Assistant Professor with the Department of Information and Communication Engineering. His research interests include non-volatile memory-based processing-in-memory architectures for deep learning, neuromorphic circuits for edge intelligence, high-speed wireline communications, and mixed-signal circuit designs. He was a recipient of the best regular paper award at the IEEE Custom Integrated Circuits Conference (CICC) in 2021.
Abstract
Energy-efficient computing architectures are imperative to employ AI systems at battery-powered edge devices. Although von Neumann architecture can support various tasks, it requires excessive power consumption to conduct massive matrix-vector multiplications (MVM) for AI systems, thereby precluding applications in edge intelligence. As a solution to the strict power budget at edge devices, neuromorphic computing and computing-in-memory (CIM) architectures have gained importance. In this talk, I will introduce approaches in neuromorphic computing and CIM architectures for edge intelligence. In particular, I will discuss a neuromorphic simultaneous localization and mapping (SLAM) accelerator employing oscillator-based pose-cells inspired by spatial recognition of rodents, and a resistive RAM (RRAM)-based CIM architecture surmounting technological challenges of RRAM.