Lita Yang

From Murmann Mixed-Signal Group

(Difference between revisions)
Jump to: navigation, search
(One intermediate revision not shown)
Line 7: Line 7:
Admitted to Ph.D. Candidacy: 2013-2014  
Admitted to Ph.D. Candidacy: 2013-2014  
-
'''Email''': [mailto:yanglita@stanford.edu yanglita AT stanford DOT edu]<br> '''Research''': Energy-efficient Memory and Communication Design for Error Tolerant Machine Learning Algorithms&nbsp;<br>  
+
'''Email''': [mailto:yanglita@stanford.edu yanglita AT stanford DOT edu]<br>  
-
<br> As transistor scaling is coming to a halt, systems today are becoming more and more power limited. Given recent trends in increasing network sizes and the need to process more data (such as Deep Learning and Big Data applications), the cost to store and move data around in a system can far exceed computation costs, prohibiting hardware implementations of machine learning algorithms in embedded applications.&nbsp;<br><br> Recently, there has been an emergence of interest in the field of Approximate Computing, which explores the performance (accuracy) of an algorithm with reduced precision. Convolutional Neural Networks (ConvNets), the current top performing image classification networks, are an example of a class of stochastic algorithms which can tolerate reduced precision for little degradation in algorithmic performance.&nbsp;We propose to reduce the system energy by exploiting error tolerance of the algorithm using approximate memory and interconnect communication design. From a memory designer’s perspective, this is rarely considered a viable option since most general purpose systems require robust storage and communication.
+
'''Research''': Approximate Memory for Energy-Efficient Machine Learning Algorithms&nbsp;  
-
From simulation, we have shown that ConvNets are tolerant to bit flips and reduction in precision [1]. To accurately quantify how allowing higher bit error rates (BERs) with low voltage memory affects classification error rate in ConvNets, a test chip has been taped out to perform stress testing on low leakage single-port SRAMs. The results from this measurement will demonstrate dynamic and leakage energy savings from operating memory at lowest possible voltage subject to a classification error rate.<br>&nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;&nbsp;[[Image:HILMemory.png]]<br>  
+
<br> As transistor scaling is coming to a halt, systems today are becoming more and more power limited. Given recent trends in increasing network sizes and the need to process more data (such as Deep Learning and Big Data applications), the cost to store and move data around in a system can far exceed computation costs, prohibiting hardware implementations of machine learning algorithms in embedded applications.
 +
 
 +
Recently, there has been an emergence of interest in the field of Approximate Computing, which explores the performance (accuracy) of an algorithm with reduced precision. Convolutional Neural Networks (ConvNets), the current top performing image classification networks, are an example of a class of stochastic algorithms which can tolerate reduced precision for little degradation in algorithmic performance.&nbsp;We propose to reduce the system energy by exploiting error tolerance of the algorithm using approximate memory. From a memory designer’s perspective, this is rarely considered a viable option since most general purpose systems require robust storage and communication.
 +
 
 +
We have shown that ConvNets are tolerant to bit flips and reduction in precision [1]. To accurately quantify the effectiveness of accepting bit errors under reduced memory supply voltages during ConvNet inference and training, we took measurements on an 8KB SRAM test chip in 28nm UTBB FD-SOI CMOS for emulating memory bit errors at low voltages [2]. The results demonstrate supply voltage reduction of 310mV on a MNIST ConvNet, resulting in 5.4x leakage power savings and 2.9x memory access power savings at 99% of floating-point classification accuracy, with no additional hardware cost.  
 +
 
 +
<br>&nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;&nbsp;[[Image:HILMemory.png]]<br>  
<br>  
<br>  
-
[1] B. Murmann, D. Bankman, E. Chai, D. Miyashita, and L. Yang, "Mixed-Signal Circuits for Embedded Machine-Learning Applications," Asilomar Conference on Signals, Systems and Computers, Asilomar, CA, Nov. 2015.
+
[1] B. Murmann, D. Bankman, E. Chai, D. Miyashita, and L. Yang, "Mixed-Signal Circuits for Embedded Machine-Learning Applications," Asilomar Conference on Signals, Systems and Computers, Asilomar, CA, Nov. 2015.
 +
 
 +
[2] L. Yang and B. Murmann, "SRAM Voltage Scaling for Energy-Efficient Convolutional Neural Networks," International Symposium on Quality Electronic Design (ISQED), Santa Clara, CA, Mar. 2017, pp. 7-12.

Revision as of 07:51, 19 July 2017

LitaYang.jpg

BSEE, California Institute of Technology, 2012

MSEE, Stanford University, 2015

Admitted to Ph.D. Candidacy: 2013-2014

Email: yanglita AT stanford DOT edu

Research: Approximate Memory for Energy-Efficient Machine Learning Algorithms 


As transistor scaling is coming to a halt, systems today are becoming more and more power limited. Given recent trends in increasing network sizes and the need to process more data (such as Deep Learning and Big Data applications), the cost to store and move data around in a system can far exceed computation costs, prohibiting hardware implementations of machine learning algorithms in embedded applications.

Recently, there has been an emergence of interest in the field of Approximate Computing, which explores the performance (accuracy) of an algorithm with reduced precision. Convolutional Neural Networks (ConvNets), the current top performing image classification networks, are an example of a class of stochastic algorithms which can tolerate reduced precision for little degradation in algorithmic performance. We propose to reduce the system energy by exploiting error tolerance of the algorithm using approximate memory. From a memory designer’s perspective, this is rarely considered a viable option since most general purpose systems require robust storage and communication.

We have shown that ConvNets are tolerant to bit flips and reduction in precision [1]. To accurately quantify the effectiveness of accepting bit errors under reduced memory supply voltages during ConvNet inference and training, we took measurements on an 8KB SRAM test chip in 28nm UTBB FD-SOI CMOS for emulating memory bit errors at low voltages [2]. The results demonstrate supply voltage reduction of 310mV on a MNIST ConvNet, resulting in 5.4x leakage power savings and 2.9x memory access power savings at 99% of floating-point classification accuracy, with no additional hardware cost.


                    HILMemory.png


[1] B. Murmann, D. Bankman, E. Chai, D. Miyashita, and L. Yang, "Mixed-Signal Circuits for Embedded Machine-Learning Applications," Asilomar Conference on Signals, Systems and Computers, Asilomar, CA, Nov. 2015.

[2] L. Yang and B. Murmann, "SRAM Voltage Scaling for Energy-Efficient Convolutional Neural Networks," International Symposium on Quality Electronic Design (ISQED), Santa Clara, CA, Mar. 2017, pp. 7-12.

Personal tools