Kang et al., 2018 - Google Patents
An in-memory VLSI architecture for convolutional neural networksKang et al., 2018
- Document ID
- 15582969763910299992
- Author
- Kang M
- Lim S
- Gonugondla S
- Shanbhag N
- Publication year
- Publication venue
- IEEE Journal on Emerging and Selected Topics in Circuits and Systems
External Links
Snippet
This paper presents an energy-efficient and high throughput architecture for convolutional neural networks (CNN). Architectural and circuit techniques are proposed to address the dominant energy and delay costs associated with data movement in CNNs. The proposed …
- 230000001537 neural 0 title abstract description 14
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F7/00—Methods or arrangements for processing data by operating upon the order or content of the data handled
- G06F7/38—Methods or arrangements for performing computations using exclusively denominational number representation, e.g. using binary, ternary, decimal representation
- G06F7/48—Methods or arrangements for performing computations using exclusively denominational number representation, e.g. using binary, ternary, decimal representation using non-contact-making devices, e.g. tube, solid state device; using unspecified devices
- G06F7/52—Multiplying; Dividing
- G06F7/523—Multiplying only
- G06F7/53—Multiplying only in parallel-parallel fashion, i.e. both operands being entered in parallel
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F7/00—Methods or arrangements for processing data by operating upon the order or content of the data handled
- G06F7/38—Methods or arrangements for performing computations using exclusively denominational number representation, e.g. using binary, ternary, decimal representation
- G06F7/48—Methods or arrangements for performing computations using exclusively denominational number representation, e.g. using binary, ternary, decimal representation using non-contact-making devices, e.g. tube, solid state device; using unspecified devices
- G06F7/544—Methods or arrangements for performing computations using exclusively denominational number representation, e.g. using binary, ternary, decimal representation using non-contact-making devices, e.g. tube, solid state device; using unspecified devices for evaluating functions by calculation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
- G06N3/06—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons
- G06N3/063—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using electronic means
- G06N3/0635—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using electronic means using analogue means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/10—Complex mathematical operations
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/50—Computer-aided design
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F15/00—Digital computers in general; Data processing equipment in general
- G06F15/76—Architectures of general purpose stored programme computers
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F1/00—Details of data-processing equipment not covered by groups G06F3/00 - G06F13/00, e.g. cooling, packaging or power supply specially adapted for computer application
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11C—STATIC STORES
- G11C11/00—Digital stores characterised by the use of particular electric or magnetic storage elements; Storage elements therefor
- G11C11/21—Digital stores characterised by the use of particular electric or magnetic storage elements; Storage elements therefor using electric elements
- G11C11/34—Digital stores characterised by the use of particular electric or magnetic storage elements; Storage elements therefor using electric elements using semiconductor devices
- G11C11/40—Digital stores characterised by the use of particular electric or magnetic storage elements; Storage elements therefor using electric elements using semiconductor devices using transistors
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N99/00—Subject matter not provided for in other groups of this subclass
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Kang et al. | An in-memory VLSI architecture for convolutional neural networks | |
Kang et al. | A multi-functional in-memory inference processor using a standard 6T SRAM array | |
Kim et al. | Colonnade: A reconfigurable SRAM-based digital bit-serial compute-in-memory macro for processing neural networks | |
Valavi et al. | A 64-tile 2.4-Mb in-memory-computing CNN accelerator employing charge-domain compute | |
Gonugondla et al. | A variation-tolerant in-memory machine learning classifier via on-chip training | |
Long et al. | ReRAM-based processing-in-memory architecture for recurrent neural network acceleration | |
Miyashita et al. | A neuromorphic chip optimized for deep learning and CMOS technology with time-domain analog and digital mixed-signal processing | |
Giacomin et al. | A robust digital RRAM-based convolutional block for low-power image processing and learning applications | |
Yin et al. | Vesti: Energy-efficient in-memory computing accelerator for deep neural networks | |
Knag et al. | A 617-TOPS/W all-digital binary neural network accelerator in 10-nm FinFET CMOS | |
Kang et al. | A 19.4-nJ/decision, 364-K decisions/s, in-memory random forest multi-class inference accelerator | |
Kang et al. | Deep in-memory architectures in SRAM: An analog approach to approximate computing | |
Yue et al. | STICKER-IM: A 65 nm computing-in-memory NN processor using block-wise sparsity optimization and inter/intra-macro data reuse | |
Sarwar et al. | Energy efficient neural computing: A study of cross-layer approximations | |
Kang et al. | An energy-efficient memory-based high-throughput VLSI architecture for convolutional networks | |
Lou et al. | A mixed signal architecture for convolutional neural networks | |
Kang et al. | In-memory computing architectures for sparse distributed memory | |
Agrawal et al. | CASH-RAM: Enabling in-memory computations for edge inference using charge accumulation and sharing in standard 8T-SRAM arrays | |
Nasrin et al. | Mf-net: Compute-in-memory sram for multibit precision inference using memory-immersed data conversion and multiplication-free operators | |
Lin et al. | A review on SRAM-based computing in-memory: Circuits, functions, and applications | |
Seo et al. | On-chip sparse learning acceleration with CMOS and resistive synaptic devices | |
Cheon et al. | A 2941-TOPS/W charge-domain 10T SRAM compute-in-memory for ternary neural network | |
Bose et al. | A 51.3-TOPS/W, 134.4-GOPS in-memory binary image filtering in 65-nm CMOS | |
Angizi et al. | A near-sensor processing accelerator for approximate local binary pattern networks | |
Liu et al. | An energy-efficient mixed-bit CNN accelerator with column parallel readout for ReRAM-based in-memory computing |