Zhou et al., 2023 - Google Patents
- ️Sun Jan 01 2023
Zhou et al., 2023
View HTML-
Document ID
- 10500402014171978066 Author
- Chen J
- Li J
- Yang L
- Li Y
- Miao X Publication year
- 2023 Publication venue
- APL Machine Learning
External Links
Snippet
In-memory computing (IMC) using emerging nonvolatile devices has received considerable attention due to its great potential for accelerating artificial neural networks and machine learning tasks. As the basic concept and operation modes of IMC are now well established …
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
- G06N3/06—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons
- G06N3/063—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using electronic means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/30—Information retrieval; Database structures therefor; File system structures therefor
- G06F17/30861—Retrieval from the Internet, e.g. browsers
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/50—Computer-aided design
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F15/00—Digital computers in general; Data processing equipment in general
- G06F15/76—Architectures of general purpose stored programme computers
- G06F15/80—Architectures of general purpose stored programme computers comprising an array of processing units with common control, e.g. single instruction multiple data processors
- G06F15/8007—Architectures of general purpose stored programme computers comprising an array of processing units with common control, e.g. single instruction multiple data processors single instruction multiple data [SIMD] multiprocessors
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for programme control, e.g. control unit
- G06F9/06—Arrangements for programme control, e.g. control unit using stored programme, i.e. using internal store of processing equipment to receive and retain programme
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F7/00—Methods or arrangements for processing data by operating upon the order or content of the data handled
- G06F7/38—Methods or arrangements for performing computations using exclusively denominational number representation, e.g. using binary, ternary, decimal representation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/12—Computer systems based on biological models using genetic models
- G06N3/126—Genetic algorithms, i.e. information processing using digital simulations of the genetic system
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N99/00—Subject matter not provided for in other groups of this subclass
- G06N99/005—Learning machines, i.e. computer in which a programme is changed according to experience gained by the machine itself during a complete run
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F12/00—Accessing, addressing or allocating within memory systems or architectures
- G06F12/02—Addressing or allocation; Relocation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/62—Methods or arrangements for recognition using electronic means
- G06K9/6267—Classification techniques
- G06K9/6268—Classification techniques relating to the classification paradigm, e.g. parametric or non-parametric approaches
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Xu et al. | 2018 | Scaling for edge inference of deep neural networks |
Bavikadi et al. | 2020 | A review of in-memory computing architectures for machine learning applications |
US10346347B2 (en) | 2019-07-09 | Field-programmable crossbar array for reconfigurable computing |
CN112101534A (en) | 2020-12-18 | Reconfigurable memory compression techniques for deep neural networks |
Jawandhiya | 2018 | Hardware design for machine learning |
Welser et al. | 2018 | Future computing hardware for AI |
Shin et al. | 2019 | The heterogeneous deep neural network processor with a non-von Neumann architecture |
Jain et al. | 2022 | A heterogeneous and programmable compute-in-memory accelerator architecture for analog-ai using dense 2-d mesh |
Dutta et al. | 2022 | Hdnn-pim: Efficient in memory design of hyperdimensional computing with feature extraction |
Gebregiorgis et al. | 2023 | Tutorial on memristor-based computing for smart edge applications |
Zhou et al. | 2023 | Bring memristive in-memory computing into general-purpose machine learning: A perspective |
US20240289597A1 (en) | 2024-08-29 | Transformer neural network in memory |
Mao et al. | 2019 | MAX 2: An ReRAM-based neural network accelerator that maximizes data reuse and area utilization |
Nugent et al. | 2018 | Thermodynamic-RAM technology stack |
Burr | 2019 | A role for analogue memory in AI hardware |
Tsakyridis et al. | 2024 | Photonic neural networks and optics-informed deep learning fundamentals |
Huang et al. | 2023 | From memristive devices to neuromorphic systems |
EP4500397A1 (en) | 2025-02-05 | Two-dimensional mesh for compute-in-memory accelerator architecture |
Song et al. | 2019 | Rebnn: in-situ acceleration of binarized neural networks in reram using complementary resistive cell |
US11776944B2 (en) | 2023-10-03 | Discrete three-dimensional processor |
Guan et al. | 2019 | Recursive binary neural network training model for efficient usage of on-chip memory |
Raha et al. | 2023 | Efficient hardware acceleration of emerging neural networks for embedded machine learning: An industry perspective |
US11527523B2 (en) | 2022-12-13 | Discrete three-dimensional processor |
Zhan et al. | 2021 | Field programmable gate array‐based all‐layer accelerator with quantization neural networks for sustainable cyber‐physical systems |
Yantir | 2018 | Efficient acceleration of computation using associative in-memory processing |