Buy Artificial Intelligence Hardware Design by Oscar Ming Kin Law
Book 1
Book 2
Book 3
Book 1
Book 2
Book 3
Book 1
Book 2
Book 3
Book 1
Book 2
Book 3
Home > Computing and Information Technology > Computer science > Artificial intelligence > Artificial Intelligence Hardware Design: Challenges and Solutions
Artificial Intelligence Hardware Design: Challenges and Solutions

Artificial Intelligence Hardware Design: Challenges and Solutions


     0     
5
4
3
2
1



International Edition


X
About the Book

ARTIFICIAL INTELLIGENCE HARDWARE DESIGN Learn foundational and advanced topics in Neural Processing Unit design with real-world examples from leading voices in the field In Artificial Intelligence Hardware Design: Challenges and Solutions, distinguished researchers and authors Drs. Albert Chun Chen Liu and Oscar Ming Kin Law deliver a rigorous and practical treatment of the design applications of specific circuits and systems for accelerating neural network processing. Beginning with a discussion and explanation of neural networks and their developmental history, the book goes on to describe parallel architectures, streaming graphs for massive parallel computation, and convolution optimization. The authors offer readers an illustration of in-memory computation through Georgia Tech’s Neurocube and Stanford’s Tetris accelerator using the Hybrid Memory Cube, as well as near-memory architecture through the embedded eDRAM of the Institute of Computing Technology, the Chinese Academy of Science, and other institutions. Readers will also find a discussion of 3D neural processing techniques to support multiple layer neural networks, as well as information like: A thorough introduction to neural networks and neural network development history, as well as Convolutional Neural Network (CNN) models Explorations of various parallel architectures, including the Intel CPU, Nvidia GPU, Google TPU, and Microsoft NPU, emphasizing hardware and software integration for performance improvement Discussions of streaming graph for massive parallel computation with the Blaize GSP and Graphcore IPU An examination of how to optimize convolution with UCLA Deep Convolutional Neural Network accelerator filter decomposition Perfect for hardware and software engineers and firmware developers, Artificial Intelligence Hardware Design is an indispensable resource for anyone working with Neural Processing Units in either a hardware or software capacity.

Table of Contents:
Author Biographies xi Preface xiii Acknowledgments xv Table of Figures xvii 1 Introduction 1 1.1 Development History 2 1.2 Neural Network Models 4 1.3 Neural Network Classification 4 1.3.1 Supervised Learning 4 1.3.2 Semi-supervised Learning 5 1.3.3 Unsupervised Learning 6 1.4 Neural Network Framework 6 1.5 Neural Network Comparison 10 Exercise 11 References 12 2 Deep Learning 13 2.1 Neural Network Layer 13 2.1.1 Convolutional Layer 13 2.1.2 Activation Layer 17 2.1.3 Pooling Layer 18 2.1.4 Normalization Layer 19 2.1.5 Dropout Layer 20 2.1.6 Fully Connected Layer 20 2.2 Deep Learning Challenges 22 Exercise 22 References 24 3 Parallel Architecture 25 3.1 Intel Central Processing Unit (CPU) 25 3.1.1 Skylake Mesh Architecture 27 3.1.2 Intel Ultra Path Interconnect (UPI) 28 3.1.3 Sub Non-unified Memory Access Clustering (SNC) 29 3.1.4 Cache Hierarchy Changes 31 3.1.5 Single/Multiple Socket Parallel Processing 32 3.1.6 Advanced Vector Software Extension 33 3.1.7 Math Kernel Library for Deep Neural Network (MKL-DNN) 34 3.2 NVIDIA Graphics Processing Unit (GPU) 39 3.2.1 Tensor Core Architecture 41 3.2.2 Winograd Transform 44 3.2.3 Simultaneous Multithreading (SMT) 45 3.2.4 High Bandwidth Memory (HBM2) 46 3.2.5 NVLink2 Configuration 47 3.3 NVIDIA Deep Learning Accelerator (NVDLA) 49 3.3.1 Convolution Operation 50 3.3.2 Single Data Point Operation 50 3.3.3 Planar Data Operation 50 3.3.4 Multiplane Operation 50 3.3.5 Data Memory and Reshape Operations 51 3.3.6 System Configuration 51 3.3.7 External Interface 52 3.3.8 Software Design 52 3.4 Google Tensor Processing Unit (TPU) 53 3.4.1 System Architecture 53 3.4.2 Multiply–Accumulate (MAC) Systolic Array 55 3.4.3 New Brain Floating-Point Format 55 3.4.4 Performance Comparison 57 3.4.5 Cloud TPU Configuration 58 3.4.6 Cloud Software Architecture 60 3.5 Microsoft Catapult Fabric Accelerator 61 3.5.1 System Configuration 64 3.5.2 Catapult Fabric Architecture 65 3.5.3 Matrix-Vector Multiplier 65 3.5.4 Hierarchical Decode and Dispatch (HDD) 67 3.5.5 Sparse Matrix-Vector Multiplication 68 Exercise 70 References 71 4 Streaming Graph Theory 73 4.1 Blaize Graph Streaming Processor 73 4.1.1 Stream Graph Model 73 4.1.2 Depth First Scheduling Approach 75 4.1.3 Graph Streaming Processor Architecture 76 4.2 Graphcore Intelligence Processing Unit 79 4.2.1 Intelligence Processor Unit Architecture 79 4.2.2 Accumulating Matrix Product (AMP) Unit 79 4.2.3 Memory Architecture 79 4.2.4 Interconnect Architecture 79 4.2.5 Bulk Synchronous Parallel Model 81 Exercise 83 References 84 5 Convolution Optimization 85 5.1 Deep Convolutional Neural Network Accelerator 85 5.1.1 System Architecture 86 5.1.2 Filter Decomposition 87 5.1.3 Streaming Architecture 90 5.1.3.1 Filter Weights Reuse 90 5.1.3.2 Input Channel Reuse 92 5.1.4 Pooling 92 5.1.4.1 Average Pooling 92 5.1.4.2 Max Pooling 93 5.1.5 Convolution Unit (CU) Engine 94 5.1.6 Accumulation (ACCU) Buffer 94 5.1.7 Model Compression 95 5.1.8 System Performance 95 5.2 Eyeriss Accelerator 97 5.2.1 Eyeriss System Architecture 97 5.2.2 2D Convolution to 1D Multiplication 98 5.2.3 Stationary Dataflow 99 5.2.3.1 Output Stationary 99 5.2.3.2 Weight Stationary 101 5.2.3.3 Input Stationary 101 5.2.4 Row Stationary (RS) Dataflow 104 5.2.4.1 Filter Reuse 104 5.2.4.2 Input Feature Maps Reuse 106 5.2.4.3 Partial Sums Reuse 106 5.2.5 Run-Length Compression (RLC) 106 5.2.6 Global Buffer 108 5.2.7 Processing Element Architecture 108 5.2.8 Network-on- Chip (NoC) 108 5.2.9 Eyeriss v2 System Architecture 112 5.2.10 Hierarchical Mesh Network 116 5.2.10.1 Input Activation HM-NoC 118 5.2.10.2 Filter Weight HM-NoC 118 5.2.10.3 Partial Sum HM-NoC 119 5.2.11 Compressed Sparse Column Format 120 5.2.12 Row Stationary Plus (RS+) Dataflow 122 5.2.13 System Performance 123 Exercise 125 References 125 6 In-Memory Computation 127 6.1 Neurocube Architecture 127 6.1.1 Hybrid Memory Cube (HMC) 127 6.1.2 Memory Centric Neural Computing (MCNC) 130 6.1.3 Programmable Neurosequence Generator (PNG) 131 6.1.4 System Performance 132 6.2 Tetris Accelerator 133 6.2.1 Memory Hierarchy 133 6.2.2 In-Memory Accumulation 133 6.2.3 Data Scheduling 135 6.2.4 Neural Network Vaults Partition 136 6.2.5 System Performance 137 6.3 NeuroStream Accelerator 138 6.3.1 System Architecture 138 6.3.2 NeuroStream Coprocessor 140 6.3.3 4D Tiling Mechanism 140 6.3.4 System Performance 141 Exercise 143 References 143 7 Near-Memory Architecture 145 7.1 DaDianNao Supercomputer 145 7.1.1 Memory Configuration 145 7.1.2 Neural Functional Unit (NFU) 146 7.1.3 System Performance 149 7.2 Cnvlutin Accelerator 150 7.2.1 Basic Operation 151 7.2.2 System Architecture 151 7.2.3 Processing Order 154 7.2.4 Zero-Free Neuron Array Format (ZFNAf) 155 7.2.5 The Dispatcher 155 7.2.6 Network Pruning 157 7.2.7 System Performance 157 7.2.8 Raw or Encoded Format (RoE) 158 7.2.9 Vector Ineffectual Activation Identifier Format (VIAI) 159 7.2.10 Ineffectual Activation Skipping 159 7.2.11 Ineffectual Weight Skipping 161 Exercise 161 References 161 8 Network Sparsity 163 8.1 Energy Efficient Inference Engine (EIE) 163 8.1.1 Leading Nonzero Detection (LNZD) Network 163 8.1.2 Central Control Unit (CCU) 164 8.1.3 Processing Element (PE) 164 8.1.4 Deep Compression 166 8.1.5 Sparse Matrix Computation 167 8.1.6 System Performance 169 8.2 Cambricon-X Accelerator 169 8.2.1 Computation Unit 171 8.2.2 Buffer Controller 171 8.2.3 System Performance 174 8.3 SCNN Accelerator 175 8.3.1 SCNN PT-IS-CP-Dense Dataflow 175 8.3.2 SCNN PT-IS-CP-Sparse Dataflow 177 8.3.3 SCNN Tiled Architecture 178 8.3.4 Processing Element Architecture 179 8.3.5 Data Compression 180 8.3.6 System Performance 180 8.4 SeerNet Accelerator 183 8.4.1 Low-Bit Quantization 183 8.4.2 Efficient Quantization 184 8.4.3 Quantized Convolution 185 8.4.4 Inference Acceleration 186 8.4.5 Sparsity-Mask Encoding 186 8.4.6 System Performance 188 Exercise 188 References 188 9 3D Neural Processing 191 9.1 3D Integrated Circuit Architecture 191 9.2 Power Distribution Network 193 9.3 3D Network Bridge 195 9.3.1 3D Network-on-Chip 195 9.3.2 Multiple-Channel High-Speed Link 195 9.4 Power-Saving Techniques 198 9.4.1 Power Gating 198 9.4.2 Clock Gating 199 Exercise 200 References 201 Appendix A: Neural Network Topology 203 Index 205

About the Author :
Albert Chun Chen Liu, PhD, is Chief Executive Officer of Kneron. He is Adjunct Associate Professor at National Tsing Hua University, National Chiao Tung University, and National Cheng Kung University. He has published over 15 IEEE papers and is an IEEE Senior Member. He is a recipient of the IBM Problem Solving Award based on the use of the EIP tool suite in 2007 and IEEE TCAS Darlington award in 2021. Oscar Ming Kin Law, PhD, is the Director of Engineering at Kneron. He works on smart robot development and in-memory architecture for neural networks. He has over twenty years of experience in the semiconductor industry working with CPU, GPU, and mobile design. He has also published over 60 patents in various areas.


Best Sellers


Product Details
  • ISBN-13: 9781119810452
  • Publisher: John Wiley & Sons Inc
  • Publisher Imprint: Wiley-IEEE Press
  • Height: 10 mm
  • No of Pages: 240
  • Returnable: N
  • Sub Title: Challenges and Solutions
  • Width: 10 mm
  • ISBN-10: 1119810450
  • Publisher Date: 07 Sep 2021
  • Binding: Hardback
  • Language: English
  • Returnable: N
  • Spine Width: 10 mm
  • Weight: 512 gr


Similar Products

Add Photo
Add Photo

Customer Reviews

REVIEWS      0     
Click Here To Be The First to Review this Product
Artificial Intelligence Hardware Design: Challenges and Solutions
John Wiley & Sons Inc -
Artificial Intelligence Hardware Design: Challenges and Solutions
Writing guidlines
We want to publish your review, so please:
  • keep your review on the product. Review's that defame author's character will be rejected.
  • Keep your review focused on the product.
  • Avoid writing about customer service. contact us instead if you have issue requiring immediate attention.
  • Refrain from mentioning competitors or the specific price you paid for the product.
  • Do not include any personally identifiable information, such as full names.

Artificial Intelligence Hardware Design: Challenges and Solutions

Required fields are marked with *

Review Title*
Review
    Add Photo Add up to 6 photos
    Would you recommend this product to a friend?
    Tag this Book Read more
    Does your review contain spoilers?
    What type of reader best describes you?
    I agree to the terms & conditions
    You may receive emails regarding this submission. Any emails will include the ability to opt-out of future communications.

    CUSTOMER RATINGS AND REVIEWS AND QUESTIONS AND ANSWERS TERMS OF USE

    These Terms of Use govern your conduct associated with the Customer Ratings and Reviews and/or Questions and Answers service offered by Bookswagon (the "CRR Service").


    By submitting any content to Bookswagon, you guarantee that:
    • You are the sole author and owner of the intellectual property rights in the content;
    • All "moral rights" that you may have in such content have been voluntarily waived by you;
    • All content that you post is accurate;
    • You are at least 13 years old;
    • Use of the content you supply does not violate these Terms of Use and will not cause injury to any person or entity.
    You further agree that you may not submit any content:
    • That is known by you to be false, inaccurate or misleading;
    • That infringes any third party's copyright, patent, trademark, trade secret or other proprietary rights or rights of publicity or privacy;
    • That violates any law, statute, ordinance or regulation (including, but not limited to, those governing, consumer protection, unfair competition, anti-discrimination or false advertising);
    • That is, or may reasonably be considered to be, defamatory, libelous, hateful, racially or religiously biased or offensive, unlawfully threatening or unlawfully harassing to any individual, partnership or corporation;
    • For which you were compensated or granted any consideration by any unapproved third party;
    • That includes any information that references other websites, addresses, email addresses, contact information or phone numbers;
    • That contains any computer viruses, worms or other potentially damaging computer programs or files.
    You agree to indemnify and hold Bookswagon (and its officers, directors, agents, subsidiaries, joint ventures, employees and third-party service providers, including but not limited to Bazaarvoice, Inc.), harmless from all claims, demands, and damages (actual and consequential) of every kind and nature, known and unknown including reasonable attorneys' fees, arising out of a breach of your representations and warranties set forth above, or your violation of any law or the rights of a third party.


    For any content that you submit, you grant Bookswagon a perpetual, irrevocable, royalty-free, transferable right and license to use, copy, modify, delete in its entirety, adapt, publish, translate, create derivative works from and/or sell, transfer, and/or distribute such content and/or incorporate such content into any form, medium or technology throughout the world without compensation to you. Additionally,  Bookswagon may transfer or share any personal information that you submit with its third-party service providers, including but not limited to Bazaarvoice, Inc. in accordance with  Privacy Policy


    All content that you submit may be used at Bookswagon's sole discretion. Bookswagon reserves the right to change, condense, withhold publication, remove or delete any content on Bookswagon's website that Bookswagon deems, in its sole discretion, to violate the content guidelines or any other provision of these Terms of Use.  Bookswagon does not guarantee that you will have any recourse through Bookswagon to edit or delete any content you have submitted. Ratings and written comments are generally posted within two to four business days. However, Bookswagon reserves the right to remove or to refuse to post any submission to the extent authorized by law. You acknowledge that you, not Bookswagon, are responsible for the contents of your submission. None of the content that you submit shall be subject to any obligation of confidence on the part of Bookswagon, its agents, subsidiaries, affiliates, partners or third party service providers (including but not limited to Bazaarvoice, Inc.)and their respective directors, officers and employees.

    Accept

    Fresh on the Shelf


    Inspired by your browsing history


    Your review has been submitted!

    You've already reviewed this product!