Skip to content

Commit e9259cd

Browse files
author
ComputeGeneralLab
committed
update dl4ca
1 parent eb60403 commit e9259cd

File tree

2 files changed

+13
-2
lines changed

2 files changed

+13
-2
lines changed

Diff for: docs/arch/deepLearning/Deep_Learning_For_Compute_Architects.md

+13-2
Original file line numberDiff line numberDiff line change
@@ -305,7 +305,7 @@ difference between unsafe optimization and approximate computing. (all unsafe op
305305
how to distinct an optimization is safe or unsafe:
306306
APPROXIMATION EXceeding metric said to be approximate as the hardware improvement come at the expense of model accuracy; conversely, as long as the optimization's impact is below the threshold, it is still considered *unsafe*, but not *approximate*
307307

308-
this metrics called : **Iso-Training Noise (ITN)**, it is a computable score that quantifies how well an approximated model performs relative to the algorithmically correct one.
308+
a distinction is made as to whether or not the optimization has a *notable* impact on the output. define a metric to quantify as "notable", this metrics called : **Iso-Training Noise (ITN)**, it is a computable score that quantifies how well an approximated model performs relative to the algorithmically correct one.
309309

310310

311311
## 4.2 Minerva: a three-pronged approach
@@ -349,11 +349,22 @@ the neural network kernel is embarrassingly parallel within a single layer, the
349349

350350
- additional safe optimization tricks:
351351
- input batching for increased locality
352-
- different architectures
352+
- different architectures (dataflow: *Eyeriss: An energyefficient reconfigurable accelerator for deep convolutional neural networks*)
353353

354354
## 4.4 Low-Power Neural networks accelerators: Unsafe Optimizations
355355
### 4.4.1 Data Type Quantization
356+
Stage3 of Minerva aggressively optimizes neural network bitwidths. the use of optimized data types is a key advantage that allows accelerators to achieve better computional efficiency than the general purpose programmable machines.
357+
358+
- Fixed-Point data type design
359+
- Optimized Fixed-Point bitwidths
356360
### 4.4.2 Selective Operation Pruning
361+
Stage4 of Minerva reduce the number of edges that must be processed in the dataflow graph. using empirical analysis of neuron activity, show that by eliminating operations involving small activity values , the number of weight read and MAC operations can be drastically reduced without impacting prediction accuracy.
362+
363+
- Analysis of Neural network activations
364+
![analysis of neuron activations and sensitivity of prediction error to pruning](neuron_activations_and_sensitivity_of_prediction_error_to_pruning.png)
365+
366+
- Predicating on insignificant operations.
367+
357368
### 4.4.3 SRAM fault mitigation
358369

359370
## 4.5 Discussion
Loading

0 commit comments

Comments
 (0)