You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/arch/deepLearning/Deep_Learning_For_Compute_Architects.md
+13-2
Original file line number
Diff line number
Diff line change
@@ -305,7 +305,7 @@ difference between unsafe optimization and approximate computing. (all unsafe op
305
305
how to distinct an optimization is safe or unsafe:
306
306
APPROXIMATION EXceeding metric said to be approximate as the hardware improvement come at the expense of model accuracy; conversely, as long as the optimization's impact is below the threshold, it is still considered *unsafe*, but not *approximate*
307
307
308
-
this metrics called : **Iso-Training Noise (ITN)**, it is a computable score that quantifies how well an approximated model performs relative to the algorithmically correct one.
308
+
a distinction is made as to whether or not the optimization has a *notable* impact on the output. define a metric to quantify as "notable", this metrics called : **Iso-Training Noise (ITN)**, it is a computable score that quantifies how well an approximated model performs relative to the algorithmically correct one.
309
309
310
310
311
311
## 4.2 Minerva: a three-pronged approach
@@ -349,11 +349,22 @@ the neural network kernel is embarrassingly parallel within a single layer, the
349
349
350
350
- additional safe optimization tricks:
351
351
- input batching for increased locality
352
-
- different architectures
352
+
- different architectures (dataflow: *Eyeriss: An energyefficient reconfigurable accelerator for deep convolutional neural networks*)
Stage3 of Minerva aggressively optimizes neural network bitwidths. the use of optimized data types is a key advantage that allows accelerators to achieve better computional efficiency than the general purpose programmable machines.
357
+
358
+
- Fixed-Point data type design
359
+
- Optimized Fixed-Point bitwidths
356
360
### 4.4.2 Selective Operation Pruning
361
+
Stage4 of Minerva reduce the number of edges that must be processed in the dataflow graph. using empirical analysis of neuron activity, show that by eliminating operations involving small activity values , the number of weight read and MAC operations can be drastically reduced without impacting prediction accuracy.
362
+
363
+
- Analysis of Neural network activations
364
+

0 commit comments