SPRACZ2 August   2022 TDA4VM , TDA4VM-Q1

ADVANCE INFORMATION  

  1.   Abstract
  2. 1Introduction
    1. 1.1 Vision Analytics
    2. 1.2 End Equipments
    3. 1.3 Deep learning: State-of-the-art
  3. 2Embedded edge AI system: Design considerations
    1. 2.1 Processors for edge AI: Technology landscape
    2. 2.2 Edge AI with TI: Energy-efficient and Practical AI
      1. 2.2.1 TDA4VM processor architecture
        1. 2.2.1.1 Development platform
    3. 2.3 Software programming
  4. 3Industry standard performance and power benchmarking
    1. 3.1 MLPerf models
    2. 3.2 Performance and efficiency benchmarking
    3. 3.3 Comparison against other SoC Architectures
      1. 3.3.1 Benchmarking against GPU-based architectures
      2. 3.3.2 Benchmarking against FPGA based SoCs
      3. 3.3.3 Summary of competitive benchmarking
  5. 4Conclusion
  6.   Revision History
  7. 5References

Embedded edge AI system: Design considerations

There are two aspects required of any neural network or a deep learning model: training and inference. The training function involves using a set of training data to train the model. Once the training is done, the model can be used for inference on a new set of input data. Typically, training is done once or a few times for a given product. Inference, on the other hand, happens all the time for a given edgeAI system.

Figure 2-1 shows the difference between the two steps. This is the reason why the inference process needs to be optimized for high performance and high energy efficiency for any embedded edge AI device.

GUID-4EA81616-C27B-41AA-BE49-CEB270813161-low.gif Figure 2-1 Training vs Inference Process

Training a deep learning model typically requires very high TOPS processing engine but for most low- to mid-end edge AI inference applications, TOPS required are in the range of 1 to 32 TOPS. This is the segment that TDA4x processor family is targeting.