During the training phase, neural networks learn to provide the desired result by modifying their weights and biases. A trained network is then deployed to solve unknown problems in the real world in a process called inferencing.
Embedding complex algorithms in constrained devices affects the embedded design, latency, and power consumption in those devices. Running AI neural networks algorithms on resource constrained devices requires the algorithm and hardware designer to co-design a solution that addresses both the data engineering and data science needs. Acceleration and compression are some of the other critical problems to be addressed while developing intelligence on embedded devices.