Course Curriculum

  1. 1

    Book Preview

    1. Book Preview Free preview
  2. 2

    Introduction

    1. (Included in full purchase)
  3. 3

    Chapter 1 : Introduction to ONNX and Edge Computing

    1. (Included in full purchase)
  4. 4

    Chapter 2 : Getting Started with ONNX

    1. (Included in full purchase)
  5. 5

    Chapter 3 : ONNX Integration with Deep Learning Frameworks

    1. (Included in full purchase)
  6. 6

    Chapter 4 : Model Optimization Using ONNX Simplifier and ONNX Runtime

    1. (Included in full purchase)
  7. 7

    Chapter 5 : Model Quantization Using ONNX Runtime

    1. (Included in full purchase)
  8. 8

    Chapter 6 : Model Pruning in Pytorch and Exporting to ONNX

    1. (Included in full purchase)
  9. 9

    Chapter 7 : Knowledge Distillation for Edge AI

    1. (Included in full purchase)
  10. 10

    Chapter 8 : Deploying ONNX Models on Edge Devices

    1. (Included in full purchase)
  11. 11

    Chapter 9 : End to End Execution of YOLOv12

    1. (Included in full purchase)
  12. 12

    Chapter 10 : End to End Execution of Whisper Speech Recognition Model

    1. (Included in full purchase)
  13. 13

    Chapter 11 : End to End Execution of SmolLM Model

    1. (Included in full purchase)
  14. 14

    Chapter 12 : ONNX Model from Scratch and Custom Operator

    1. (Included in full purchase)
  15. 15

    Chapter 13 : Real-World Applications, Best Practices, Security, and Future Trends in ONNX for Edge AI

    1. (Included in full purchase)
  16. 16

    Index

    1. (Included in full purchase)

About the Course

ONNX has emerged as the de facto standard for deploying portable, framework-agnostic machine learning models across diverse hardware platforms. Ultimate ONNX for Deep Learning Optimization provides a structured, end-to-end guide to the ONNX ecosystem, starting with ONNX fundamentals, model representation, and framework integration. You will learn how to export models from PyTorch, TensorFlow, and Scikit-Learn, inspect and modify ONNX graphs, and leverage ONNX Runtime and ONNX Simplifier for inference optimization. Each chapter builds technical depth, equipping you with the tools required to move models beyond experimentation. The book focuses on performance-critical optimization techniques, including quantization, pruning, and knowledge distillation, followed by practical deployment on edge devices such as Raspberry Pi. Through complete, real-world case studies covering object detection, speech recognition, and compact language models, you can implement custom operators, follow deployment best practices, and understand production constraints. Thus, by the end of this book, you will be capable of designing, optimizing, and deploying efficient ONNX-based AI systems for edge environments. 

About the Author

Meet Patel is a machine learning engineer with over seven years of expertise dedicated to a singular challenge, that is, making Artificial Intelligence (AI) faster, smaller, and more efficient. His passion lies in unlocking the potential of AI on resource-constrained devices, pushing models from the lab into the real world. His transition into AI from a mechanical engineering background underscores a journey fueled by curiosity and self-motivation. He was driven by a passion to master the intricacies of machine learning. Meet has extensive hands-on experience in taking models from initial research and training through advanced optimization techniques such as quantization, pruning, and knowledge distillation, all the way to compiler level enhancements and final deployment.