Please log in to watch this conference skillscast.
Getting machine learning models ready for use on-device is a major challenge. Drag-and-drop training tools can get you started, but the models they produce aren’t small enough or fast enough to ship. In this talk, you’ll learn optimization, pruning, and compression techniques that keep app sizes small and inference speeds high. You’ll start by discovering flexible model architectures that meet performance and accuracy requirements across devices and platforms. You’ll then learn pruning and distillation techniques to optimize model performance and use quantization tools to compress models to a fraction of their original size. Finally, you’ll explore a practical example of this process as you create an artistic style transfer model that’s just 17kb. All of these techniques will be applied to mobile machine learning frameworks for Android, i.e. TensorFlow Lite.
YOU MAY ALSO LIKE:
- Functional Programming in Java, Groovy, and Kotlin (Online Meetup on 4th August 2020)
- Ceci n’est pas un canard – Machine Learning and Generative Adversarial Networks (Online Meetup on 6th August 2020)
- Digital Discrimination: Cognitive Bias in Machine Learning (SkillsCast recorded in June 2020)
- Using A.I. to Help Humanitarian Causes (SkillsCast recorded in May 2020)
Creating smaller, faster, production-worthy mobile machine learning models for Android
Dr. Jameson Toole is the CEO and cofounder of Fritz—helping developers teach devices how to see, hear, sense, and think.