4th Generation Intel® Xeon® Scalable Processors Deep Dive Video

Document Library
Reference architectures, white papers, and solutions briefs to help build and enhance your network infrastructure, at any level of deployment.
Engagement / Document Library / Optimization of Model Load Time using Model Caching Capability in Intel® Distribution of OpenVINO™ Toolkit
Optimization of Model Load Time using Model Caching Capability in Intel® Distribution of OpenVINO™ Toolkit
Last Updated: Mar 24, 2025
This document presents the BKM for reducing the mode load time by enabling model caching in Intel® Distribution of OpenVINO™ Toolkit. The following figure illustrates the steps involved in implementing the inference pipeline in the user application using OpenVINO™ Runtime API.
Related Content
I Agree to Share My Information
By submitting this form, you agree to share your personal data with Intel for this business request.
Verify Your Email to Continue
To continue viewing the document, please verify your email address. Check your inbox and follow the instructions.
If you didn’t receive the email, click here to generate a new verification link.
Thank You!
Thanks for registering as a member in the Intel® Industry Solution Builders program. Your account has been activated. Please set your password by clicking on the link sent to your registered email address.
Note: If you do not see an email from our site in your inbox, please check your email account's Spam or Junk folder to ensure the message was not filtered. If the message was filtered, please 'Mark as good' or 'Not spam' or 'Add sender to white-list.'