Edge Microvisor Toolkit Versions#
Edge Microvisor Toolkit is available in several pre-configured versions that serve different purposes. Some are published as binaries, others are available from a custom build. This document will help you select the version that best suits your needs. To do so, check out:
How to select the right EMT.
The diagram below will help you select the toolkit version right for your workflow.
How EMT differs between versions.
Version
Real Time
Stable Kernel
Available for opt-in
✓
✓
Optional
✓
✓
Available for opt-in
✓
✓
-
✓
–
How usage scenarios affect EMT setup.
Scenario
Description
Primary outcomes
Technology areas
Real-time & deterministic workloads
Run latency-sensitive workloads with guaranteed bounded jitter and repeatable execution timelines across one or more hosts, maintainable under steady-state and failure-recovery conditions
- Bounded end-to-end latency & jitter
- Repeatable scheduling windows under load
- Cross-host timing consistency for distributed stages
- Fast, predictable recovery without violating SLOs
- PREEMPT_RT kernel
- Resource Director Technologies
- Intel GPU RT
- CPU & Scheduler Isolation
- Memory Determinism
- Time & Clocks
- Network Determinism (TSN)VM-based workloads on Kubernetes with shared GPUs
Run multiple virtual machines on Kubernetes that concurrently share one or more physical GPUs, with predictable fairness, isolation, and policy-driven placement—using a KubeVirt stack extended for GPU sharing
- Stable, repeatable GPU performance per VM under contention
- Hard/soft sharing policies (fair-share, priority tiers, or quotas)
- Safe isolation between tenants/VMs (memory, contexts, resets)
- Schedulable resources with clear admission signals (no surprise fails)
- Operational guardrails: health checks, graceful drain/eviction, rollback
- SRIOV
- Intel GPU
- kubevirt
- Host virtualization
- Intel GPU device pluginAI & Vision workloads
Enable AI inference and computer-vision workloads on edge nodes using Intel GPU and NPU acceleration, exposing unified hardware-assisted pipelines through standard APIs and user-space libraries
- Efficient execution of deep-learning and vision inference on-device without cloud dependency
- Unified GPU/NPU compute abstraction for developers (OpenVINO backend, media pipelines)
- Deterministic frame-rate and latency for multi-stream analytics workloads (e.g., camera ingest)
- Seamless integration with containers or pods, including dynamic device discovery and sharing
- Stable ABI/API interface across OS updates and driver versions
- Edge AI packages
- OpenVino
- Intel GPU and NPU drivers
- Intel GPU device pluginHow to build your own version of EMT.
You can create your own custom version of Edge Microvisor Toolkit by following the guide. You can also try and learn how to build your own solution and deploy it on edge.