#openvino
7 articles
Advanced
NPU Architecture and GPU+NPU Co-Inference
#intel
#npu
#openvino
#hetero
#multi-device
#co-inference
Advanced
OpenVINO Graph Optimization Pipeline
#intel
#openvino
#graph-optimization
#model-compilation
#plugin
Advanced
LLM Inference on NPU: KV Cache and the Software Stack
#intel
#npu
#llm
#kv-cache
#openvino
#npuw
#static-shape
Intermediate
Impact of Optimization on Accuracy
#benchmark
#quantization
#accuracy
#perplexity
#openvino
#lm-eval-harness
#llama-cpp
Intermediate
Intel Model Optimization Stack: Choosing Between Optimum Intel, NNCF, and OpenVINO
#intel
#optimum
#nncf
#openvino
#quantization
#model-conversion
Intermediate
Quantization and Model Conversion Toolchain Landscape
#quantization
#model-conversion
#toolchain
#optimum
#nncf
#openvino
#gguf
#onnx
Intermediate
Hands-On: HF β GGUF / ONNX / OpenVINO β Three End-to-End Paths
#quantization
#model-conversion
#hands-on
#llama-cpp
#onnx
#openvino
#intel-igpu