x11xnn
X11xnn is an open-source neural network inference engine and compiler designed to deliver efficient execution of feedforward models on a range of CPU and GPU backends. The project aims to provide a compact, portable runtime with a focus on small footprint and ease of integration for desktop, server, and embedded applications. It accepts models from common formats such as ONNX and compiles them into optimized kernels that run on available hardware.
Key components include a graph compiler that performs operator fusion and constant folding, a runtime for executing
Supported features typically include common neural network operators (convolutions, depthwise convolutions, fully connected layers, activations, pooling),
Licensing and development: x11xnn is released under an open-source license (project pages indicate MIT- or Apache-style