Fast inference engine for Transformer models
-
Updated
Apr 8, 2025 - C++
Fast inference engine for Transformer models
oneAPI Deep Neural Network Library (oneDNN)
High-performance Spiking Neural Networks Library Written From Scratch with C++ and Python Interfaces.
[WIP] edge-optimized neural style transfer using Intel oneDNN
high-level Rust bindings to the oneDNN C api
Add a description, image, and links to the onednn topic page so that developers can more easily learn about it.
To associate your repository with the onednn topic, visit your repo's landing page and select "manage topics."