• Stars
    star
    3,576
  • Rank 12,403 (Top 0.3 %)
  • Language
    C++
  • License
    Apache License 2.0
  • Created over 8 years ago
  • Updated 2 months ago

Reviews

There are no reviews yet. Be the first to send feedback to the community and the maintainers!

Repository Details

oneAPI Deep Neural Network Library (oneDNN)

UXL Foundation Logo

oneAPI Deep Neural Network Library (oneDNN)

OpenSSF Best Practices OpenSSF Scorecard

oneAPI Deep Neural Network Library (oneDNN) is an open-source cross-platform performance library of basic building blocks for deep learning applications. oneDNN project is part of the UXL Foundation and is an implementation of the oneAPI specification for oneDNN component.

The library is optimized for Intel(R) Architecture Processors, Intel Graphics, and Arm* 64-bit Architecture (AArch64)-based processors. oneDNN has experimental support for the following architectures: NVIDIA* GPU, AMD* GPU, OpenPOWER* Power ISA (PPC64), IBMz* (s390x), and RISC-V.

oneDNN is intended for deep learning applications and framework developers interested in improving application performance on CPUs and GPUs. Deep learning practitioners should use one of the applications enabled with oneDNN.

Table of Contents

Documentation

  • Developer Guide explains the programming model, supported functionality, and implementation details, and includes annotated examples.
  • API Reference provides a comprehensive reference of the library API.

Installation

Binary distribution of this software is available in:

The packages do not include library dependencies and these need to be resolved in the application at build time. See the System Requirements section below and the Build Options section in the Developer Guide for more details on CPU and GPU runtimes.

If the configuration you need is not available, you can build the library from source.

System Requirements

oneDNN supports platforms based on the following architectures:

WARNING

Power ISA (PPC64), IBMz (s390x), and RISC-V (RV64) support is experimental with limited testing validation.

The library is optimized for the following CPUs:

  • Intel Atom(R) processor (at least Intel SSE4.1 support is required)
  • Intel Core(TM) processor (at least Intel SSE4.1 support is required)
  • Intel Core Ultra processors (formerly Meteor Lake)
  • Intel Xeon(R) processor E3, E5, and E7 family (formerly Sandy Bridge, Ivy Bridge, Haswell, and Broadwell)
  • Intel Xeon Scalable processor (formerly Skylake, Cascade Lake, Cooper Lake, Ice Lake, Sapphire Rapids, and Emerald Rapids)
  • Intel Xeon CPU Max Series (formerly Sapphire Rapids HBM)
  • future Intel Xeon Scalable processors (code name Sierra Forest and Granite Rapids)

On a CPU based on Intel 64 or on AMD64 architecture, oneDNN detects the instruction set architecture (ISA) at runtime and uses just-in-time (JIT) code generation to deploy the code optimized for the latest supported ISA. Future ISAs may have initial support in the library disabled by default and require the use of run-time controls to enable them. See CPU dispatcher control for more details.

On a CPU based on Arm AArch64 architecture, oneDNN can be built with Arm Compute Library (ACL) integration. ACL is an open-source library for machine learning applications and provides AArch64 optimized implementations of core functions. This functionality currently requires that ACL is downloaded and built separately; see Build from Source. oneDNN only supports Compute Library versions 23.11 or later.

WARNING

On macOS, applications that use oneDNN may need to request special entitlements if they use the hardened runtime. See the Linking Guide for more details.

The library is optimized for the following GPUs:

  • Intel Graphics for 11th-14th Generation Intel Core Processors
  • Intel Graphics for Intel Core Ultra processors (formerly Meteor Lake)
  • Intel Iris Xe MAX Graphics (formerly DG1)
  • Intel Arc(TM) graphics (formerly Alchemist)
  • Intel Data Center GPU Flex Series (formerly Arctic Sound)
  • Intel Data Center GPU Max Series (formerly Ponte Vecchio)

Requirements for Building from Source

oneDNN supports systems meeting the following requirements:

  • Operating system with Intel 64 / Arm 64 / Power / IBMz architecture support
  • C++ compiler with C++11 standard support
  • CMake 2.8.12 or later
  • Arm Compute Library (ACL) for builds using ACL on AArch64.

The following tools are required to build oneDNN documentation:

Configurations of CPU and GPU engines may introduce additional build time dependencies.

CPU Engine

oneDNN CPU engine is used to execute primitives on Intel Architecture Processors, 64-bit Arm Architecture (AArch64) processors, 64-bit Power ISA (PPC64) processors, IBMz (s390x), and compatible devices.

The CPU engine is built by default but can be disabled at build time by setting DNNL_CPU_RUNTIME to NONE. In this case, GPU engine must be enabled. The CPU engine can be configured to use the OpenMP, TBB or SYCL runtime. The following additional requirements apply:

Some implementations rely on OpenMP 4.0 SIMD extensions. For the best performance results on Intel Architecture Processors we recommend using the Intel C++ Compiler.

GPU Engine

Intel Processor Graphics and Xe Architecture graphics are supported by the oneDNN GPU engine. The GPU engine is disabled in the default build configuration. The following additional requirements apply when GPU engine is enabled:

WARNING

Linux will reset GPU when kernel runtime exceeds several seconds. The user can prevent this behavior by disabling hangcheck for Intel GPU driver. Windows has built-in timeout detection and recovery mechanism that results in similar behavior. The user can prevent this behavior by increasing the TdrDelay value.

WARNING

NVIDIA GPU support is experimental. General information, build instructions, and implementation limitations are available in the NVIDIA backend readme.

WARNING

AMD GPU support is experimental. General information, build instructions, and implementation limitations are available in the AMD backend readme.

Runtime Dependencies

When oneDNN is built from source, the library runtime dependencies and specific versions are defined by the build environment.

Linux

Common dependencies:

  • GNU C Library (libc.so)
  • GNU Standard C++ Library v3 (libstdc++.so)
  • Dynamic Linking Library (libdl.so)
  • C Math Library (libm.so)
  • POSIX Threads Library (libpthread.so)

Runtime-specific dependencies:

Runtime configuration Compiler Dependency
DNNL_CPU_RUNTIME=OMP GCC GNU OpenMP runtime (libgomp.so)
DNNL_CPU_RUNTIME=OMP Intel C/C++ Compiler Intel OpenMP runtime (libiomp5.so)
DNNL_CPU_RUNTIME=OMP Clang Intel OpenMP runtime (libiomp5.so)
DNNL_CPU_RUNTIME=TBB any TBB (libtbb.so)
DNNL_CPU_RUNTIME=SYCL Intel oneAPI DPC++ Compiler Intel oneAPI DPC++ Compiler runtime (libsycl.so), TBB (libtbb.so), OpenCL loader (libOpenCL.so)
DNNL_GPU_RUNTIME=OCL any OpenCL loader (libOpenCL.so)
DNNL_GPU_RUNTIME=SYCL Intel oneAPI DPC++ Compiler Intel oneAPI DPC++ Compiler runtime (libsycl.so), OpenCL loader (libOpenCL.so), oneAPI Level Zero loader (libze_loader.so)

Windows

Common dependencies:

  • Microsoft Visual C++ Redistributable (msvcrt.dll)

Runtime-specific dependencies:

Runtime configuration Compiler Dependency
DNNL_CPU_RUNTIME=OMP Microsoft Visual C++ Compiler No additional requirements
DNNL_CPU_RUNTIME=OMP Intel C/C++ Compiler Intel OpenMP runtime (iomp5.dll)
DNNL_CPU_RUNTIME=TBB any TBB (tbb.dll)
DNNL_CPU_RUNTIME=SYCL Intel oneAPI DPC++ Compiler Intel oneAPI DPC++ Compiler runtime (sycl.dll), TBB (tbb.dll), OpenCL loader (OpenCL.dll)
DNNL_GPU_RUNTIME=OCL any OpenCL loader (OpenCL.dll)
DNNL_GPU_RUNTIME=SYCL Intel oneAPI DPC++ Compiler Intel oneAPI DPC++ Compiler runtime (sycl.dll), OpenCL loader (OpenCL.dll), oneAPI Level Zero loader (ze_loader.dll)

macOS

Common dependencies:

  • System C/C++ runtime (libc++.dylib, libSystem.dylib)

Runtime-specific dependencies:

Runtime configuration Compiler Dependency
DNNL_CPU_RUNTIME=OMP Intel C/C++ Compiler Intel OpenMP runtime (libiomp5.dylib)
DNNL_CPU_RUNTIME=TBB any TBB (libtbb.dylib)

Validated Configurations

CPU engine was validated on RedHat* Enterprise Linux 8 with

on Windows Server* 2019 with

on macOS 11 (Big Sur) with

GPU engine was validated on Ubuntu* 22.04 with

on Windows Server 2019 with

Applications Enabled with oneDNN

Support

Submit questions, feature requests, and bug reports on the GitHub issues page.

You can also contact oneDNN developers via UXL Foundation Slack using #onednn channel.

Governance

oneDNN project is governed by the UXL Foundation and you can get involved in this project in multiple ways. It is possible to join the AI Special Interest Group (SIG) meetings where the groups discuss and demonstrate work using this project. Members can also join the Open Source and Specification Working Group meetings.

You can also join the mailing lists for the UXL Foundation to be informed of when meetings are happening and receive the latest information and discussions.

Contributing

We welcome community contributions to oneDNN. If you have an idea on how to improve the library:

For additional details, see contribution guidelines. You can also contact oneDNN developers and maintainers via UXL Foundation Slack using #onednn channel.

This project is intended to be a safe, welcoming space for collaboration, and contributors are expected to adhere to the Contributor Covenant code of conduct.

License

oneDNN is licensed under Apache License Version 2.0. Refer to the "LICENSE" file for the full license text and copyright notice.

This distribution includes third party software governed by separate license terms.

3-clause BSD license:

2-clause BSD license:

Apache License Version 2.0:

Boost Software License, Version 1.0:

MIT License:

This third party software, even if included with the distribution of the Intel software, may be governed by separate license terms, including without limitation, third party license terms, other Intel software license terms, and open source software license terms. These separate license terms govern your use of the third party programs as set forth in the "THIRD-PARTY-PROGRAMS" file.

Security

Security Policy outlines our guidelines and procedures for ensuring the highest level of Security and trust for our users who consume oneDNN.

Trademark Information

Intel, the Intel logo, Arc, Intel Atom, Intel Core, Iris, OpenVINO, the OpenVINO logo, Pentium, VTune, and Xeon are trademarks of Intel Corporation or its subsidiaries.

* Other names and brands may be claimed as the property of others.

Microsoft, Windows, and the Windows logo are trademarks, or registered trademarks of Microsoft Corporation in the United States and/or other countries.

OpenCL and the OpenCL logo are trademarks of Apple Inc. used by permission by Khronos.

(C) Intel Corporation

More Repositories

1

oneTBB

oneAPI Threading Building Blocks (oneTBB)
C++
5,603
star
2

oneAPI-samples

Samples for Intel® oneAPI Toolkits
C++
922
star
3

oneDPL

oneAPI DPC++ Library (oneDPL) https://software.intel.com/content/www/us/en/develop/tools/oneapi/components/dpc-library.html
C++
720
star
4

oneDAL

oneAPI Data Analytics Library (oneDAL)
C++
607
star
5

oneMKL

oneAPI Math Kernel Library (oneMKL) Interfaces
C++
606
star
6

SYCLomatic

LLVM
221
star
7

level-zero

oneAPI Level Zero Specification Headers and Loader
C++
210
star
8

oneCCL

oneAPI Collective Communications Library (oneCCL)
C++
188
star
9

oneVPL

oneAPI Video Processing Library (oneVPL) dispatcher, tools, and examples
C++
173
star
10

oneAPI-spec

oneAPI Specification source files
Python
165
star
11

oneapi-ci

Sample configuration files for using oneAPI in CI systems
Shell
92
star
12

oneVPL-intel-gpu

C++
86
star
13

oneAPI-tab

oneAPI Technical Advisory Board (TAB) Meeting Notes
71
star
14

distributed-ranges

Distributed ranges is a generalization of C++ ranges for distributed data structures.
C++
46
star
15

level-zero-tests

oneAPI Level Zero Conformance & Performance test content
C++
45
star
16

Velocity-Bench

C++
42
star
17

unified-runtime

C++
31
star
18

unified-memory-framework

A library for constructing allocators and memory pools. It also contains broadly useful abstractions and utilities for memory management. UMF allows users to manage multiple memory pools characterized by different attributes, allowing certain allocation types to be isolated from others and allocated using different hardware resources as required.
C
31
star
19

oneVPL-cpu

oneAPI Video Processing Library (oneVPL) CPU implementation. This GitHub repository is no longer active. See ReadMe for more information.
C++
25
star
20

level-zero-spec

Python
17
star
21

ishmem

Intel® SHMEM - Device initiated shared memory based communication library
C++
15
star
22

drone-navigation-inspection

AI Starter Kit for AI applications in Drone technology using Intel® Optimized Tensorflow*
Python
13
star
23

predictive-asset-health-analytics

AI Starter Kit for Predictive Asset Maintenance using Intel® optimized version of XGBoost
HTML
13
star
24

SYCLomatic-test

LLVM
13
star
25

text-data-generation

AI Starter Kit for AI Unstructured Synthetic Data Generation using Intel® Extension for Pytorch
Python
10
star
26

traffic-camera-object-detection

AI Starter Kit for traffic camera object detection using Intel® Extension for Pytorch
Python
10
star
27

invoice-to-cash-automation

Ai starter kit for trade promotion and claim documents categorization using pytorch* and Tensorflow*
Python
7
star
28

demand-forecasting

AI Starter Kit for demand forecasting using Intel® Optimized Tensorflow*
Python
7
star
29

disease-prediction

AI Starter Kit for the implementation of AI-based NLP Disease Prediction system using Intel® Extension for PyTorch* and Intel® Neural Compressor
Python
7
star
30

computational-fluid-dynamics

AI Starter Kit for fluid Flow Profiling using Intel® Optimized Tensorflow*
Python
6
star
31

historical-assets-document-process

AI Starter Kit for Historical Assets document processing using Intel® Extension for Pytorch
Python
6
star
32

network-intrusion-detection

AI Starter Kit for Network Intrusion Detection using Intel® Extension for Scikit-learn*
Python
6
star
33

ai-transcribe

AI Starter Kit for the implementation of an AI transcribe system using Intel® Extension for PyTorch*
Python
6
star
34

level-zero-intel-gpu

5
star
35

structural-damage-assessment

AI Starter Kit for applications in Satellite Image processing using Intel® Extension for Pytorch
Python
5
star
36

digital-twin

AI Starter Kit to build a MOSFET Digital Twin for Design Exploration using Intel® optimized version of XGBoost
Python
4
star
37

medical-imaging-diagnostics

AI Starter Kit for image-based abnormalities for different diseases classification using Intel® Optimized Tensorflow*
Python
4
star
38

visual-quality-inspection

AI Starter Kit for Quality Visual Inspection using Intel® Extension for Pytorch
Python
4
star
39

customer-chatbot

AI Starter Kit for Customer Chatbot using Intel® Extension for Pytorch
Python
3
star
40

distributed-ranges-tutorial

C++
3
star
41

purchase-prediction

AI Starter Kit for Purchase Prediction model using Intel® Extension for Scikit-learn*
Python
3
star
42

customer-segmentation

AI Starter Kit for Customer Segmentation for Online Retail using Intel® Extension for Scikit-learn*
Python
3
star
43

powerline-fault-detection

AI Starter Kit for detect faulty signals in power line voltage using Intel® Extension for Scikit-learn*
Python
3
star
44

image-data-generation

AI Starter Kit for Synthetic Image Generation using Intel® Optimized Tensorflow*
Python
2
star
45

intelligent-indexing

AI Starter Kit for Intelligent Indexing of Incoming Correspondence using Intel® Extension for Scikit-learn*
Python
2
star
46

unified-runtime-spec

2
star
47

visual-process-discovery

AI Starter Kit for Visual Process Discovery using Intel® Extension for Pytorch
Python
2
star
48

vertical-search-engine

AI Starter Kit for Semantic Vertical Search Engines using Intel® Extension for Pytorch
Python
2
star
49

document-automation

AI Starter Kit for Named Entity Recognition using Intel® Optimized Tensorflow (version 2.9.0 with oneDNN)
Python
2
star
50

ai-structured-data-generation

AI Starter Kit to generate structured synthetic data using Intel® Distribution of Modin
Python
1
star
51

voice-data-generation

AI Starter Kit for Synthetic Voice and Audio Generation using Intel® Extension for Pytorch
Python
1
star
52

order-to-delivery-time-forecasting

AI Starter Kit of a delivery time forecasting solution using Intel® optimized version of XGBoost
1
star
53

product-recommendations

AI Starter Kit for product recommendation system using Intel® Extension for Scikit-learn*
Jupyter Notebook
1
star
54

customer-churn-prediction

AI Starter Kit for customer churn prediction using Intel® Extension for Scikit-learn*
Python
1
star
55

credit-card-fraud-detection

AI Starter Kit for Credit Card Fraud Detection model using Intel® Extension for Scikit-learn*
Python
1
star
56

loan-default-risk-prediction

AI Starter Kit to predict probability of a loan default from client using Intel® optimized version of XGBoost
Python
1
star
57

ai-data-protection

AI Starter Kit for Personal Identifiable Information Anonymization using Intel® Extension for Pytorch
Python
1
star
58

engineering-design-optimization

AI Starter Kit for Engineering Design Optimization using Intel® Extension for Pytorch
Python
1
star
59

data-streaming-anomaly-detection

AI Starter Kit for Data Streaming Anomaly Detection using Intel® Optimized Tensorflow*
Python
1
star