• Stars
    star
    573
  • Rank 77,865 (Top 2 %)
  • Language
    C++
  • Created over 7 years ago
  • Updated over 1 year ago

Reviews

There are no reviews yet. Be the first to send feedback to the community and the maintainers!

Repository Details

Compute Library for Deep Neural Networks (clDNN)

DISCONTINUATION OF PROJECT

This project will no longer be maintained by Intel.

Intel has ceased development and contributions including, but not limited to, maintenance, bug fixes, new releases, or updates, to this project.

Intel no longer accepts patches to this project.

If you have an ongoing need to use this project, are interested in independently developing it, or would like to maintain patches for the open source software community, please create your own fork of this project.

Contact: [email protected] 

Compute Library for Deep Neural Networks (clDNN)


Discontinued repository

This project now is an integral part of Intel® Distribution of OpenVino™ Toolkit. It's content and development has been moved to DLDT repo.

To get latest clDNN sources please refer to DLDT repo.


Apache License Version 2.0 v1.0

Compute Library for Deep Neural Networks (clDNN) is an open source performance library for Deep Learning (DL) applications intended for acceleration of DL Inference on Intel® Processor Graphics – including HD Graphics and Iris® Graphics.
clDNN includes highly optimized building blocks for implementation of convolutional neural networks (CNN) with C and C++ interfaces. We created this project to enable the DL community to innovate on Intel® processors.

Usages supported: Image recognition, image detection, and image segmentation.

Validated Topologies: AlexNet*, VGG(16,19)*, GoogleNet(v1,v2,v3)*, ResNet(50,101,152)* Faster R-CNN*, Squeezenet*, SSD_googlenet*, SSD_VGG*, PVANET*, PVANET_REID*, age_gender*, FCN* and yolo*.

As with any technical preview, APIs may change in future updates.

License

clDNN is licensed is licensed under Apache License Version 2.0.

Attached licenses

clDNN uses 3rd-party components licensed under following licenses:

Documentation

The latest clDNN documentation is at GitHub pages.

There is also inline documentation available that can be generated with Doxygen.

Accelerate Deep Learning Inference with Intel® Processor Graphics whitepaper link.

Intel® OpenVino™ Toolkit and clDNN

clDNN is released also together with Intel® OpenVino™ Toolkit, which contains:

  • Model Optimizer a Python*-based command line tool, which imports trained models from popular deep learning frameworks such as Caffe*, TensorFlow*, and Apache MXNet*.
  • Inference Engine an execution engine which uses a common API to deliver inference solutions on the platform of your choice (for example GPU with clDNN library)

You can find more information here.

Changelog

Drop 14.1

New features:
- network serialization
- 3D support for: Acitvation, Reorder, Eltwise, Reshape, Deconvolution
Bug fixes:
- concatenation fix for different input formats
UX:
- added 2019.4 intel ocl icd
- refactored bfyx_f16 format
- added i32 and i64 support for select primitive

Drop 14.0

New features:
- 3 spatial dimensions support in convolution primitive (3D convolution)
- reverse primitive
- arg_max_min support for i8/s8/i32/i64 types
- concatenation support for bfzyx (5D) format
Bug fixes:
- fixes in primitive fusing pass (for i8/s8 types)
- fixes in graph optimizer (reshape primitive)
- overflow/underflow fixes for eltwise (i8/s8)
- fixes for convolution-eltwise primitive
- fixes for convolution primitive (depth-wise case)
- perf fixes for events pool
- fixes for pooling primitive (u8)
- fixes for deconvolution primitive
- fixes for fc primitive
- fixes for batch_norm primitive
UX:
- refactored and cleaned up JIT constants generation mechanism
- refactored kernel selection mechanism
- removed legacy device info mechanism
Performance:
- convolution primitive optimizations (for byxf, for MMAD-based, for byxf fp16, for bfyx fp16)
- fc primitive optimizations (for byxf)
- pooling primitive optimizations (for byxf, bfyx)
- convolution-relu primitive fusing (i8 -> s8 case)
- eltwise primitive optimizations (for byxf)
- fused convolution-eltwise primitive optimizations (IMAD-based)
- block-based optimizations for fp16 primitives

Drop 13.1

New features:
- added max mode for contract primitive
- added one_hot primitive
- optional explicit output data type support for all primitives
Bug fixes:
- fix for graph optimizer (crop primitive)
- fix for processing order (deconvolution primitive)
- fix for convolution-eltwise primitive
UX:
- cache.json is searched in to library directory
Performance:
- optimizations for lstm_gemm primitive

Drop 13.0

New features:
- events pool
- group support in convolution and deconvolution primitives
- broadcastable inputs support for eltwise primitive
- asymmetric padding for convolution primitive
- fused convolution-eltwise primitive (API extension)
- auto-calculated output shape support for reshape primitive
- crop support for i8/s8/i32/i64 types
- broadcast axis support for broadcast primitive
- logic and comparison operations support for eltwise primitive
Bug fixes:
- added required alignment checks for some fc implementations
- added lstm support for f16 (half) type
- reorders for fc moved to graph compiler
- primitive fusing and reorder fixes
UX:
- added internal core tests project
- refactored optimizations pass manager and passes
Performance:
- optimized concatenation during upsampling (unpool)
- IMAD-based optimizations for convolution, fc, eltwise and pooling primitives (i8/s8)
- convolution-eltwise fusing optimizations
- partial writes optimizations for block-based kernels

Drop 12.1

- gtests code refactor
- buildbreak fix

Drop 12.0

New features:
- pyramidRoiAlign primitive
- multiple axes support for reverse mode in index_select
- eltwise min/max/mod support for i8/i32/i64
- broadcast support for i32/i64
Bug fixes:
- memory leak fixes
- in-place reshape
- no padding for output primitives
UX:
- RapidJSON library for auto-tune cache
- less dependencies in program.cpp
- do not throw error, when device not validated
- global pooling in c API
- optimized padding for convolution

Drop 11.0

New features:
- throttle hints
- extended border and tile
- GPU implementation of Detection Output
- More cases for BatchNorm primitive
Bug fixes:
- GEMM fix (align with ONNX)
- memory leak fix in memory pool
- increase FC precision for fp16 (fp32 accu) 
Performance:
- cache for new topologies and devices
- conv1x1 with stride >1 into eltwise optimization 

Drop 10.0

New features:
- condition primitive
- fused convolution with bn and scale (backprop)
- scale/shit and mean/var as an output in batch norm
- add LSTM output selection
Bug fixes:
- memory pool fixes 
UX:
- downgrade to cxx11
- add support for u8 data type in custom primitive 
- library size optimizations
Performance:
- in place concatenation optimization 
- conv1x1 with stride >1 into eltwise optimization 

Drop 9.2

New features
- local convolution
- eltwise with strie

Drop 9.1

New features:
- select index primitive
- gemm primitive
Bug fixes:
- fix for output format in fully connected primitive

Drop 9.0

New features:
- log2 activation function
- support for i32 and i64 types
- select primitive
- border primitive
- tile primitive
Bug fixes:
- dilation > input size fix

Drop 8.0

New features:
- lstm primitive
- average unpooling primitive
- serialization - dump weights, biases and kernels
- scale grad for input and weights primitive
Bug fixes:
- wrong gws in concatenation
- int8 layers
- convolution depthwise bias concatenation
- params in engine_info
- mutable_data filler
- momentum calculation
UX:
- kernel selector renaming
- bfyx_yxfb batched reorder
- code cleanups
- primitives allocation order

Drop 7.0

New features:
- support for img_info=4 in proposal_gpu
- support images format in winograd
- support for 2 or more inputs in eltwise
- priority and throttle hints
- deconvolution_grad_input primitive
- fc_grad_input and fc_grad_weights primitives
Bug fixes:
- tensor fixes (i.e. less operator fix)
- cascade concat fixes
- winograd fixes for bfyx format
- auto-tuning fixes for weights calculation
UX:
- memory pool (reusing memory buffers)
- added choosen kernel name in graph dump
- flush memory functionality
Performance:
- graph optimizations
- depth-concatenation with fused relu optimization
- winograd optimizations
- deconvolution optimizations (i.e bfyx opt)

Drop 6.0

New features:
- fused winograd
- image support for weights
- yolo_region primitive support
- yolo_reorg primitive support
Bug fixes:
- winograd bias fix
- mean subtract fix
UX:
- update boost to 1.64.0
- extend graph dumps
Performance:
- update offline caches for newer drivers
- conv1x1 byxf optimization
- conv1x1 with images
- cascade depth concatenation fuse optimization

Drop 5.0

New features:
- split primitive
- upsampling primitive
- add preliminary Coffe Lake support
- uint8 weights support
- versioning
- offline autotuner cache
- Winograd phase 1 - not used yet
Bug fixes:
- in-place crop optimization bug fix
- output spatial padding in yxfb kernels fix
- local work sizes fix in softmax
- underflow fix in batch normalization
- average pooling corner case fix
UX:
- graph logger, dumps graphwiz format files
- extended documentation with API diagram and graph compilation steps
Performance:
- softmax optimization
- lrn within channel optimization
- priorbox optimization
- constant propagation

Drop 4.0

New features:
- OOOQ execution model implementation
- depthwise separable convolution implementation
- kernel auto-tuner implementation
Bug fixes:
- dump hidden layer fix
- run single layer fix
- reshape fix
UX:
- enable RTTI
- better error handling/reporting
Performance:
- lrn optimization
- dynamic pruning for sparse fc layers
- reorder optimization
- concatenation optimization
- eltwise optimization
- activation fusing 

Drop 3.0

Added:
- kernel selector
- custom layer
Changed:
- performance improvments
- bug fixes (deconvolution, softmax, reshape)
- apply fixes from community reported issues

Drop 2.0

Added:
- step by step tutorial
Changed:
- perfomance optimization for: softmax, fully connected, eltwise, reshape
- bug fixes (conformance)

Drop 1.0

- initial drop of clDNN

Support

Please report issues and suggestions GitHub issues.

How to Contribute

We welcome community contributions to clDNN. If you have an idea how to improve the library:

  • Share your proposal via GitHub issues
  • Ensure you can build the product and run all the examples with your patch
  • In the case of a larger feature, create a test
  • Submit a pull request

We will review your contribution and, if any additional fixes or modifications are necessary, may provide feedback to guide you. When accepted, your pull request will be merged into our internal and GitHub repositories.

System Requirements

clDNN supports Intel® HD Graphics and Intel® Iris® Graphics and is optimized for

  • Codename Skylake:
    • Intel® HD Graphics 510 (GT1, client market)
    • Intel® HD Graphics 515 (GT2, client market)
    • Intel® HD Graphics 520 (GT2, client market)
    • Intel® HD Graphics 530 (GT2, client market)
    • Intel® Iris® Graphics 540 (GT3e, client market)
    • Intel® Iris® Graphics 550 (GT3e, client market)
    • Intel® Iris® Pro Graphics 580 (GT4e, client market)
    • Intel® HD Graphics P530 (GT2, server market)
    • Intel® Iris® Pro Graphics P555 (GT3e, server market)
    • Intel® Iris® Pro Graphics P580 (GT4e, server market)
  • Codename Apollolake:
    • Intel® HD Graphics 500
    • Intel® HD Graphics 505
  • Codename Kabylake:
    • Intel® HD Graphics 610 (GT1, client market)
    • Intel® HD Graphics 615 (GT2, client market)
    • Intel® HD Graphics 620 (GT2, client market)
    • Intel® HD Graphics 630 (GT2, client market)
    • Intel® Iris® Graphics 640 (GT3e, client market)
    • Intel® Iris® Graphics 650 (GT3e, client market)
    • Intel® HD Graphics P630 (GT2, server market)
    • Intel® Iris® Pro Graphics 630 (GT2, server market)

clDNN currently uses OpenCL™ with multiple Intel® OpenCL™ extensions and requires Intel® Graphics Driver to run.

clDNN requires CPU with Intel® SSE/Intel® AVX support.


The software dependencies are:

  • CMake* 3.9 or later
  • C++ compiler with partial or full C++11 standard support compatible with:
    • GNU* Compiler Collection 4.8.2
    • clang 3.5 or later
    • Intel® C++ Compiler 17.0 or later
    • Visual C++ 2015 (MSVC++ 19.0) or later

Intel® CPU intrinsics header (<immintrin.h>) must be available during compilation.

  • python™ 2.7 or later (scripts are both compatible with python™ 2.7.x and python™ 3.x)
  • (optional) Doxygen* 1.8.13 or later
    Needed for manual generation of documentation from inline comments or running docs custom target which will generate it automatically.

GraphViz* (2.38 or later) is also recommended to generate documentation with all embedded diagrams.
(Make sure that dot application is visible in the PATH environment variable.)


We recommend to use latest for Linux link and 24.20 driver for Windows link.

Installation

Building

Download clDNN source code or clone the repository to your system:

    git clone  https://github.com/intel/cldnn.git

Satisfy all software dependencies and ensure that the versions are correct before building.

clDNN uses multiple 3rd-party components. They are stored in binary form in common subdirectory. Currently they are prepared for MSVC++ and GCC*. They will be cloned with repository.


clDNN uses a CMake-based build system. You can use CMake command-line tool or CMake GUI (cmake-gui) to generate required solution.
For Windows system, you can call in cmd (or powershell):

    @REM Generate 32-bit solution (solution contains multiple build configurations)...
    cmake -E make_directory build && cd build && cmake -G "Visual Studio 14 2015" ..
    @REM Generate 64-bit solution (solution contains multiple build configurations)...
    cmake -E make_directory build && cd build && cmake -G "Visual Studio 14 2015 Win64" ..

Created solution can be opened in Visual Studio 2015 or built using appropriate msbuild tool (you can also use cmake --build . to select build tool automatically).

For Unix and Linux systems:

    @REM Create GNU makefile for release clDNN and build it...
    cmake -E make_directory build && cd build && cmake -DCMAKE_BUILD_TYPE=Release .. && make
    @REM Create Ninja makefile for debug clDNN and build it...
    cmake -E make_directory build && cd build && cmake -G Ninja -DCMAKE_BUILD_TYPE=Debug .. && ninja -k 20

You can call also scripts in main directory of project which will create solutions/makefiles for clDNN (they will generate solutions/makefiles in build subdirectory and binary outputs will be written to build/out subdirectory):

  • create_msvc_mscc.bat (Windows*, Visual Studio* 2015)
  • create_unixmake_gcc.sh [Y|N] [<devtoolset-version>] (Linux*, GNU* or Ninja* makefiles, optional devtoolset support)
    • If you specify the first parameter as Y, the Ninja makefiles will be generated.
    • If you specify second parameter (number), the CMake will be called via scl with selected devtoolset version.

CMake solution offers multiple options which you can specify using normal CMake syntax (-D<option-name>=<value>):

CMake option Type Description
CMAKE_BUILD_TYPE STRING Build configuration that will be used by generated makefiles (it does not affect multi-configuration generators like generators for Visual Studio solutions). Currently supported: Debug (default), Release
CMAKE_INSTALL_PREFIX PATH Install directory prefix.
CLDNN__ARCHITECTURE_TARGET STRING Architecture of target system (where binary output will be deployed). CMake will try to detect it automatically (based on selected generator type, host OS and compiler properties). Specify this option only if CMake has problem with detection. Currently supported: Windows32, Windows64, Linux64
CLDNN__OUTPUT_DIR (CLDNN__OUTPUT_BIN_DIR, CLDNN__OUTPUT_LIB_DIR) PATH Location where built artifacts will be written to. It is set automatically to roughly build/out/<arch-target>/<build-type> subdirectory. For more control use: CLDNN__OUTPUT_LIB_DIR (specifies output path for static libraries) or CLDNN__OUTPUT_BIN_DIR (for shared libs and executables).
CMake advanced option Type Description
PYTHON_EXECUTABLE FILEPATH Path to Python interpreter. CMake will try to detect Python. Specify this option only if CMake has problem with locating Python.
CLDNN__IOCL_ICD_USE_EXTERNAL BOOL Use this option to enable use of external Intel® OpenCL™ SDK as a source for ICD binaries and headers (based on INTELOCLSDKROOT environment variable). Default: OFF
CLDNN__IOCL_ICD_VERSION STRING Version of Intel® OpenCL™ ICD binaries and headers to use (from common subdirectory). It is automatically setected by CMake (highest version). Specify, if you have multiple versions and want to use different than automatically selected.
CLDNN__COMPILE_LINK_ALLOW_UNSAFE_SIZE_OPT BOOL Allow unsafe optimizations during linking (like aggressive dead code elimination, etc.). Default: ON
CLDNN__COMPILE_LINK_USE_STATIC_RUNTIME BOOL Link with static C++ runtime. Default: OFF (shared C++ runtime is used)
CLDNN__INCLUDE_CORE BOOL Include core clDNN library project in generated makefiles/solutions. Default: ON
CLDNN__INCLUDE_TESTS BOOL Include tests application project (based on googletest framework) in generated makefiles/solutions . Default: ON
CLDNN__RUN_TESTS BOOL Run tests after building tests project. This option requires CLDNN__INCLUDE_TESTS option to be ON. Default: OFF
CLDNN__CMAKE_DEBUG BOOL Enable extended debug messages in CMake. Default: OFF

clDNN includes unit tests implemented using the googletest framework. To validate your build, run tests target, e.g.:

    make tests

(Make sure that both CLDNN__INCLUDE_TESTS and CLDNN__RUN_TESTS were set to ON when invoking CMake.)

Generating documentation

Documentation is provided inline and can be generated in HTML format with Doxygen. We recommend to use latest Doxygen* and GraphViz*.

Documentation templates and configuration files are stored in docs subdirectory. You can simply call:

    cd docs && doxygen

to generate HTML documentation in docs/html subdirectory.

There is also custom CMake target named docs which will generate documentation in CLDNN__OUTPUT_BIN_DIR/html directory. For example, when using Unix makefiles, you can run:

    make docs

in order to create it.

Deployment

Special install target will place the API header files and libraries in /usr/local (C:/Program Files/clDNN or C:/Program Files (x86)/clDNN on Windows). To change the installation path, use the option -DCMAKE_INSTALL_PREFIX=<prefix> when invoking CMake.


* Other names and brands may be claimed as the property of others.

Copyright © 2017, Intel® Corporation

More Repositories

1

hyperscan

High-performance regular expression matching library
C++
4,478
star
2

acat

Assistive Context-Aware Toolkit (ACAT)
C#
3,191
star
3

haxm

Intel® Hardware Accelerated Execution Manager (Intel® HAXM)
C
3,029
star
4

appframework

The definitive HTML5 mobile javascript framework
CSS
2,435
star
5

neural-compressor

SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX Runtime
Python
2,182
star
6

intel-extension-for-transformers

⚡ Build your chatbot within minutes on your favorite device; offer SOTA compression techniques for LLMs; run LLMs efficiently on Intel Platforms⚡
Python
2,122
star
7

pcm

Intel® Performance Counter Monitor (Intel® PCM)
C++
2,083
star
8

intel-extension-for-pytorch

A Python package for extending the official PyTorch that can easily obtain performance on Intel platform
Python
1,203
star
9

linux-sgx

Intel SGX for Linux*
C++
1,180
star
10

scikit-learn-intelex

Intel(R) Extension for Scikit-learn is a seamless way to speed up your Scikit-learn application
Python
954
star
11

llvm

Intel staging area for llvm.org contribution. Home for Intel LLVM-based projects.
918
star
12

nemu

ARCHIVED: Modern Hypervisor for the Cloud. See https://github.com/cloud-hypervisor/cloud-hypervisor instead
C
915
star
13

compute-runtime

Intel® Graphics Compute Runtime for oneAPI Level Zero and OpenCL™ Driver
C++
912
star
14

caffe

This fork of BVLC/Caffe is dedicated to improving performance of this deep learning framework when running on CPU, in particular Intel® Xeon processors.
C++
850
star
15

isa-l

Intelligent Storage Acceleration Library
C
816
star
16

media-driver

C
783
star
17

cve-bin-tool

The CVE Binary Tool helps you determine if your system includes known vulnerabilities. You can scan binaries for over 200 common, vulnerable components (openssl, libpng, libxml2, expat and others), or if you know the components used, you can get a list of known vulnerabilities associated with an SBOM or a list of components and versions.
Python
721
star
18

intel-cmt-cat

User space software for Intel(R) Resource Director Technology
C
630
star
19

fastuidraw

C++
603
star
20

optimization-manual

Contains the source code examples described in the "Intel® 64 and IA-32 Architectures Optimization Reference Manual"
Assembly
602
star
21

libipt

libipt - an Intel(R) Processor Trace decoder library
C
594
star
22

libxcam

libXCam is a project for extended camera(not limited in camera) features and focus on image quality improvement and video analysis. There are lots features supported in image pre-processing, image post-processing and smart analysis. This library makes GPU/CPU/ISP working together to improve image quality. OpenCL is used to improve performance in different platforms.
C++
590
star
23

libva

Libva is an implementation for VA-API (Video Acceleration API)
C
558
star
24

intel-graphics-compiler

C++
503
star
25

wds

Wireless Display Software For Linux OS (WDS)
C++
496
star
26

thermal_daemon

Thermal daemon for IA
C++
485
star
27

x86-simd-sort

C++ header file library for high performance SIMD based sorting algorithms for primitive datatypes
C++
485
star
28

Intel-Linux-Processor-Microcode-Data-Files

466
star
29

gvt-linux

C
463
star
30

kernel-fuzzer-for-xen-project

Kernel Fuzzer for Xen Project (KF/x) - Hypervisor-based fuzzing using Xen VM forking, VMI & AFL
C
441
star
31

tinycbor

Concise Binary Object Representation (CBOR) Library
C
432
star
32

openfl

An open framework for Federated Learning.
Python
427
star
33

cc-oci-runtime

OCI (Open Containers Initiative) compatible runtime for Intel® Architecture
C
415
star
34

tinycrypt

tinycrypt is a library of cryptographic algorithms with a focus on small, simple implementation.
C
373
star
35

compile-time-init-build

C++ library for composing modular firmware at compile-time.
C++
372
star
36

ARM_NEON_2_x86_SSE

The platform independent header allowing to compile any C/C++ code containing ARM NEON intrinsic functions for x86 target systems using SIMD up to SSE4 intrinsic functions
C
369
star
37

yarpgen

Yet Another Random Program Generator
C++
357
star
38

intel-device-plugins-for-kubernetes

Collection of Intel device plugins for Kubernetes
Go
356
star
39

QAT_Engine

Intel QuickAssist Technology( QAT) OpenSSL Engine (an OpenSSL Plug-In Engine) which provides cryptographic acceleration for both hardware and optimized software using Intel QuickAssist Technology enabled Intel platforms. https://developer.intel.com/quickassist
C
356
star
40

linux-sgx-driver

Intel SGX Linux* Driver
C
334
star
41

safestringlib

C
328
star
42

xess

C
313
star
43

idlf

Intel® Deep Learning Framework
C++
311
star
44

ad-rss-lib

Library implementing the Responsibility Sensitive Safety model (RSS) for Autonomous Vehicles
C++
298
star
45

intel-vaapi-driver

VA-API user mode driver for Intel GEN Graphics family
C
289
star
46

ipp-crypto

C
269
star
47

rohd

The Rapid Open Hardware Development (ROHD) framework is a framework for describing and verifying hardware in the Dart programming language. ROHD enables you to build and traverse a graph of connectivity between module objects using unrestricted software.
Dart
256
star
48

opencl-intercept-layer

Intercept Layer for Debugging and Analyzing OpenCL Applications
C++
255
star
49

FSP

Intel(R) Firmware Support Package (FSP)
C
244
star
50

dffml

The easiest way to use Machine Learning. Mix and match underlying ML libraries and data set sources. Generate new datasets or modify existing ones with ease.
Python
244
star
51

userspace-cni-network-plugin

Go
242
star
52

intel-ipsec-mb

Intel(R) Multi-Buffer Crypto for IPSec
C
238
star
53

isa-l_crypto

Assembly
232
star
54

confidential-computing-zoo

Confidential Computing Zoo provides confidential computing solutions based on Intel SGX, TDX, HEXL, etc. technologies.
CMake
229
star
55

bmap-tools

BMAP Tools
Python
227
star
56

intel-extension-for-tensorflow

Intel® Extension for TensorFlow*
C++
226
star
57

ozone-wayland

Wayland implementation for Chromium Ozone classes
C++
214
star
58

intel-qs

High-performance simulator of quantum circuits
C++
202
star
59

SGXDataCenterAttestationPrimitives

C++
202
star
60

intel-sgx-ssl

Intel® Software Guard Extensions SSL
C
197
star
61

msr-tools

C
195
star
62

depth-camera-web-demo

JavaScript
194
star
63

rmd

Go
189
star
64

CPU-Manager-for-Kubernetes

Kubernetes Core Manager for NFV workloads
Python
187
star
65

asynch_mode_nginx

C
186
star
66

hexl

Intel®️ Homomorphic Encryption Acceleration Library accelerates modular arithmetic operations used in homomorphic encryption
C++
181
star
67

ros_object_analytics

C++
177
star
68

zephyr.js

JavaScript* Runtime for Zephyr* OS
C
176
star
69

generic-sensor-demos

HTML
175
star
70

ipmctl

C
172
star
71

sgx-ra-sample

C++
171
star
72

lmbench

C
171
star
73

cri-resource-manager

Kubernetes Container Runtime Interface proxy service with hardware resource aware workload placement policies
Go
170
star
74

platform-aware-scheduling

Enabling Kubernetes to make pod placement decisions with platform intelligence.
Go
165
star
75

virtual-storage-manager

Python
165
star
76

PerfSpect

System performance characterization tool based on linux perf
Python
164
star
77

he-transformer

nGraph-HE: Deep learning with Homomorphic Encryption (HE) through Intel nGraph
C++
163
star
78

systemc-compiler

This tool translates synthesizable SystemC code to synthesizable SystemVerilog.
C++
155
star
79

webml-polyfill

Deprecated, the Web Neural Network Polyfill project has been moved to https://github.com/webmachinelearning/webnn-polyfill
Python
153
star
80

pmem-csi

Persistent Memory Container Storage Interface Driver
Go
151
star
81

libyami

Yet Another Media Infrastructure. it is core part of media codec with hardware acceleration, it is yummy to your video experience on Linux like platform.
C++
148
star
82

ros_openvino_toolkit

C++
147
star
83

rib

Rapid Interface Builder (RIB) is a browser-based design tool for quickly prototyping and creating the user interface for web applications. Layout your UI by dropping widgets onto a canvas. Run the UI in an interactive "Preview mode". Export the generated HTML and Javascript. It's that simple!
JavaScript
147
star
84

ideep

Intel® Optimization for Chainer*, a Chainer module providing numpy like API and DNN acceleration using MKL-DNN.
C++
145
star
85

libva-utils

Libva-utils is a collection of tests for VA-API (VIdeo Acceleration API)
C
144
star
86

gmmlib

C++
141
star
87

numatop

NumaTOP is an observation tool for runtime memory locality characterization and analysis of processes and threads running on a NUMA system.
C
139
star
88

ros2_grasp_library

C++
138
star
89

XBB

C++
133
star
90

tdx-tools

Cloud Stack and Tools for Intel TDX (Trust Domain Extension)
C
131
star
91

ros2_intel_realsense

This project is deprecated and no more maintained. Please visit https://github.com/IntelRealSense/realsense-ros for ROS2 wrapper.
C++
131
star
92

linux-intel-lts

C
131
star
93

CeTune

Python
130
star
94

cm-compiler

C++
130
star
95

pti-gpu

Profiling Tools Interfaces for GPU (PTI for GPU) is a set of Getting Started Documentation and Tools Library to start performance analysis on Intel(R) Processor Graphics easily
C++
129
star
96

fMBT

Free Model Based tool
Python
129
star
97

zlib

C
128
star
98

ros_intel_movidius_ncs

C++
126
star
99

mpi-benchmarks

C
125
star
100

mOS

C
124
star