Skip to content

AutoKernel 是一个简单易用,低门槛的自动算子优化工具,提高深度学习算法部署效率。

License

Notifications You must be signed in to change notification settings

dreamplayer-zhang/AutoKernel

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

33 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

English | 简体中文

AutoKernel

Introduction

Neural networks are now used in a wide variety of applications. Efficient execution of Neural networks on various devices plays a critical role for these applications. Facing the rapid evolution of deep learning algorithms, there're limited qualified programmers to write hand optimized low-level kernels on different hardware platforms. Using automatic optimization tools to generate high-performance implementations become a promising solution.

AutoKernel began as a research project at OPEN AI LAB. The project is now open source. AutoKernel is an operator optimzation tools for automatically generating high-performance low-level codes for diverse hardware backends. It aims to accelerate the development of high performance operators on various hardware including specialized accelerators.

AutoKernel Architecture

AutoKernel arch

AutoKernel consists of three modules:

  • Operator Generator:

    This module uses the open source project Halide. Halide is a domain specific language (DSL), embedded in C++, designed to make it easier to write high-performance image processing code on modern machines. Halide seperates the algorithm description from its schedule. The input of this module is the algorithm description of operator, and the output is compiled optimized assembly code/object file for corresponding back-ends.

  • AutoSearch

    AutoSearch is an automatic module for searching optimized schedules for halide operators, using multiple optimization algorithms (greedy algorithm, reinforce learning, marchine learning, ...). It supports searching optimized schedules on both CPU and GPU, and generate code files running on different platforms (x86 or arm). This module is still under developping.

  • AutoKernel Plugin:

    AutoKernel Plugin realizes one-click integration of auto-generated optimized operator codes into Tengine, without modifying the core code base of Tengine. AutoKernel plugin realizes the one-click deployment of the automatic generated operator implements.

Features

  • Automated
  • Efficient
  • User-friendly

Docker

We provide following dockers with Halide and Tengine installed:

  • cpu: openailab/autokernel
  • cuda: openailab/autokernel:cuda
  • opencl: openailab/autokernel:opencl

Detail Dockerfiles, see Dockerfiles

[NOTE]: if using the cuda image, you need use nvidia-docker instead of docker, here's nvidia-docker install-guide.

nvidia-docker pull openailab/autokernel:cuda
nvidia-docker run -it openailab/autokernel:cuda /bin/bash

License

Discussion

  • Github issues
  • QQ group: 829565581

About

AutoKernel 是一个简单易用,低门槛的自动算子优化工具,提高深度学习算法部署效率。

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages

  • C++ 87.3%
  • C 9.8%
  • Python 1.5%
  • Makefile 0.4%
  • Shell 0.4%
  • CMake 0.3%
  • Jupyter Notebook 0.3%