Skip to content

Research and Materials on Hardware implementation of BERT (Bidirectional Encoder Representations from Transformers) Model

License

Notifications You must be signed in to change notification settings

nedaraad/bert-on-silicon

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

33 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

BERT Model on Silicon

Research and Materials on Hardware implementation of BERT (Bidirectional Encoder Representations from Transformers) Model

BERT Model

Description

BERT is a method of pre-training language representations, meaning that we train a general-purpose "language understanding" model on a large text corpus (like Wikipedia) and then use that model for downstream NLP tasks. BERT was created and published in 2018 by Jacob Devlin and his colleagues from Google. BERT is conceptually simple and empirically powerful. It obtains new state-of-the-art results on eleven natural language processing tasks, including pushing the GLUE score to 80.5% (7.7% point absolute improvement), MultiNLI accuracy to 86.7 (5.6% absolute improvement), SQuAD v1.1 question answering Test F1 to 93.2 (1.5% absolute improvement) and SQuAD v2.0 Test F1 to 83.1 (5.1% absolute improvement).

Architecture

BERT is a Transformer-based model. The architecture of BERT is similar to the original Transformer model, except that BERT has two separate Transformer models: one for the left-to-right direction (the “encoder”) and one for the right-to-left direction (the “encoder”). The output of each model is the hidden state output by the final Transformer layer. The two models are pre-trained jointly on a large corpus of unlabeled text. The pre-training task is a simple and straightforward masked language modeling objective. The pre-trained BERT model can then be fine-tuned with just one additional output layer to create state-of-the-art models for a wide range of tasks, such as question answering and language inference, without substantial task-specific architecture modifications.


Reference Papers

1. Attention Is All You Need

DOI-Link PDF-Download

Code-Link Code-Link

The dominant sequence transduction models are based on complex recurrent or convolutional neural networks that include an encoder and a decoder. The best performing models also connect the encoder and decoder through an attention mechanism. We propose a new simple network architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two machine translation tasks show these models to be superior in quality while being more parallelizable and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014 Englishto-German translation task, improving over the existing best results, including ensembles, by over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training costs of the best models from the literature. We show that the Transformer generalizes well to other tasks by applying it successfully to English constituency parsing both with large and limited training data.

2. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding

DOI-Link PDF-Download Code-Link Code-Link

We introduce a new language representation model called BERT, which stands for Bidirectional Encoder Representations from Transformers. Unlike recent language representation models (Peters et al., 2018a; Radford et al., 2018), BERT is designed to pretrain deep bidirectional representations from unlabeled text by jointly conditioning on both left and right context in all layers. As a result, the pre-trained BERT model can be finetuned with just one additional output layer to create state-of-the-art models for a wide range of tasks, such as question answering and language inference, without substantial taskspecific architecture modifications.
BERT is conceptually simple and empirically powerful. It obtains new state-of-the-art results on eleven natural language processing tasks, including pushing the GLUE score to 80.5% (7.7% point absolute improvement), MultiNLI accuracy to 86.7% (4.6% absolute improvement), SQuAD v1.1 question answering Test F1 to 93.2 (1.5 point absolute improvement) and SQuAD v2.0 Test F1 to 83.1 (5.1 point absolute improvement).

Important Papers

Distilling the Knowledge in a Neural Network

DOI-Link PDF-Download

Distilling Knowledge Learned in BERT for Text Generation

DOI-Link PDF-Download

DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter

DOI-Link PDF-Download

TinyBERT: Distilling BERT for Natural Language Understanding

DOI-Link PDF-Download

Distilling the Knowledge in a Neural Network

DOI-Link PDF-Download

FastBERT: a Self-distilling BERT with Adaptive Inference Time

DOI-Link PDF-Download

Distilling Task-Specific Knowledge from BERT into Simple Neural Networks

DOI-Link PDF-Download

Patient Knowledge Distillation for BERT Model Compression

DOI-Link PDF-Download

MobileBERT: a Compact Task-Agnostic BERT for Resource-Limited Devices

DOI-Link PDF-Download

Improving Multi-Task Deep Neural Networks via Knowledge Distillation for Natural Language Understanding

DOI-Link PDF-Download


BERT on Silicon

2018

Algorithm-Hardware Co-Design of Single Shot Detector for Fast Object Detection on FPGAs

DOI-Link


2019

An Evaluation of Transfer Learning for Classifying Sales Engagement Emails at Large Scale

DOI-Link

Pre-trained bert-gru model for relation extraction

DOI-Link

Q8BERT: Quantized 8Bit BERT

DOI-Link PDF-Link

Structured pruning of a BERT-based question answering model

DOI-Link PDF-Link

Structured pruning of large language models

DOI-Link PDF-Link

Tinybert: Distilling bert for natural language understanding

DOI-Link PDF-Link


2020

A Low-Cost Reconfigurable Nonlinear Core for Embedded DNN Applications

DOI-Link

A^3: Accelerating Attention Mechanisms in Neural Networks with Approximation

DOI-Link

Accelerating event detection with DGCNN and FPGAS

DOI-Link

An Empirical Analysis of BERT Embedding for Automated Essay Scoring

DOI-Link

An investigation on different underlying quantization schemes for pre-trained language models

DOI-Link

ATT: A Fault-Tolerant ReRAM Accelerator for Attention-based Neural Networks

DOI-Link

Binarybert: Pushing the limit of bert quantization

DOI-Link

Capuchin: Tensor-based GPU Memory Management for Deep Learning

DOI-Link

CATBERT: Context-Aware Tiny BERT for Detecting Social Engineering Emails

DOI-Link

CatBERT: Context-Aware Tiny BERT for Detecting Targeted Social Engineering Emails

DOI-Link

ColBERT: Efficient and Effective Passage Search via Contextualized Late Interaction over BERT

DOI-Link

Combining Feature Selection Methods with BERT: An In-depth Experimental Study of Long Text Classification

DOI-Link

Comparison of Deep Learning Models and Various Text Pre-Processing Techniques for the Toxic Comments Classification

DOI-Link

Compressing BERT: Studying the Effects of Weight Pruning on Transfer Learning

DOI-Link

Deep Learning Acceleration with Neuron-to-Memory Transformation

DOI-Link

Efficient algorithms and hardware for natural language processing

DOI-Link

Efficient transformer-based large scale language representations using hardware-friendly block structured pruning

DOI-Link

FARM: A flexible accelerator for recurrent and memory augmented neural networks

DOI-Link

Fastformers: Highly efficient transformer models for natural language understanding

DOI-Link

FTRANS: energy-efficient acceleration of transformers using FPGA

DOI-Link

Hardware accelerator for multi-head attention and position-wise feed-forward in the transformer

DOI-Link

Improving Accuracy and Speeding Up Document Image Classification Through Parallel Systems

DOI-Link PDF-Link

Improving post training neural quantization: Layer-wise calibration and integer programming

DOI-Link PDF-Link

Integer quantization for deep learning inference: Principles and empirical evaluation

DOI-Link PDF-Link

Look-Up Table based Energy Efficient Processing in Cache Support for Neural Network Acceleration

DOI-Link PDF-Link

Poor Man's BERT: Smaller and Faster Transformer Models

DOI-Link PDF-Link

PoWER-BERT: Accelerating BERT Inference via Progressive Word-vector Elimination

DOI-Link PDF-Link

Pruning Redundant Mappings in Transformer Models via Spectral-Normalized Identity Prior

DOI-Link PDF-Link

Q-BERT: Hessian Based Ultra Low Precision Quantization of BERT

DOI-Link PDF-Link

ReTransformer: ReRAM-based processing-in-memory architecture for transformer acceleration

DOI-Link PDF-Link

SqueezeBERT: What can computer vision teach NLP about efficient neural networks?

DOI-Link PDF-Link

TernaryBERT: Distillation-aware Ultra-low Bit BERT

DOI-Link PDF-Link

Training Large Neural Networks with Constant Memory using a New Execution Algorithm

DOI-Link PDF-Link

Ultron-AutoML: An open-source, distributed, scalable framework for efficient hyper-parameter optimization

DOI-Link PDF-Link


2021

A Framework for Area-efficient Multi-task BERT Execution on ReRAM-based Accelerators

DOI-Link

A Full-Stack Search Technique for Domain Optimized Deep Learning Accelerators

DOI-Link

A Microcontroller is All You Need: Enabling Transformer Execution on Low-Power IoT Endnodes

DOI-Link

Accelerated Device Placement Optimization with Contrastive Learning

DOI-Link

Accelerating bandwidth-bound deep learning inference with main-memory accelerators

DOI-Link

Accelerating Emerging Neural Workloads

DOI-Link

Accelerating Framework of Transformer by Hardware Design and Model Compression Co-Optimization

DOI-Link

Accelerating Transformer-based Deep Learning Models on FPGAs using Column Balanced Block Pruning

DOI-Link PDF-Link

Accommodating Transformer onto FPGA: Coupling the Balanced Model Compression and FPGA-Implementation Optimization

DOI-Link

Adaptive Inference through Early-Exit Networks: Design, Challenges and Directions

DOI-Link

Adaptive Spatio-Temporal Graph Enhanced Vision-Language Representation for Video QA

DOI-Link

Algorithm-hardware Co-design of Attention Mechanism on FPGA Devices

DOI-Link

Aquabolt-XL: Samsung HBM2-PIM with in-memory processing for ML accelerators and beyond

DOI-Link

AUBER: Automated BERT regularization

DOI-Link

Automatic Mixed-Precision Quantization Search of BERT

DOI-Link

BERMo: What can BERT learn from ELMo?

DOI-Link

BERT Model for Classification of Fake News using the Cloud Processing Capacity

DOI-Link

Bertinho: Galician BERT representations

DOI-Link

BERxiT: Early exiting for BERT with better fine-tuning and extension to regression

DOI-Link

Beyond preserved accuracy: Evaluating loyalty and robustness of BERT compression

DOI-Link

Binary Complex Neural Network Acceleration on FPGA : (Invited Paper)

DOI-Link

Biomedical Named Entity Recognition at Scale

DOI-Link

Block pruning for faster transformers

DOI-Link

Compressing Large-Scale Transformer-Based Models: A Case Study on BERT

DOI-Link

DAP-BERT: Differentiable Architecture Pruning of BERT

DOI-Link

Demystifying BERT: Implications for Accelerator Design

DOI-Link

Dynamic-TinyBERT: Boost TinyBERT's Inference Efficiency by Dynamic Sequence Length

DOI-Link

EAGLE: Expedited Device Placement with Automatic Grouping for Large Models

DOI-Link

EBERT: Efficient BERT Inference with Dynamic Structured Pruning

DOI-Link

EdgeBERT: Sentence-Level Energy Optimizations for Latency-Aware Multi-Task NLP Inference

DOI-Link

ELSA: Hardware-Software co-design for efficient, lightweight self-attention mechanism in neural networks

DOI-Link

Enabling energy-efficient DNN training on hybrid GPU-FPGA accelerators

DOI-Link

Energy efficiency boost in the AI-infused POWER10 processor

DOI-Link

Fixed-point Quantization for Vision Transformer

DOI-Link

FlexACC: A Programmable Accelerator with Application-Specific ISA for Flexible Deep Neural Network Inference

DOI-Link

Gemmini: Enabling systematic deep-learning architecture evaluation via full-stack integration

DOI-Link

Gobo: Quantizing attention-based nlp models for low latency and energy efficient inference

DOI-Link

Hardware Acceleration of Fully Quantized BERT for Efficient Natural Language Processing

DOI-Link

Hardware acceleration of sparse and irregular tensor computations of ml models: A survey and insights

DOI-Link

HMC-TRAN: A Tensor-core Inspired Hierarchical Model Compression for Transformer-based DNNs on GPU

DOI-Link

I-BERT: Integer-only BERT Quantization

DOI-Link PDF-Link

Improving the efficiency of transformers for resource-constrained devices

DOI-Link PDF-Link

KAISA: An adaptive second-order optimizer framework for deep neural networks

DOI-Link PDF-Link

Kunlun: A 14nm High-Performance AI Processor for Diversified Workloads

DOI-Link

Layerweaver: Maximizing Resource Utilization of Neural Processing Units via Layer-Wise Scheduling

DOI-Link PDF-Link

M2M: Learning to Enhance Low-Light Image from Model to Mobile FPGA

DOI-Link

NeuralScale: A RISC-V Based Neural Processor Boosting AI Inference in Clouds

DOI-Link PDF-Link

NLP-Fast: A Fast, Scalable, and Flexible System to Accelerate Large-Scale Heterogeneous NLP Models

DOI-Link

NPE: An FPGA-based Overlay Processor for Natural Language Processing

DOI-Link PDF-Link

Predicting Efficiency/Effectiveness Trade-offs for Dense vs. Sparse Retrieval Strategy Selection

DOI-Link

PTQ4ViT: Post-Training Quantization Framework for Vision Transformers with Twin Uniform Quantization

DOI-Link PDF-Link

Randomly Wired Network Based on RoBERTa and Dialog History Attention for Response Selection

DOI-Link

Re2PIM: A Reconfigurable ReRAM-Based PIM Design for Variable-Sized Vector-Matrix Multiplication

DOI-Link

RISC-VTF: RISC-V Based Extended Instruction Set for Transformer

DOI-Link

RMSMP: A Novel Deep Neural Network Quantization Framework with Row-wise Mixed Schemes and Multiple Precisions

DOI-Link PDF-Link

Simplified TinyBERT: Knowledge Distillation for Document Retrieval

DOI-Link PDF-Link

SmaQ: Smart Quantization for DNN Training by Exploiting Value Clustering

DOI-Link PDF-Link

Softermax: Hardware/Software Co-Design of an Efficient Softmax for Transformers

DOI-Link PDF-Link

SpAtten: Efficient Sparse Attention Architecture with Cascade Token and Head Pruning

DOI-Link PDF-Link

SQuAT: Sharpness- and Quantization-Aware Training for BERT

DOI-Link PDF-Link

Stochastic precision ensemble: self-knowledge distillation for quantized deep neural networks

DOI-Link PDF-Link

Talos: A Weighted Speedup-Aware Device Placement of Deep Learning Models

DOI-Link

TR-BERT: Dynamic Token Reduction for Accelerating BERT Inference

DOI-Link PDF-Link

Training with Quantization Noise for Extreme Model Compression

DOI-Link PDF-Link

Transformer Acceleration with Dynamic Sparse Attention

DOI-Link PDF-Link

Understanding and Overcoming the Challenges of Efficient Transformer Quantization

DOI-Link PDF-Link

Vis-TOP: Visual Transformer Overlay Processor

DOI-Link PDF-Link


2022

A 28nm 27.5TOPS/W Approximate-Computing-Based Transformer Processor with Asymptotic Sparsity Speculating and Out-of-Order Computing

DOI-Link

A 40nm 5.6TOPS/W 239GOPS/mm2 Self-Attention Processor with Sign Random Projection-based Approximation

DOI-Link

A Dual-Mode Similarity Search Accelerator based on Embedding Compression for Online Cross-Modal Image-Text Retrieval

DOI-Link

A Fast and Flexible FPGA-based Accelerator for Natural Language Processing Neural Networks

DOI-Link

A Framework for Accelerating Transformer-Based Language Model on ReRAM-Based Architecture

DOI-Link

A length adaptive algorithm-hardware co-design of transformer on FPGA through sparse attention and dynamic pipelining

DOI-Link

A Resource-Saving Energy-Efficient Reconfigurable Hardware Accelerator for BERT-based Deep Neural Network Language Models using FFT Multiplication

DOI-Link

A Self-Attention Network for Deep JSCCM: The Design and FPGA Implementation

DOI-Link

Accelerating attention mechanism on fpgas based on efficient reconfigurable systolic array

DOI-Link

Accelerating attention through gradient-based learned runtime pruning

DOI-Link

Accelerating NLP Tasks on FPGA with Compressed BERT and a Hardware-Oriented Early Exit Method

DOI-Link

Accelerating Transformer Networks through Recomposing Softmax Layers

DOI-Link PDF-Link

Achieving the Performance of All-Bank In-DRAM PIM With Standard Memory Interface: Memory-Computation Decoupling

DOI-Link PDF-Link

Adaptable Butterfly Accelerator for Attention-based NNs via Hardware and Algorithm Co-design

DOI-Link PDF-Link

AlphaTuning: Quantization-Aware Parameter-Efficient Adaptation of Large-Scale Pre-Trained Language Models

DOI-Link

Alternative non-BERT model choices for the textual classification in low-resource languages and environments

DOI-Link

An Algorithm-Hardware Co-Optimized Framework for Accelerating N:M Sparse Transformers

DOI-Link

An Automatic and Efficient BERT Pruning for Edge AI Systems

DOI-Link

An Efficient Hardware Accelerator for Sparse Transformer Neural Networks

DOI-Link

An Energy-Efficient Transformer Processor Exploiting Dynamic Weak Relevances in Global Attention

DOI-Link

An FPGA-Based Transformer Accelerator Using Output Block Stationary Dataflow for Object Recognition Applications

DOI-Link

Analog-memory-based 14nm Hardware Accelerator for Dense Deep Neural Networks including Transformers

DOI-Link

Answer Fast: Accelerating BERT on the Tensor Streaming Processor

DOI-Link

ANT: Exploiting Adaptive Numerical Data Type for Low-bit Deep Neural Network Quantization

DOI-Link

APT: The master-copy-free training method for quantised neural network on edge devices

DOI-Link

Auto-ViT-Acc: An FPGA-Aware Automatic Acceleration Framework for Vision Transformer with Mixed-Scheme Quantization

DOI-Link

Balance Multi-Head Attention based on Software and Hardware Co-design

DOI-Link

BEBERT: Efficient and robust binary ensemble BERT

DOI-Link

BERT model optimization methods for inference: a comparative study of five alternative BERT-model implementations

DOI-Link

BERT on a Data Diet: Finding Important Examples by Gradient-Based Pruning

DOI-Link

BERTPerf: Inference Latency Predictor for BERT on ARM big.LITTLE Multi-Core Processors

DOI-Link

BiBERT: Accurate Fully Binarized BERT

DOI-Link

Bigger&Faster: Two-stage Neural Architecture Search for Quantized Transformer Models

DOI-Link

BiT: Robustly Binarized Multi-distilled Transformer

DOI-Link

Boosting Distributed Training Performance of the Unpadded BERT Model

DOI-Link

Compact Token Representations with Contextual Quantization for Efficient Document Re-ranking

DOI-Link

Compressing Pre-trained Transformers via Low-Bit NxM Sparsity for Natural Language Understanding

DOI-Link

Compression of Generative Pre-trained Language Models via Quantization

DOI-Link

CPSAA: Accelerating Sparse Attention using Crossbar-based Processing-In-Memory Architecture

DOI-Link

Demystifying BERT: System Design Implications

DOI-Link

DFX: A Low-latency Multi-FPGA Appliance for Accelerating Transformer-based Text Generation

DOI-Link

DiVIT: Algorithm and architecture co-design of differential attention in vision transformer

DOI-Link

DOTA: Detect and Omit Weak Attentions for Scalable Transformer Acceleration

DOI-Link

DQ-BART: Efficient Sequence-to-Sequence Model via Joint Distillation and Quantization

DOI-Link

DTQAtten: Leveraging Dynamic Token-based Quantization for Efficient Attention Architecture

DOI-Link

Dynamic Precision Analog Computing for Neural Networks

DOI-Link

EFA-Trans: An Efficient and Flexible Acceleration Architecture for Transformers

DOI-Link

Elastic Processing and Hardware Architectures for Machine Learning

DOI-Link

Enabling and Accelerating Dynamic Vision Transformer Inference for Real-Time Applications

DOI-Link

Enabling Efficient Large-Scale Deep Learning Training with Cache Coherent Disaggregated Memory Systems

DOI-Link

Enabling Energy-Efficient Inference for Self-Attention Mechanisms in Neural Networks

DOI-Link

Enabling fast uncertainty estimation: accelerating bayesian transformers via algorithmic and hardware optimizations

DOI-Link

Enabling Fast Uncertainty Estimation: Exploiting Structured Sparsity in Bayesian Transformers

DOI-Link

Ensemble Model Compression for Fast and Energy-Efficient Ranking on FPGAs

DOI-Link

Extending the ONNX Runtime Framework for the Processing-in-Memory Execution

DOI-Link

Fast Heterogeneous Task Mapping for Reducing Edge DNN Latency

DOI-Link

FILM-QNN: Efficient FPGA Acceleration of Deep Neural Networks with Intra-Layer, Mixed-Precision Quantization

DOI-Link

FPGA-aware automatic acceleration framework for vision transformer with mixed-scheme quantization: late breaking results

DOI-Link

FPGA-based design and implementation of the location attention mechanism in neural networks

DOI-Link

Future Scaling of Memory Hierarchy for Tensor Cores and Eliminating Redundant Shared Memory Traffic Using Inter-Warp Multicastin

DOI-Link

Greedy-layer pruning: Speeding up transformer models for natural language processing

DOI-Link

GuardNN: secure accelerator architecture for privacy-preserving deep learning

DOI-Link

Handling heavy-tailed input of transformer inference on GPUs

DOI-Link

Hardware Acceleration of Transformer Networks using FPGAs

DOI-Link

Hardware and Software Co-design for Soft Switch in ViT Variants Processing Unit

DOI-Link

Hardware and Software Co-optimization for Windows Attention

DOI-Link

Improving Oversubscribed GPU Memory Performance in the PyTorch Framework

DOI-Link

LLM.int8(): 8-bit Matrix Multiplication for Transformers at Scale

DOI-Link PDF-Link

Low-Precision Quantization Techniques for Hardware-Implementation-Friendly BERT Models

DOI-Link

MKQ-BERT: Quantized BERT with 4-bits Weights and Activations

DOI-Link PDF-Link

Mokey: enabling narrow fixed-point inference for out-of-the-box floating-point transformer models

DOI-Link PDF-Link

Near-Optimal Sparse Allreduce for Distributed Deep Learning

DOI-Link PDF-Link

Optimal Brain Compression: A framework for accurate post-training quantization and pruning

DOI-Link PDF-Link

PipeBERT: High-throughput BERT Inference for ARM Big.LITTLE Multi-core Processors

DOI-Link

Post-Training Quantization for Longformer with Chunkwise Quantization Granularity and Optimized Percentile

DOI-Link

Pre-trained Language Model with Feature Reduction and No Fine-Tuning

DOI-Link

Privacy-Preserving Text Classification on BERT Embeddings with Homomorphic Encryption

DOI-Link PDF-Link

ProSE: the architecture and design of a protein discovery engine

DOI-Link PDF-Link

QDrop: Randomly Dropping Quantization for Extremely Low-bit Post-Training Quantization

DOI-Link PDF-Link

QuaLA-MiniLM: a Quantized Length Adaptive MiniLM

DOI-Link PDF-Link

RCT: Resource Constrained Training for Edge AI

DOI-Link PDF-Link

ReAAP: A Reconfigurable and Algorithm-Oriented Array Processor With Compiler-Architecture Co-Design

DOI-Link PDF-Link

Row-wise Accelerator for Vision Transformer

DOI-Link PDF-Link

S4: a High-sparsity, High-performance AI Accelerator

DOI-Link PDF-Link

SALO: an efficient spatial accelerator enabling hybrid sparse attention mechanisms for long sequences

DOI-Link PDF-Link

SensiMix: Sensitivity-Aware 8-bit index & 1-bit value mixed precision quantization for BERT compression

DOI-Link PDF-Link

Sentiment Analysis Using Pre-Trained Language Model With No Fine-Tuning and Less Resource

DOI-Link PDF-Link

Software and Hardware Fusion Multi-Head Attention

DOI-Link

Sparse Attention Acceleration with Synergistic In-Memory Pruning and On-Chip Recomputation

DOI-Link PDF-Link

SwiftPruner: Reinforced Evolutionary Pruning for Efficient Ad Relevance

DOI-Link PDF-Link

T-OPU: An FPGA-based Overlay Processor for Natural Language Processing

DOI-Link PDF-Link

The Optimal BERT Surgeon: Scalable and Accurate Second-Order Pruning for Large Language Models

DOI-Link PDF-Link

Towards efficient post-training quantization of pre-trained language models

DOI-Link PDF-Link

Train Flat, Then Compress: Sharpness-Aware Minimization Learns More Compressible Models

DOI-Link PDF-Link

TranCIM: Full-Digital Bitline-Transpose CIM-based Sparse Transformer Accelerator With Pipeline/Parallel Reconfigurable Modes

DOI-Link

TransPIM: A Memory-based Acceleration via Software-Hardware Co-Design for Transformer

DOI-Link PDF-Link

VAQF: Fully Automatic Software-Hardware Co-Design Framework for Low-Bit Vision Transformer

DOI-Link PDF-Link

Varuna: Scalable, Low-cost Training of Massive Deep Learning Models

DOI-Link

ViA: A Novel Vision-Transformer Accelerator Based on FPGA

DOI-Link


2023

An Efficient Transformer Inference Engine on DSP

DOI-Link

CHARM: Composing Heterogeneous Accelerators for Matrix Multiply on Versal ACAP Architecture

DOI-Link

DTATrans: Leveraging Dynamic Token-Based Quantization With Accuracy Compensation Mechanism for Efficient Transformer Architecture

DOI-Link

HAMMER: Hardware-friendly Approximate Computing for Self-attention with Mean-redistribution and Linearization

DOI-Link

ViTA: A Vision Transformer Inference Accelerator for Edge Applications

DOI-Link PDF-Link

TRON: Transformer Neural Network Acceleration with Non-Coherent Silicon Photonics

DOI-Link PDF-Link

SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models

DOI-Link PDF-Link

Sparse*BERT: Sparse Models Generalize To New tasks and Domains

DOI-Link PDF-Link

Teacher Intervention: Improving Convergence of Quantization Aware Training for Ultra-Low Precision Transformers

DOI-Link PDF-Link

TiC-SAT: Tightly-Coupled Systolic Accelerator for Transformers

DOI-Link PDF-Link

ViTALiTy: Unifying Low-rank and Sparse Approximation for Vision Transformer Acceleration with a Linear Taylor Attention

DOI-Link PDF-Link

Trends in AI inference energy consumption: Beyond the performance-vs-parameter laws of deep learning

DOI-Link PDF-Link

TransCODE: Co-design of Transformers and Accelerators for Efficient Training and Inference

DOI-Link PDF-Link

Architecting High Performance Silicon Systems for Accurate and Efficient On-Chip Deep Learning

DOI-Link PDF-Link

--

About

Research and Materials on Hardware implementation of BERT (Bidirectional Encoder Representations from Transformers) Model

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%