The implementation of DeBERTa
-
Updated
Sep 29, 2023 - Python
The implementation of DeBERTa
Pre-training of Deep Bidirectional Transformers for Language Understanding: pre-train TextCNN
[NeurIPS‘2021] "TransGAN: Two Pure Transformers Can Make One Strong GAN, and That Can Scale Up", Yifan Jiang, Shiyu Chang, Zhangyang Wang
Implementation of Transformer Model in Tensorflow
[IGARSS'22]: A Transformer-Based Siamese Network for Change Detection
Seq2SeqSharp is a tensor based fast & flexible deep neural network framework written by .NET (C#). It has many highlighted features, such as automatic differentiation, different network types (Transformer, LSTM, BiLSTM and so on), multi-GPUs supported, cross-platforms (Windows, Linux, x86, x64, ARM), multimodal model for text and images and so on.
pretrained BERT model for cyber security text, learned CyberSecurity Knowledge
Multi-module Recurrent Convolutional Neural Network with Transformer Encoder for ECG Arrhythmia Classification
This repository contains PyTorch implementations of the models from the paper An Empirical Study MIME: MIMicking Emotions for Empathetic Response Generation.
The repo is for the Heart Disease classification project using Transformer Encoders in PyTorch.
题目知识点预测标注。Question knowledge point prediction.
PyTorch implementation of RealFormer: Transformer Likes Residual Attention
Universal Transforming Geometric Network for protein structure prediction.
Code for the ACL 2019 paper "Observing Dialogue in Therapy: Categorizing and Forecasting Behavioral Codes"
Transformer OCR is a Optical Character Recognition tookit built for researchers working on both OCR for both Vietnamese and English. This project only focused on variants of vanilla Transformer (Conformer) and Feature Extraction (CNN-based approach).
Official Pytorch implementation of (Roles and Utilization of Attention Heads in Transformer-based Neural Language Models), ACL 2020
Contains pytorch implementation of Transformer and EvolvedTransformer architectures. WIP
A Transformer Implementation that is easy to understand and customizable.
CASPR is a deep learning framework applying transformer architecture to learn and predict from tabular data at scale.
Generating English Rock lyrics using BERT
Add a description, image, and links to the transformer-encoder topic page so that developers can more easily learn about it.
To associate your repository with the transformer-encoder topic, visit your repo's landing page and select "manage topics."