Huang-Cheng Chou's Projects
A library for performing coverage guided fuzzing of neural networks
Data loaders and abstractions for text and NLP
:deciduous_tree:从中文文本中自动提取关键词和摘要
A list of all named GANs!
Transformer seq2seq model, program that can build a language translator from parallel corpus
Code for the paper "Efficient Adaption of Pretrained Transformers for Abstractive Summarization"
Official PyTorch implementation of Long-Short Transformer (NeurIPS 2021).
Sequence labeling base on universal transformer (Transformer encoder) and CRF; 基于Universal Transformer + CRF 的中文分词和词性标注
This repository includes the official project of TransUNet, presented in our paper: TransUNet: Transformers Make Strong Encoders for Medical Image Segmentation.
Trax — Deep Learning with Clear Code and Speed
Code to support characterizing sources of uncertainty to proxy calibration and disambiguate annotator and data bias.
High-quality implementations of standard and SOTA methods on a variety of tasks.
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
Implementation of Universal Transformer in Pytorch
Calibration library and code for the paper: Verified Uncertainty Calibration. Ananya Kumar, Percy Liang, Tengyu Ma. NeurIPS 2019 (Spotlight).
Code for the ICML 2021 (long talk) paper: "ViLT: Vision-and-Language Transformer Without Convolution or Region Supervision"
Code and data for the framework in "Multimodal Pretraining Unmasked: A Meta-Analysis and a Unified Framework of Vision-and-Language BERTs", TACL 2021.
A large-scale multilingual speech corpus for representation learning, semi-supervised learning and interpretation
Implementation of VQ-VAE for audio
Code and Experiments for ACL-IJCNLP 2021 Paper "Mind Your Outliers! Investigating the Negative Impact of Outliers on Active Learning for Visual Question Answering."
How to use our public wav2vec2 dimensional emotion model
WaveNet auto-ancoders for ZeroSpeech challenge 2020
This repository contains lists of state-or-art weakly supervised semantic segmentation works
Scalable, Portable and Distributed Gradient Boosting (GBDT, GBRT or GBM) Library, for Python, R, Java, Scala, C++ and more. Runs on single machine, Hadoop, Spark, Flink and DataFlow
XLNet: Generalized Autoregressive Pretraining for Language Understanding