ist-daslab Goto Github PK
Name: IST Austria Distributed Algorithms and Systems Lab
Type: Organization
Name: IST Austria Distributed Algorithms and Systems Lab
Type: Organization
Code for reproducing "AC/DC: Alternating Compressed/DeCompressed Training of Deep Neural Networks" (NeurIPS 2021)
Repository for Correlation Aware Prune (NeurIPS23) source and experimental code
ConvNet training using pytorch
Code for reproducing the results from "CrAM: A Compression-Aware Minimizer" accepted at ICLR 2023
Deep Learning Examples
Neural Network Distiller by Intel AI Lab: a Python package for neural network compression research. https://nervanasystems.github.io/distiller
working repository for the the IST daslab micronet challenge
The repository contains code to reproduce the experiments from our paper Error Feedback Can Accurately Compress Preconditioners available below:
GPTQ with finetuning
Stripped version of gcomp_sim for ML course
Code for the ICLR 2023 paper "GPTQ: Accurate Post-training Quantization of Generative Pretrained Transformers".
GridSearcher simplifies running grid searches for machine learning projects in Python, emphasizing parallel execution and GPU scheduling without dependencies on SLURM or other workload managers.
Distributed training framework for TensorFlow, Keras, PyTorch, and Apache MXNet.
Code for the experiments in Knowledge Distillation Performs Partial Variance Reduction, NeurIPS 2023
LLM training code for Databricks foundation models
Efficient reference implementations of the static & dynamic M-FAC algorithms (for pruning and optimization)
FP16xINT4 LLM inference kernel that can achieve near-ideal ~4x speedups up to medium batchsizes of 16-32 tokens.
Code for the paper "Mathador-LM: A Dynamic Benchmark for Mathematical Reasoning on LLMs".
This repository contains code for the MicroAdam paper.
Code for the NeurIPS 2022 paper "Optimal Brain Compression: A Framework for Accurate Post-Training Quantization and Pruning".
A fork of the PEFT library, supporting Robust Adaptation (RoSA)
Code for reproducing the paper "Bias in Pruned Vision Models: In-Depth Analysis and Countermeasures"
Repository for CPU Kernel Generation for LLM Inference
Code for the paper "QMoE: Practical Sub-1-Bit Compression of Trillion-Parameter Models".
Repository for the implementation of "Distributed Principal Component Analysis with Limited Communication" (Alimisis et al., NeurIPS 2021). Parts of this code were originally based on code from "Communication-Efficient Distributed PCA by Riemannian Optimization" (Huang and Pan, ICML 2020).
Repository for the QUIK project, enabling the use of 4bit kernels for generative inference
A declarative, efficient, and flexible JavaScript library for building user interfaces.
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google ❤️ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.