Git Product home page Git Product logo

l2l_mlprague23's Introduction

Learning to Learn workshop at MLPrague 2023

This repo contains link to the materials for the Learning to Learn workshop on Machine Learning Prague 2023.

Getting started

The primary playground env for the exercises below is Google Colab. The linked Colab notebooks contain the resolution of dependences, but if you'd like to run the exercises elsewhere, simply install the attached requirements.txt into any environment:

git clone https://github.com/gaussalgo/L2L_MLPrague23.git
pip install -r L2L_MLPrague23/requirements.txt

Outline

1. Intro to Transformers

Open in Colab

  • Architectures

    • Difference to other arch's (attention layer)
    • Tasks (=objectives)
  • Pre-training & Fine-tuning

  • Inputs and outputs

    • Single token prediction
  • Generation

    • Iterative prediction
    • Other generation strategies
    • [Hands-on] Constraining generated output (forcing & disabling)

2. In-context Learning and Few-shot Learning with Transformers

Open in Colab

  • Problem definition (usage)
  • Contrast with Supervised ML
  • Zero-shot vs few-shot
    • Examples
    • [Hands-on] comparison of zero-shot vs. few-shot performance (of some chosen ICL)

3. Methods for Improving ICL

Inference

  • Demonstrations heterogeneity
  • Prompt engineering
    • Promptsource - database of prompts?
    • [Hands-on] prompt engineering (inspired by the training data?)

Training Strategies

Open in Colab

  • Training strategies + existing models
    • training in explicit fewshot format (QA)
    • Instruction tuning
    • Multitask learning
    • Chain-of-Thought
    • Pre-training on a code
    • Fine-tuning with human feedback

Theory behind - why does ICL exist?

  • Data properties fostering ICL
  • Experiments
  • Explanations of the existing models?

4. Hands-on in Improving Few-shot ICL

Open in Colab

  • [Hands-on] Customizing Few-shot ICL to specialized data
  • Practical training pipeline
    • Overview of the training pipeline
    • Adaptor example

Models evaluation & competition [Optional]

If you trained your own great few-shot ICL model, it would be a pity not to test it on some unseen reasoning tasks.

See the competition readme for how to evaluate the model and if it beats the baseline, how to spread the word!


l2l_mlprague23's People

Contributors

stefanik12 avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.