arrow-left

All pages
gitbookPowered by GitBook
1 of 10

Loading...

Loading...

Loading...

Loading...

Loading...

Loading...

Loading...

Loading...

Loading...

Loading...

Week 1 - General Introduction

hashtag
This week you will...

  • get all details about the structure and form of the course.

  • get to know your fellow course participants.

  • get a general introduction into the current use of transformer models and their basic idea.

hashtag
Learning Resources

hashtag
Until next week you should...

file-pdf
3MB
221026_General Introduction.pdf
PDF
arrow-up-right-from-squareOpen
Presentation Slides
video 1arrow-up-right
video 2arrow-up-right
chapter 1arrow-up-right

Requirements for a Certificate of Achievement or ECTS

The conditions to be met in order to receive a Certificate of Achievement (and ECTS) are:

hashtag
Attendance:

If you attend via Zoom, please make sure to use your full name, which should be the same that you used to register at edu.opencampus.sh. Otherwise your attendance will not be recorded!

hashtag
Projects:

Check the Projects section to learn more about the projects.

Application of Transformer Models

Week 2 - Self-Attention and Prompt Design

hashtag
This week you will...

  • get to know the mechanism underlying the self-attention approach.

  • get to know the basics of prompt design and how to apply them in a playground.

  • get to know examples for the application of transformer models.

hashtag
Learning Resources

Additional resource explaining the Transformer model:

  • Text: by Jay Alammar

  • Video:

hashtag
Until next week you should...

Week 5 - The Datasets Library

hashtag
This week you will...

  • insights on the benefits of the Hugging Face Datasets library

  • discuss the literature review on project tasks

  • get some ideas on how to visualize sequence data

hashtag
Learning Resources

hashtag
Until next class you should...

Week 3 - Introduction to Transformer Models

hashtag
This week you will...

  • learn about the basic structure of the transformer models

  • get to know what tokenization is about and why it is important.

hashtag
Learning Resources

  • To get an idea about the number of tokens a text or word is made off in the GPT models, you may want to checkout .

hashtag
Until next week you should...

Week 4 - Fine-Tuning Pretrained Models

hashtag
This week you will...

  • learn how to fine-tune a Hugging Face model with your own data

hashtag
Learning Resources

hashtag
Until next week you should...

file-pdf
5MB
221102_Self-Attention and Prompt Design.pdf
PDF
arrow-up-right-from-squareOpen
The Illustrated Transformerarrow-up-right
Transformer Neural Networks - EXPLAINED!arrow-up-right
video 3arrow-up-right
video 4arrow-up-right
chapter 2arrow-up-right

create filter variables to group your input data according to special characteristics

  • consider the following questions:

    • What are potential biases in your training data?

    • Are there outliers in the dataset?

    • Are the classes balanced? (If you deal with a classification task.)

  • file-pdf
    3MB
    221123_The Datasets Library.pdf
    PDF
    arrow-up-right-from-squareOpen
    chapter arrow-up-right
    6arrow-up-right
    file-pdf
    2MB
    221109_Introduction to Transformers.pdf
    PDF
    arrow-up-right-from-squareOpen
    this sitearrow-up-right
    chapter 3arrow-up-right

    Focus on the structure of the input and of the output

  • Are there pretrained models that you can use?

  • Which type of model is best suited? Do you need tokenization?

  • Do you need a type of embedding layer?

  • file-pdf
    2MB
    221116_Fine-Tuning Transformers.pdf
    PDF
    arrow-up-right-from-squareOpen
    chapter 4arrow-up-right
    chapter 5arrow-up-right

    Week 7 - Main NLP Tasks

    hashtag
    This week you will...

    • get insights on the preprocessing of different NLP and sequence classification tasks.

    • get an idea on plausible hyperparameters to fit transformer models for different tasks.

    • learn about different metrics to evaluate NLP models.

    hashtag
    Learning Resources

    hashtag
    Until next class you should...

    • prepare your final presentation as described in week 8 .

    Week 8 - Presentation of the Final Projects

    hashtag
    This week you will...

    • present your project in the final presentation. :-)

    The presentation should take about 15 minutes and include the following aspects that correspond to the content of a model card for the Hugging Face library:

    • model description

    • intended uses & limitations

    • training data

    Check this section on for more details.

    Additionally, please also include into your presentation any things that you tried out and that didn't work, so we can all learn from your experiences.

    Week 6 - The Tokenizers Library

    hashtag
    This week you will...

    • get insights on the benefits of the Hugging Face Tokenizers library

    • discuss the characteristics of the data you use in your project

    hashtag
    Learning Resources

    hashtag
    Until next class you should...

    file-pdf
    4MB
    230104_Main NLP Tasks.pdf
    PDF
    arrow-up-right-from-squareOpen
    here
    training procedure
  • variable and Metrics

  • evaluation results

  • Building a model cardarrow-up-right
    file-pdf
    2MB
    221130_The Tokenizers Library.pdf
    PDF
    arrow-up-right-from-squareOpen
    chapter 7arrow-up-right