LLM Course documentation
End-of-chapter quiz
0. Setup
1. Transformer models
2. Using 🤗 Transformers
IntroductionBehind the pipelineModelsTokenizersHandling multiple sequencesPutting it all togetherBasic usage completed!Optimized Inference DeploymentEnd-of-chapter quiz
3. Fine-tuning a pretrained model
4. Sharing models and tokenizers
5. The 🤗 Datasets library
6. The 🤗 Tokenizers library
7. Classical NLP tasks
8. How to ask for help
9. Building and sharing demos
10. Curate high-quality datasets
11. Fine-tune Large Language Models
12. Build Reasoning Models new
Course Events
End-of-chapter quiz
1. What is the order of the language modeling pipeline?
2. How many dimensions does the tensor output by the base Transformer model have, and what are they?
3. Which of the following is an example of subword tokenization?
4. What is a model head?
5. What is an AutoModel?
6. What are the techniques to be aware of when batching sequences of different lengths together?
7. What is the point of applying a SoftMax function to the logits output by a sequence classification model?
8. What method is most of the tokenizer API centered around?
9. What does the result variable contain in this code sample?
from transformers import AutoTokenizer
tokenizer = AutoTokenizer.from_pretrained("bert-base-cased")
result = tokenizer.tokenize("Hello!")10. Is there something wrong with the following code?
from transformers import AutoTokenizer, AutoModel
tokenizer = AutoTokenizer.from_pretrained("bert-base-cased")
model = AutoModel.from_pretrained("gpt2")
encoded = tokenizer("Hey!", return_tensors="pt")
result = model(**encoded)