1 | 7 | ||||||||||||||||||||||
2 | |||||||||||||||||||||||
3 | 8 | ||||||||||||||||||||||
4 | |||||||||||||||||||||||
5 | |||||||||||||||||||||||
6 | |||||||||||||||||||||||
1. The process of interpreting the meaning of words and phrases in context.
2. A large set of texts used to train a model to understand and generate language.
3. A modeling error that occurs when a model learns the training data too well, failing to generalize to new data.
4. A method of further training a pre-trained model on a specific dataset to improve performance on a particular task.
5. The internal variables of a model that are adjusted during training to minimize prediction error.
6. Where a text is represented as an unordered collection of words.
7. An architecture that uses self-attention mechanisms to process and generate sequences of data.
8. The process of breaking text into smaller pieces, called tokens, which can be words or subwords.