Intro

img

A large language model (LLM) is a language model trained with self-supervised machine learning on a vast amount of text, designed for natural language processing tasks, especially language generation

Intro to Large Language Models | slides

LLM Training

img

we can use these pre-trained weights as initial weights for a task-specific fine-tuning training process with a smaller amount of labelled data

Pre-training

Pre-training is the initial phase of training an LLM on a large corpus of text dataset, enabling the model to learn general language patterns, often by predicting the next token.

Fine-tuning

Fine-tuning involves training the base model on a more specific dataset to adapt it to tasks not covered during pre-training

Inference

Inference is the process of running a trained model to to generate outputs

img

Whether it’s answering a question, generating content, or analyzing new data, inference is the phase where the AI “thinks” and produces results

Related