π Theory
ποΈ AST
AST stands for Abstract Syntax Tree. It is a tree representation of code.
ποΈ Beam search
When picking the next token, instead of considering the probability of the next
ποΈ Choosing parameters for Lora
To train with QLoRA, one needs to pick a lot of parameters
ποΈ Completion API
It is one of the GPT API of OpenAI. The documentation can be found here:
ποΈ Embedding models
An embedding model is a model that is designed to convert text to
ποΈ F-score measure
Classification results can be presented as a confusion matrix:
ποΈ Fine-tuning
See also
ποΈ Introduction to Deep Learning
Deep learning is the area of machine learning that deals with
ποΈ Loss function
To train a network, you need to reduce the error the network makes. This
ποΈ ML versus DL
There is three term that we use interchangeably but are subtly different.
ποΈ MPS
MPS stands for Metal Performance Shader. It's the name of the shader language
ποΈ Matrix
A matrix is a 2d table of number. It has a size called the dimension. An example
ποΈ Neural network
Neural networks are a family of algorithm that are able to learn from examples.
ποΈ PCA
PCA stands for principal component analysis. It's a way to convert vectors with
ποΈ Parameter
Consider the following function from $\mathbb$ to $\mathbb{R}$.
ποΈ Pre-prompt
It's a prompt given to the model to indicate how to behave with the user.
ποΈ QLoRA
Source
ποΈ Rank
The rank of a Matrix of size $n \times n$ is an integer
ποΈ Tokens
Computers are famously good with numbers. But they are bad when dealing with
ποΈ Train a neural network
Training a Neural network means changing its
ποΈ Transformer decoding methods
Source
ποΈ Translate texts before embeddings
When using Embedding models to vectorize words
ποΈ Vector
A vector are elements that can be added together and scaled by a constant. It is