S

Empirical relationships showing how AI model performance improves predictably as model size, dataset size, or compute budget increases. Scaling laws have guided the development of increasingly large f...

Definition
S

Artificially generated data that mimics the statistical properties of real-world data. Synthetic data is used to train AI models when real data is scarce, expensive, sensitive, or biased, and can be g...

Definition
T

A parameter that controls the randomness of an AI model's output. Lower temperature values produce more deterministic and focused responses, while higher values increase creativity and diversity but m...

Definition
T

An AI technology that converts written text into natural-sounding spoken audio. Modern TTS systems use deep learning to produce highly realistic voice output with appropriate intonation, rhythm, and e...

Definition
T

The basic unit of text that language models process. A token can be a word, part of a word (subword), or a single character, depending on the tokenizer. Token counts determine model input limits, outp...

Definition
T

The process of breaking text into smaller units called tokens, which can be words, subwords, or characters. Tokenization is the first step in natural language processing pipelines and determines how t...

Definition
T

A deep learning architecture introduced in the 2017 paper "Attention Is All You Need" that relies on self-attention mechanisms instead of recurrence. Transformers are the foundation of modern LLMs lik...

Definition
U

A modeling error that occurs when a machine learning model is too simple to capture the underlying patterns in the data. An underfit model performs poorly on both training and test data, indicating in...

Definition
W

A learned representation of text where words with similar meanings are mapped to nearby points in a vector space. Pioneered by Word2Vec and GloVe, word embeddings capture semantic and syntactic relati...

Definition