Language models
Based on the GPT architecture, SambaStudio’s transformer language models enable computers to understand text by recognizing learned patterns. By implementing computational techniques and theories of linguistics, our language models are able to understand, generate, translate, analyze, and interpret natural language texts.
Model architecture
The table below describes the ML App and architecture for the SambaStudio provided language models. The linked document for each model provides the following information:
-
Data preparation
-
Prompt guidelines
-
Usage guidelines
-
Hyperparameter and settings
-
Inference settings
Filed of application | ML App | Architecture | Models |
---|---|---|---|
Language |
Generative Tuning 13B |
GPT 13B |
|
Language |
Generative Tuning 13B |
GPT 13B |
|
Language |
Bloom 176B |
Bloom 176B |
Data preparation
The Generative data preparation document describes how to access the generative data preparation package, used for training all SambaStudio language models, on the SambaNova public GitHub . Additionally, each language model’s document contains data preparation instructions specific to that model.