Generative data preparation
The generative data preparation package describes how to prepare data to be used to train SambaNova’s generative NLP text models. It offers an efficient way to convert input text files into tokenized sequences that are packed into a fixed sequence length. The resulting output directory can be directly used for training. This package features many styles of packing text of any length into tokenized sequences, compressed hdf5 file outputs, efficient multiprocessing, shuffling any sized dataset, splitting your data into train/dev/test, and specifying what tokens are attended to during training.
To access the generative data preparation package, including its associated documentation, please visit the SambaNova public GitHub using the following link: