├── .gitignore ├── README.md ├── datatools ├── __init__.py ├── io_utils.py ├── load.py ├── merge_index.py ├── process.py └── scripts │ ├── merge_index.py │ ├── pack.py │ ├── peek.py │ ├── tokenize.py │ ├── tokenizers │ ├── llama2_tokenizer.model │ ├── llama2_tokenizer.py │ ├── llama3_tokenizer.model │ └── llama3_tokenizer.py │ └── wrangle.py └── setup.py /.gitignore: -------------------------------------------------------------------------------- 1 | # Byte-compiled / optimized / DLL files 2 | __pycache__/ 3 | *.py[cod] 4 | *$py.class 5 | 6 | # C extensions 7 | *.so 8 | 9 | # Distribution / packaging 10 | .Python 11 | build/ 12 | develop-eggs/ 13 | dist/ 14 | downloads/ 15 | eggs/ 16 | .eggs/ 17 | lib/ 18 | lib64/ 19 | parts/ 20 | sdist/ 21 | var/ 22 | wheels/ 23 | share/python-wheels/ 24 | *.egg-info/ 25 | .installed.cfg 26 | *.egg 27 | MANIFEST 28 | 29 | # PyInstaller 30 | # Usually these files are written by a python script from a template 31 | # before PyInstaller builds the exe, so as to inject date/other infos into it. 32 | *.manifest 33 | *.spec 34 | 35 | # Installer logs 36 | pip-log.txt 37 | pip-delete-this-directory.txt 38 | 39 | # Unit test / coverage reports 40 | htmlcov/ 41 | .tox/ 42 | .nox/ 43 | .coverage 44 | .coverage.* 45 | .cache 46 | nosetests.xml 47 | coverage.xml 48 | *.cover 49 | *.py,cover 50 | .hypothesis/ 51 | .pytest_cache/ 52 | cover/ 53 | 54 | # Translations 55 | *.mo 56 | *.pot 57 | 58 | # Django stuff: 59 | *.log 60 | local_settings.py 61 | db.sqlite3 62 | db.sqlite3-journal 63 | 64 | # Flask stuff: 65 | instance/ 66 | .webassets-cache 67 | 68 | # Scrapy stuff: 69 | .scrapy 70 | 71 | # Sphinx documentation 72 | docs/_build/ 73 | 74 | # PyBuilder 75 | .pybuilder/ 76 | target/ 77 | 78 | # Jupyter Notebook 79 | .ipynb_checkpoints 80 | 81 | # IPython 82 | profile_default/ 83 | ipython_config.py 84 | 85 | # pyenv 86 | # For a library or package, you might want to ignore these files since the code is 87 | # intended to run in multiple environments; otherwise, check them in: 88 | # .python-version 89 | 90 | # pipenv 91 | # According to pypa/pipenv#598, it is recommended to include Pipfile.lock in version control. 92 | # However, in case of collaboration, if having platform-specific dependencies or dependencies 93 | # having no cross-platform support, pipenv may install dependencies that don't work, or not 94 | # install all needed dependencies. 95 | #Pipfile.lock 96 | 97 | # poetry 98 | # Similar to Pipfile.lock, it is generally recommended to include poetry.lock in version control. 99 | # This is especially recommended for binary packages to ensure reproducibility, and is more 100 | # commonly ignored for libraries. 101 | # https://python-poetry.org/docs/basic-usage/#commit-your-poetrylock-file-to-version-control 102 | #poetry.lock 103 | 104 | # pdm 105 | # Similar to Pipfile.lock, it is generally recommended to include pdm.lock in version control. 106 | #pdm.lock 107 | # pdm stores project-wide configurations in .pdm.toml, but it is recommended to not include it 108 | # in version control. 109 | # https://pdm.fming.dev/#use-with-ide 110 | .pdm.toml 111 | 112 | # PEP 582; used by e.g. github.com/David-OConnor/pyflow and github.com/pdm-project/pdm 113 | __pypackages__/ 114 | 115 | # Celery stuff 116 | celerybeat-schedule 117 | celerybeat.pid 118 | 119 | # SageMath parsed files 120 | *.sage.py 121 | 122 | # Environments 123 | .env 124 | .venv 125 | env/ 126 | venv/ 127 | ENV/ 128 | env.bak/ 129 | venv.bak/ 130 | 131 | # Spyder project settings 132 | .spyderproject 133 | .spyproject 134 | 135 | # Rope project settings 136 | .ropeproject 137 | 138 | # mkdocs documentation 139 | /site 140 | 141 | # mypy 142 | .mypy_cache/ 143 | .dmypy.json 144 | dmypy.json 145 | 146 | # Pyre type checker 147 | .pyre/ 148 | 149 | # pytype static type analyzer 150 | .pytype/ 151 | 152 | # Cython debug symbols 153 | cython_debug/ 154 | 155 | # PyCharm 156 | # JetBrains specific template is maintained in a separate JetBrains.gitignore that can 157 | # be found at https://github.com/github/gitignore/blob/main/Global/JetBrains.gitignore 158 | # and can be added to the global gitignore or merged into this file. For a more nuclear 159 | # option (not recommended) you can uncomment the following to ignore the entire idea folder. 160 | #.idea/ 161 | 162 | 163 | checkpoints 164 | datasets 165 | .cache 166 | slurm 167 | init 168 | 169 | 170 | wandb 171 | 172 | 173 | # Utils for creating math datasets 174 | math_utils/ -------------------------------------------------------------------------------- /README.md: -------------------------------------------------------------------------------- 1 | # 🛠️ *datatools*: Simple utilities for common data actions 2 | 3 | Minimal scripts and reusable functions for implementing common data operations (tokenization, splitting, subsampling, packing, and more). 4 | 5 | Built with special support for [Mosaic Streaming Datasets (MDS)](https://docs.mosaicml.com/projects/streaming/en/stable/index.html). 6 | 7 | ## Table of contents 8 | - [Installation](#installation) 9 | - [Library](#library) 10 | - [Core Functions](#core-functions) 11 | - [Example](#example) 12 | - [Scripts](#scripts) 13 | 14 | ## Installation 15 | 16 | Clone this repo and install via `pip install -e .` or install from PyPI via `pip install datatools-py`. 17 | 18 | ## Library 19 | 20 | *datatools* provides core libraries that can be used to easily build custom data pipelines, specifically through `from datatools import load, process`. 21 | 22 | ### Core functions 23 | 24 | ```python 25 | load(path, load_options) 26 | ``` 27 | Loads the dataset at the path and **automatically infers its format** (e.g., compressed JSON, PyArrow, MDS, etc.) based on clues from the file format and directory structure. 28 | 29 | --- 30 | 31 | ```python 32 | process(input_dataset, process_fn, output_path, process_options) 33 | ``` 34 | Processes an input dataset and writes the results to disk. It supports: 35 | 36 | 1. **Multi-processing** with many CPUs, e.g. `ProcessOptions(num_proc=16)` (or as flag `-w 16`) 37 | 2. **Slurm array parallelization**, e.g. `ProcessOptions(slurm_array=True)` (or `--slurm_array`) automatically sets up `job_id` and `num_jobs` using Slurm environment variables 38 | 3. **Custom indexing**, e.g. only working on a subset `--index_range 0 30` or using a custom index file `--index_path path/to/index.npy` 39 | See [ProcessOptions](https://github.com/CodeCreator/datatools/blob/main/datatools/process.py#L30) for details. 40 | 4. By default, output is written as mosaic-streaming MDS shards, which are merged into a single MDS dataset when the job finishes. The code also supports writing to JSONL files (`--jsonl`) and ndarray files for each column (`--ndarray`). The shards for these output formats are not automatically merged. 41 | 42 | The `process_fn` should be a function that takes one to three arguments: 43 | 1. A subset of the data with `len(...)` and `.[...]` access 44 | 2. The global indices corresponding to the subset (optional) 45 | 3. The `process_id` for logging or sharding purposes (optional) 46 | 47 | ### Example 48 | 49 | ```python 50 | from datatools import load, process, ProcessOptions 51 | from transformers import AutoTokenizer 52 | 53 | # Load dataset (can be JSON, Parquet, MDS, etc.) 54 | dataset = load("path/to/dataset") 55 | 56 | # Setup tokenizer and processing function 57 | tokenizer = AutoTokenizer.from_pretrained("meta-llama/Llama-3.1-8B") 58 | def tokenize_docs(data_subset): 59 | for item in data_subset: 60 | # Tokenize text and return dict with tokens and length 61 | tokens = tokenizer.encode(item["text"], add_special_tokens=False) 62 | 63 | # Chunk the text into 1024 token chunks 64 | for i in range(0, len(tokens), 1024): 65 | yield { 66 | "input_ids": tokens[i:i+1024], 67 | "length": len(tokens[i:i+1024]) 68 | } 69 | 70 | # Process dataset with 4 workers and write to disk 71 | process(dataset, tokenize_docs, "path/to/output", process_options=ProcessOptions(num_proc=4)) 72 | ``` 73 | 74 | ## Scripts 75 | 76 | *datatools* comes with the following default scripts: 77 | 78 | * `tokenize`: Tokenize datasets per document 79 | * `pack`: Pack tokenized documents into fixed sequences 80 | * `peek`: Print datasets as JSON to stdout 81 | * `wrangle`: Subsample, merge datasets, make random splits (e.g., train/test/validation), etc. 82 | * `merge_index`: Merge Mosaic streaming datasets in subfolders into a larger dataset 83 | 84 | Run `