├── README.md ├── added_tokens.json ├── config.json ├── generation_config.json ├── special_tokens_map.json ├── spiece.model └── tokenizer_config.json /README.md: -------------------------------------------------------------------------------- 1 | # Stock Market QA Chatbot with Text-to-Text Transfer Transformer(T5) 2 | 3 | ## 📌 Overview 4 | 5 | This repository hosts the quantized version of the T5 model fine-tuned for question-answer tasks related to stock market. The model has been trained on the stock_trading_QA dataset from Hugging Face. The model is quantized to Float16 (FP16) to optimize inference speed and efficiency while maintaining high performance. 6 | 7 | ## 🏗 Model Details 8 | 9 | - **Model Architecture:** t5-base 10 | - **Task:** QA Chatbot for Stock Market 11 | - **Dataset:** Hugging Face's `stock_trading_QA` 12 | - **Quantization:** Float16 (FP16) for optimized inference 13 | - **Fine-tuning Framework:** Hugging Face Transformers 14 | 15 | ## 🚀 Usage 16 | 17 | ### Installation 18 | 19 | ```bash 20 | pip install transformers torch 21 | ``` 22 | 23 | ### Loading the Model 24 | 25 | ```python 26 | from transformers import T5Tokenizer, T5ForConditionalGeneration 27 | import torch 28 | 29 | device = "cuda" if torch.cuda.is_available() else "cpu" 30 | 31 | model_name = "AventIQ-AI/t5-stockmarket-qa-chatbot" 32 | model = T5ForConditionalGeneration.from_pretrained(model_name).to(device) 33 | tokenizer = T5Tokenizer.from_pretrained(model_name) 34 | ``` 35 | 36 | ### Question Answer Example 37 | 38 | ```python 39 | question = "How can I start investing in stocks?" 40 | input_text = "question: " + question 41 | input_ids = tokenizer.encode(input_text, return_tensors="pt").to(model.device) 42 | 43 | with torch.no_grad(): 44 | outputs = model.generate(input_ids, max_length=50) 45 | answer = tokenizer.decode(outputs[0], skip_special_tokens=True) 46 | 47 | print(f"Question: {question}") 48 | print(f"Predicted Answer: {answer}") 49 | ``` 50 | 51 | ## 📊 Evaluation Metric: BLEU Score 52 | 53 | For question answer tasks, a high BLEU score indicates that the model’s corrected sentences closely match human-annotated corrections. 54 | 55 | ## **Interpreting Our BLEU Score** 56 | Our model achieved a **BLEU score of 0.7888**, which indicates: 57 | ✅ **Good answer generating ability** 58 | ✅ **Moderate sentence fluency** 59 | 60 | BLEU is computed by comparing the **1-gram, 2-gram, 3-gram, and 4-gram overlaps** between the model’s output and the reference sentence while applying a **brevity penalty** if the model generates shorter sentences. 61 | 62 | ### **BLEU Score Ranges for Chatbot** 63 | 64 | | BLEU Score | Interpretation | 65 | | --- | --- | 66 | | **0.8 - 1.0** | Near-perfect corrections, closely matching human annotations. | 67 | | **0.7 - 0.8** | High-quality corrections, minor variations in phrasing. | 68 | | **0.6 - 0.7** | Good corrections, but with some grammatical errors or missing words. | 69 | | **0.5 - 0.6** | Decent corrections, noticeable mistakes, lacks fluency. | 70 | | **Below 0.5** | Needs improvement, frequent incorrect corrections. | 71 | 72 | 73 | ## ⚡ Quantization Details 74 | 75 | Post-training quantization was applied using PyTorch's built-in quantization framework. The model was quantized to Float16 (FP16) to reduce model size and improve inference efficiency while balancing accuracy. 76 | 77 | ## 📂 Repository Structure 78 | 79 | ``` 80 | . 81 | ├── model/ # Contains the quantized model files 82 | ├── tokenizer_config/ # Tokenizer configuration and vocabulary files 83 | ├── model.safetensors/ # Quantized Model 84 | ├── README.md # Model documentation 85 | ``` 86 | 87 | ## ⚠️ Limitations 88 | 89 | - The model may struggle with highly ambiguous sentences. 90 | - Quantization may lead to slight degradation in accuracy compared to full-precision models. 91 | - Performance may vary across different writing styles and sentence structures. 92 | 93 | ## 🤝 Contributing 94 | 95 | Contributions are welcome! Feel free to open an issue or submit a pull request if you have suggestions or improvements. 96 | -------------------------------------------------------------------------------- /added_tokens.json: -------------------------------------------------------------------------------- 1 | { 2 | "": 32099, 3 | "": 32089, 4 | "": 32088, 5 | "": 32087, 6 | "": 32086, 7 | "": 32085, 8 | "": 32084, 9 | "": 32083, 10 | "": 32082, 11 | "": 32081, 12 | "": 32080, 13 | "": 32098, 14 | "": 32079, 15 | "": 32078, 16 | "": 32077, 17 | "": 32076, 18 | "": 32075, 19 | "": 32074, 20 | "": 32073, 21 | "": 32072, 22 | "": 32071, 23 | "": 32070, 24 | "": 32097, 25 | "": 32069, 26 | "": 32068, 27 | "": 32067, 28 | "": 32066, 29 | "": 32065, 30 | "": 32064, 31 | "": 32063, 32 | "": 32062, 33 | "": 32061, 34 | "": 32060, 35 | "": 32096, 36 | "": 32059, 37 | "": 32058, 38 | "": 32057, 39 | "": 32056, 40 | "": 32055, 41 | "": 32054, 42 | "": 32053, 43 | "": 32052, 44 | "": 32051, 45 | "": 32050, 46 | "": 32095, 47 | "": 32049, 48 | "": 32048, 49 | "": 32047, 50 | "": 32046, 51 | "": 32045, 52 | "": 32044, 53 | "": 32043, 54 | "": 32042, 55 | "": 32041, 56 | "": 32040, 57 | "": 32094, 58 | "": 32039, 59 | "": 32038, 60 | "": 32037, 61 | "": 32036, 62 | "": 32035, 63 | "": 32034, 64 | "": 32033, 65 | "": 32032, 66 | "": 32031, 67 | "": 32030, 68 | "": 32093, 69 | "": 32029, 70 | "": 32028, 71 | "": 32027, 72 | "": 32026, 73 | "": 32025, 74 | "": 32024, 75 | "": 32023, 76 | "": 32022, 77 | "": 32021, 78 | "": 32020, 79 | "": 32092, 80 | "": 32019, 81 | "": 32018, 82 | "": 32017, 83 | "": 32016, 84 | "": 32015, 85 | "": 32014, 86 | "": 32013, 87 | "": 32012, 88 | "": 32011, 89 | "": 32010, 90 | "": 32091, 91 | "": 32009, 92 | "": 32008, 93 | "": 32007, 94 | "": 32006, 95 | "": 32005, 96 | "": 32004, 97 | "": 32003, 98 | "": 32002, 99 | "": 32001, 100 | "": 32000, 101 | "": 32090 102 | } 103 | -------------------------------------------------------------------------------- /config.json: -------------------------------------------------------------------------------- 1 | { 2 | "_name_or_path": "/kaggle/working/qa-stock-model", 3 | "architectures": [ 4 | "T5ForConditionalGeneration" 5 | ], 6 | "classifier_dropout": 0.0, 7 | "d_ff": 3072, 8 | "d_kv": 64, 9 | "d_model": 768, 10 | "decoder_start_token_id": 0, 11 | "dense_act_fn": "relu", 12 | "dropout_rate": 0.1, 13 | "eos_token_id": 1, 14 | "feed_forward_proj": "relu", 15 | "initializer_factor": 1.0, 16 | "is_encoder_decoder": true, 17 | "is_gated_act": false, 18 | "layer_norm_epsilon": 1e-06, 19 | "model_type": "t5", 20 | "n_positions": 512, 21 | "num_decoder_layers": 12, 22 | "num_heads": 12, 23 | "num_layers": 12, 24 | "output_past": true, 25 | "pad_token_id": 0, 26 | "relative_attention_max_distance": 128, 27 | "relative_attention_num_buckets": 32, 28 | "task_specific_params": { 29 | "summarization": { 30 | "early_stopping": true, 31 | "length_penalty": 2.0, 32 | "max_length": 200, 33 | "min_length": 30, 34 | "no_repeat_ngram_size": 3, 35 | "num_beams": 4, 36 | "prefix": "summarize: " 37 | }, 38 | "translation_en_to_de": { 39 | "early_stopping": true, 40 | "max_length": 300, 41 | "num_beams": 4, 42 | "prefix": "translate English to German: " 43 | }, 44 | "translation_en_to_fr": { 45 | "early_stopping": true, 46 | "max_length": 300, 47 | "num_beams": 4, 48 | "prefix": "translate English to French: " 49 | }, 50 | "translation_en_to_ro": { 51 | "early_stopping": true, 52 | "max_length": 300, 53 | "num_beams": 4, 54 | "prefix": "translate English to Romanian: " 55 | } 56 | }, 57 | "torch_dtype": "float16", 58 | "transformers_version": "4.47.0", 59 | "use_cache": true, 60 | "vocab_size": 32128 61 | } 62 | -------------------------------------------------------------------------------- /generation_config.json: -------------------------------------------------------------------------------- 1 | { 2 | "_from_model_config": true, 3 | "decoder_start_token_id": 0, 4 | "eos_token_id": 1, 5 | "pad_token_id": 0, 6 | "transformers_version": "4.47.0" 7 | } 8 | -------------------------------------------------------------------------------- /special_tokens_map.json: -------------------------------------------------------------------------------- 1 | { 2 | "additional_special_tokens": [ 3 | "", 4 | "", 5 | "", 6 | "", 7 | "", 8 | "", 9 | "", 10 | "", 11 | "", 12 | "", 13 | "", 14 | "", 15 | "", 16 | "", 17 | "", 18 | "", 19 | "", 20 | "", 21 | "", 22 | "", 23 | "", 24 | "", 25 | "", 26 | "", 27 | "", 28 | "", 29 | "", 30 | "", 31 | "", 32 | "", 33 | "", 34 | "", 35 | "", 36 | "", 37 | "", 38 | "", 39 | "", 40 | "", 41 | "", 42 | "", 43 | "", 44 | "", 45 | "", 46 | "", 47 | "", 48 | "", 49 | "", 50 | "", 51 | "", 52 | "", 53 | "", 54 | "", 55 | "", 56 | "", 57 | "", 58 | "", 59 | "", 60 | "", 61 | "", 62 | "", 63 | "", 64 | "", 65 | "", 66 | "", 67 | "", 68 | "", 69 | "", 70 | "", 71 | "", 72 | "", 73 | "", 74 | "", 75 | "", 76 | "", 77 | "", 78 | "", 79 | "", 80 | "", 81 | "", 82 | "", 83 | "", 84 | "", 85 | "", 86 | "", 87 | "", 88 | "", 89 | "", 90 | "", 91 | "", 92 | "", 93 | "", 94 | "", 95 | "", 96 | "", 97 | "", 98 | "", 99 | "", 100 | "", 101 | "", 102 | "" 103 | ], 104 | "eos_token": { 105 | "content": "", 106 | "lstrip": false, 107 | "normalized": false, 108 | "rstrip": false, 109 | "single_word": false 110 | }, 111 | "pad_token": { 112 | "content": "", 113 | "lstrip": false, 114 | "normalized": false, 115 | "rstrip": false, 116 | "single_word": false 117 | }, 118 | "unk_token": { 119 | "content": "", 120 | "lstrip": false, 121 | "normalized": false, 122 | "rstrip": false, 123 | "single_word": false 124 | } 125 | } 126 | -------------------------------------------------------------------------------- /spiece.model: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/Avent-IQ/t5-stockmarket-qa-chatbot/e0f2a40ab927c530ec3fbe761c50926b883d2441/spiece.model -------------------------------------------------------------------------------- /tokenizer_config.json: -------------------------------------------------------------------------------- 1 | { 2 | "add_prefix_space": true, 3 | "added_tokens_decoder": { 4 | "0": { 5 | "content": "", 6 | "lstrip": false, 7 | "normalized": false, 8 | "rstrip": false, 9 | "single_word": false, 10 | "special": true 11 | }, 12 | "1": { 13 | "content": "", 14 | "lstrip": false, 15 | "normalized": false, 16 | "rstrip": false, 17 | "single_word": false, 18 | "special": true 19 | }, 20 | "2": { 21 | "content": "", 22 | "lstrip": false, 23 | "normalized": false, 24 | "rstrip": false, 25 | "single_word": false, 26 | "special": true 27 | }, 28 | "32000": { 29 | "content": "", 30 | "lstrip": false, 31 | "normalized": false, 32 | "rstrip": false, 33 | "single_word": false, 34 | "special": true 35 | }, 36 | "32001": { 37 | "content": "", 38 | "lstrip": false, 39 | "normalized": false, 40 | "rstrip": false, 41 | "single_word": false, 42 | "special": true 43 | }, 44 | "32002": { 45 | "content": "", 46 | "lstrip": false, 47 | "normalized": false, 48 | "rstrip": false, 49 | "single_word": false, 50 | "special": true 51 | }, 52 | "32003": { 53 | "content": "", 54 | "lstrip": false, 55 | "normalized": false, 56 | "rstrip": false, 57 | "single_word": false, 58 | "special": true 59 | }, 60 | "32004": { 61 | "content": "", 62 | "lstrip": false, 63 | "normalized": false, 64 | "rstrip": false, 65 | "single_word": false, 66 | "special": true 67 | }, 68 | "32005": { 69 | "content": "", 70 | "lstrip": false, 71 | "normalized": false, 72 | "rstrip": false, 73 | "single_word": false, 74 | "special": true 75 | }, 76 | "32006": { 77 | "content": "", 78 | "lstrip": false, 79 | "normalized": false, 80 | "rstrip": false, 81 | "single_word": false, 82 | "special": true 83 | }, 84 | "32007": { 85 | "content": "", 86 | "lstrip": false, 87 | "normalized": false, 88 | "rstrip": false, 89 | "single_word": false, 90 | "special": true 91 | }, 92 | "32008": { 93 | "content": "", 94 | "lstrip": false, 95 | "normalized": false, 96 | "rstrip": false, 97 | "single_word": false, 98 | "special": true 99 | }, 100 | "32009": { 101 | "content": "", 102 | "lstrip": false, 103 | "normalized": false, 104 | "rstrip": false, 105 | "single_word": false, 106 | "special": true 107 | }, 108 | "32010": { 109 | "content": "", 110 | "lstrip": false, 111 | "normalized": false, 112 | "rstrip": false, 113 | "single_word": false, 114 | "special": true 115 | }, 116 | "32011": { 117 | "content": "", 118 | "lstrip": false, 119 | "normalized": false, 120 | "rstrip": false, 121 | "single_word": false, 122 | "special": true 123 | }, 124 | "32012": { 125 | "content": "", 126 | "lstrip": false, 127 | "normalized": false, 128 | "rstrip": false, 129 | "single_word": false, 130 | "special": true 131 | }, 132 | "32013": { 133 | "content": "", 134 | "lstrip": false, 135 | "normalized": false, 136 | "rstrip": false, 137 | "single_word": false, 138 | "special": true 139 | }, 140 | "32014": { 141 | "content": "", 142 | "lstrip": false, 143 | "normalized": false, 144 | "rstrip": false, 145 | "single_word": false, 146 | "special": true 147 | }, 148 | "32015": { 149 | "content": "", 150 | "lstrip": false, 151 | "normalized": false, 152 | "rstrip": false, 153 | "single_word": false, 154 | "special": true 155 | }, 156 | "32016": { 157 | "content": "", 158 | "lstrip": false, 159 | "normalized": false, 160 | "rstrip": false, 161 | "single_word": false, 162 | "special": true 163 | }, 164 | "32017": { 165 | "content": "", 166 | "lstrip": false, 167 | "normalized": false, 168 | "rstrip": false, 169 | "single_word": false, 170 | "special": true 171 | }, 172 | "32018": { 173 | "content": "", 174 | "lstrip": false, 175 | "normalized": false, 176 | "rstrip": false, 177 | "single_word": false, 178 | "special": true 179 | }, 180 | "32019": { 181 | "content": "", 182 | "lstrip": false, 183 | "normalized": false, 184 | "rstrip": false, 185 | "single_word": false, 186 | "special": true 187 | }, 188 | "32020": { 189 | "content": "", 190 | "lstrip": false, 191 | "normalized": false, 192 | "rstrip": false, 193 | "single_word": false, 194 | "special": true 195 | }, 196 | "32021": { 197 | "content": "", 198 | "lstrip": false, 199 | "normalized": false, 200 | "rstrip": false, 201 | "single_word": false, 202 | "special": true 203 | }, 204 | "32022": { 205 | "content": "", 206 | "lstrip": false, 207 | "normalized": false, 208 | "rstrip": false, 209 | "single_word": false, 210 | "special": true 211 | }, 212 | "32023": { 213 | "content": "", 214 | "lstrip": false, 215 | "normalized": false, 216 | "rstrip": false, 217 | "single_word": false, 218 | "special": true 219 | }, 220 | "32024": { 221 | "content": "", 222 | "lstrip": false, 223 | "normalized": false, 224 | "rstrip": false, 225 | "single_word": false, 226 | "special": true 227 | }, 228 | "32025": { 229 | "content": "", 230 | "lstrip": false, 231 | "normalized": false, 232 | "rstrip": false, 233 | "single_word": false, 234 | "special": true 235 | }, 236 | "32026": { 237 | "content": "", 238 | "lstrip": false, 239 | "normalized": false, 240 | "rstrip": false, 241 | "single_word": false, 242 | "special": true 243 | }, 244 | "32027": { 245 | "content": "", 246 | "lstrip": false, 247 | "normalized": false, 248 | "rstrip": false, 249 | "single_word": false, 250 | "special": true 251 | }, 252 | "32028": { 253 | "content": "", 254 | "lstrip": false, 255 | "normalized": false, 256 | "rstrip": false, 257 | "single_word": false, 258 | "special": true 259 | }, 260 | "32029": { 261 | "content": "", 262 | "lstrip": false, 263 | "normalized": false, 264 | "rstrip": false, 265 | "single_word": false, 266 | "special": true 267 | }, 268 | "32030": { 269 | "content": "", 270 | "lstrip": false, 271 | "normalized": false, 272 | "rstrip": false, 273 | "single_word": false, 274 | "special": true 275 | }, 276 | "32031": { 277 | "content": "", 278 | "lstrip": false, 279 | "normalized": false, 280 | "rstrip": false, 281 | "single_word": false, 282 | "special": true 283 | }, 284 | "32032": { 285 | "content": "", 286 | "lstrip": false, 287 | "normalized": false, 288 | "rstrip": false, 289 | "single_word": false, 290 | "special": true 291 | }, 292 | "32033": { 293 | "content": "", 294 | "lstrip": false, 295 | "normalized": false, 296 | "rstrip": false, 297 | "single_word": false, 298 | "special": true 299 | }, 300 | "32034": { 301 | "content": "", 302 | "lstrip": false, 303 | "normalized": false, 304 | "rstrip": false, 305 | "single_word": false, 306 | "special": true 307 | }, 308 | "32035": { 309 | "content": "", 310 | "lstrip": false, 311 | "normalized": false, 312 | "rstrip": false, 313 | "single_word": false, 314 | "special": true 315 | }, 316 | "32036": { 317 | "content": "", 318 | "lstrip": false, 319 | "normalized": false, 320 | "rstrip": false, 321 | "single_word": false, 322 | "special": true 323 | }, 324 | "32037": { 325 | "content": "", 326 | "lstrip": false, 327 | "normalized": false, 328 | "rstrip": false, 329 | "single_word": false, 330 | "special": true 331 | }, 332 | "32038": { 333 | "content": "", 334 | "lstrip": false, 335 | "normalized": false, 336 | "rstrip": false, 337 | "single_word": false, 338 | "special": true 339 | }, 340 | "32039": { 341 | "content": "", 342 | "lstrip": false, 343 | "normalized": false, 344 | "rstrip": false, 345 | "single_word": false, 346 | "special": true 347 | }, 348 | "32040": { 349 | "content": "", 350 | "lstrip": false, 351 | "normalized": false, 352 | "rstrip": false, 353 | "single_word": false, 354 | "special": true 355 | }, 356 | "32041": { 357 | "content": "", 358 | "lstrip": false, 359 | "normalized": false, 360 | "rstrip": false, 361 | "single_word": false, 362 | "special": true 363 | }, 364 | "32042": { 365 | "content": "", 366 | "lstrip": false, 367 | "normalized": false, 368 | "rstrip": false, 369 | "single_word": false, 370 | "special": true 371 | }, 372 | "32043": { 373 | "content": "", 374 | "lstrip": false, 375 | "normalized": false, 376 | "rstrip": false, 377 | "single_word": false, 378 | "special": true 379 | }, 380 | "32044": { 381 | "content": "", 382 | "lstrip": false, 383 | "normalized": false, 384 | "rstrip": false, 385 | "single_word": false, 386 | "special": true 387 | }, 388 | "32045": { 389 | "content": "", 390 | "lstrip": false, 391 | "normalized": false, 392 | "rstrip": false, 393 | "single_word": false, 394 | "special": true 395 | }, 396 | "32046": { 397 | "content": "", 398 | "lstrip": false, 399 | "normalized": false, 400 | "rstrip": false, 401 | "single_word": false, 402 | "special": true 403 | }, 404 | "32047": { 405 | "content": "", 406 | "lstrip": false, 407 | "normalized": false, 408 | "rstrip": false, 409 | "single_word": false, 410 | "special": true 411 | }, 412 | "32048": { 413 | "content": "", 414 | "lstrip": false, 415 | "normalized": false, 416 | "rstrip": false, 417 | "single_word": false, 418 | "special": true 419 | }, 420 | "32049": { 421 | "content": "", 422 | "lstrip": false, 423 | "normalized": false, 424 | "rstrip": false, 425 | "single_word": false, 426 | "special": true 427 | }, 428 | "32050": { 429 | "content": "", 430 | "lstrip": false, 431 | "normalized": false, 432 | "rstrip": false, 433 | "single_word": false, 434 | "special": true 435 | }, 436 | "32051": { 437 | "content": "", 438 | "lstrip": false, 439 | "normalized": false, 440 | "rstrip": false, 441 | "single_word": false, 442 | "special": true 443 | }, 444 | "32052": { 445 | "content": "", 446 | "lstrip": false, 447 | "normalized": false, 448 | "rstrip": false, 449 | "single_word": false, 450 | "special": true 451 | }, 452 | "32053": { 453 | "content": "", 454 | "lstrip": false, 455 | "normalized": false, 456 | "rstrip": false, 457 | "single_word": false, 458 | "special": true 459 | }, 460 | "32054": { 461 | "content": "", 462 | "lstrip": false, 463 | "normalized": false, 464 | "rstrip": false, 465 | "single_word": false, 466 | "special": true 467 | }, 468 | "32055": { 469 | "content": "", 470 | "lstrip": false, 471 | "normalized": false, 472 | "rstrip": false, 473 | "single_word": false, 474 | "special": true 475 | }, 476 | "32056": { 477 | "content": "", 478 | "lstrip": false, 479 | "normalized": false, 480 | "rstrip": false, 481 | "single_word": false, 482 | "special": true 483 | }, 484 | "32057": { 485 | "content": "", 486 | "lstrip": false, 487 | "normalized": false, 488 | "rstrip": false, 489 | "single_word": false, 490 | "special": true 491 | }, 492 | "32058": { 493 | "content": "", 494 | "lstrip": false, 495 | "normalized": false, 496 | "rstrip": false, 497 | "single_word": false, 498 | "special": true 499 | }, 500 | "32059": { 501 | "content": "", 502 | "lstrip": false, 503 | "normalized": false, 504 | "rstrip": false, 505 | "single_word": false, 506 | "special": true 507 | }, 508 | "32060": { 509 | "content": "", 510 | "lstrip": false, 511 | "normalized": false, 512 | "rstrip": false, 513 | "single_word": false, 514 | "special": true 515 | }, 516 | "32061": { 517 | "content": "", 518 | "lstrip": false, 519 | "normalized": false, 520 | "rstrip": false, 521 | "single_word": false, 522 | "special": true 523 | }, 524 | "32062": { 525 | "content": "", 526 | "lstrip": false, 527 | "normalized": false, 528 | "rstrip": false, 529 | "single_word": false, 530 | "special": true 531 | }, 532 | "32063": { 533 | "content": "", 534 | "lstrip": false, 535 | "normalized": false, 536 | "rstrip": false, 537 | "single_word": false, 538 | "special": true 539 | }, 540 | "32064": { 541 | "content": "", 542 | "lstrip": false, 543 | "normalized": false, 544 | "rstrip": false, 545 | "single_word": false, 546 | "special": true 547 | }, 548 | "32065": { 549 | "content": "", 550 | "lstrip": false, 551 | "normalized": false, 552 | "rstrip": false, 553 | "single_word": false, 554 | "special": true 555 | }, 556 | "32066": { 557 | "content": "", 558 | "lstrip": false, 559 | "normalized": false, 560 | "rstrip": false, 561 | "single_word": false, 562 | "special": true 563 | }, 564 | "32067": { 565 | "content": "", 566 | "lstrip": false, 567 | "normalized": false, 568 | "rstrip": false, 569 | "single_word": false, 570 | "special": true 571 | }, 572 | "32068": { 573 | "content": "", 574 | "lstrip": false, 575 | "normalized": false, 576 | "rstrip": false, 577 | "single_word": false, 578 | "special": true 579 | }, 580 | "32069": { 581 | "content": "", 582 | "lstrip": false, 583 | "normalized": false, 584 | "rstrip": false, 585 | "single_word": false, 586 | "special": true 587 | }, 588 | "32070": { 589 | "content": "", 590 | "lstrip": false, 591 | "normalized": false, 592 | "rstrip": false, 593 | "single_word": false, 594 | "special": true 595 | }, 596 | "32071": { 597 | "content": "", 598 | "lstrip": false, 599 | "normalized": false, 600 | "rstrip": false, 601 | "single_word": false, 602 | "special": true 603 | }, 604 | "32072": { 605 | "content": "", 606 | "lstrip": false, 607 | "normalized": false, 608 | "rstrip": false, 609 | "single_word": false, 610 | "special": true 611 | }, 612 | "32073": { 613 | "content": "", 614 | "lstrip": false, 615 | "normalized": false, 616 | "rstrip": false, 617 | "single_word": false, 618 | "special": true 619 | }, 620 | "32074": { 621 | "content": "", 622 | "lstrip": false, 623 | "normalized": false, 624 | "rstrip": false, 625 | "single_word": false, 626 | "special": true 627 | }, 628 | "32075": { 629 | "content": "", 630 | "lstrip": false, 631 | "normalized": false, 632 | "rstrip": false, 633 | "single_word": false, 634 | "special": true 635 | }, 636 | "32076": { 637 | "content": "", 638 | "lstrip": false, 639 | "normalized": false, 640 | "rstrip": false, 641 | "single_word": false, 642 | "special": true 643 | }, 644 | "32077": { 645 | "content": "", 646 | "lstrip": false, 647 | "normalized": false, 648 | "rstrip": false, 649 | "single_word": false, 650 | "special": true 651 | }, 652 | "32078": { 653 | "content": "", 654 | "lstrip": false, 655 | "normalized": false, 656 | "rstrip": false, 657 | "single_word": false, 658 | "special": true 659 | }, 660 | "32079": { 661 | "content": "", 662 | "lstrip": false, 663 | "normalized": false, 664 | "rstrip": false, 665 | "single_word": false, 666 | "special": true 667 | }, 668 | "32080": { 669 | "content": "", 670 | "lstrip": false, 671 | "normalized": false, 672 | "rstrip": false, 673 | "single_word": false, 674 | "special": true 675 | }, 676 | "32081": { 677 | "content": "", 678 | "lstrip": false, 679 | "normalized": false, 680 | "rstrip": false, 681 | "single_word": false, 682 | "special": true 683 | }, 684 | "32082": { 685 | "content": "", 686 | "lstrip": false, 687 | "normalized": false, 688 | "rstrip": false, 689 | "single_word": false, 690 | "special": true 691 | }, 692 | "32083": { 693 | "content": "", 694 | "lstrip": false, 695 | "normalized": false, 696 | "rstrip": false, 697 | "single_word": false, 698 | "special": true 699 | }, 700 | "32084": { 701 | "content": "", 702 | "lstrip": false, 703 | "normalized": false, 704 | "rstrip": false, 705 | "single_word": false, 706 | "special": true 707 | }, 708 | "32085": { 709 | "content": "", 710 | "lstrip": false, 711 | "normalized": false, 712 | "rstrip": false, 713 | "single_word": false, 714 | "special": true 715 | }, 716 | "32086": { 717 | "content": "", 718 | "lstrip": false, 719 | "normalized": false, 720 | "rstrip": false, 721 | "single_word": false, 722 | "special": true 723 | }, 724 | "32087": { 725 | "content": "", 726 | "lstrip": false, 727 | "normalized": false, 728 | "rstrip": false, 729 | "single_word": false, 730 | "special": true 731 | }, 732 | "32088": { 733 | "content": "", 734 | "lstrip": false, 735 | "normalized": false, 736 | "rstrip": false, 737 | "single_word": false, 738 | "special": true 739 | }, 740 | "32089": { 741 | "content": "", 742 | "lstrip": false, 743 | "normalized": false, 744 | "rstrip": false, 745 | "single_word": false, 746 | "special": true 747 | }, 748 | "32090": { 749 | "content": "", 750 | "lstrip": false, 751 | "normalized": false, 752 | "rstrip": false, 753 | "single_word": false, 754 | "special": true 755 | }, 756 | "32091": { 757 | "content": "", 758 | "lstrip": false, 759 | "normalized": false, 760 | "rstrip": false, 761 | "single_word": false, 762 | "special": true 763 | }, 764 | "32092": { 765 | "content": "", 766 | "lstrip": false, 767 | "normalized": false, 768 | "rstrip": false, 769 | "single_word": false, 770 | "special": true 771 | }, 772 | "32093": { 773 | "content": "", 774 | "lstrip": false, 775 | "normalized": false, 776 | "rstrip": false, 777 | "single_word": false, 778 | "special": true 779 | }, 780 | "32094": { 781 | "content": "", 782 | "lstrip": false, 783 | "normalized": false, 784 | "rstrip": false, 785 | "single_word": false, 786 | "special": true 787 | }, 788 | "32095": { 789 | "content": "", 790 | "lstrip": false, 791 | "normalized": false, 792 | "rstrip": false, 793 | "single_word": false, 794 | "special": true 795 | }, 796 | "32096": { 797 | "content": "", 798 | "lstrip": false, 799 | "normalized": false, 800 | "rstrip": false, 801 | "single_word": false, 802 | "special": true 803 | }, 804 | "32097": { 805 | "content": "", 806 | "lstrip": false, 807 | "normalized": false, 808 | "rstrip": false, 809 | "single_word": false, 810 | "special": true 811 | }, 812 | "32098": { 813 | "content": "", 814 | "lstrip": false, 815 | "normalized": false, 816 | "rstrip": false, 817 | "single_word": false, 818 | "special": true 819 | }, 820 | "32099": { 821 | "content": "", 822 | "lstrip": false, 823 | "normalized": false, 824 | "rstrip": false, 825 | "single_word": false, 826 | "special": true 827 | } 828 | }, 829 | "additional_special_tokens": [ 830 | "", 831 | "", 832 | "", 833 | "", 834 | "", 835 | "", 836 | "", 837 | "", 838 | "", 839 | "", 840 | "", 841 | "", 842 | "", 843 | "", 844 | "", 845 | "", 846 | "", 847 | "", 848 | "", 849 | "", 850 | "", 851 | "", 852 | "", 853 | "", 854 | "", 855 | "", 856 | "", 857 | "", 858 | "", 859 | "", 860 | "", 861 | "", 862 | "", 863 | "", 864 | "", 865 | "", 866 | "", 867 | "", 868 | "", 869 | "", 870 | "", 871 | "", 872 | "", 873 | "", 874 | "", 875 | "", 876 | "", 877 | "", 878 | "", 879 | "", 880 | "", 881 | "", 882 | "", 883 | "", 884 | "", 885 | "", 886 | "", 887 | "", 888 | "", 889 | "", 890 | "", 891 | "", 892 | "", 893 | "", 894 | "", 895 | "", 896 | "", 897 | "", 898 | "", 899 | "", 900 | "", 901 | "", 902 | "", 903 | "", 904 | "", 905 | "", 906 | "", 907 | "", 908 | "", 909 | "", 910 | "", 911 | "", 912 | "", 913 | "", 914 | "", 915 | "", 916 | "", 917 | "", 918 | "", 919 | "", 920 | "", 921 | "", 922 | "", 923 | "", 924 | "", 925 | "", 926 | "", 927 | "", 928 | "", 929 | "" 930 | ], 931 | "clean_up_tokenization_spaces": false, 932 | "eos_token": "", 933 | "extra_ids": 100, 934 | "extra_special_tokens": {}, 935 | "legacy": true, 936 | "model_max_length": 1000000000000000019884624838656, 937 | "pad_token": "", 938 | "sp_model_kwargs": {}, 939 | "tokenizer_class": "T5Tokenizer", 940 | "unk_token": "" 941 | } 942 | --------------------------------------------------------------------------------