├── .gitignore ├── LICENSE ├── README-zhcn.md ├── README-zhtw.md ├── README.md ├── cpp_format_convert ├── 1-chinese-llama-7b-plus-combined.sh ├── 2-convert-merged-llama-to-ggml.sh ├── 3-convert-fp16-to-4bit.sh ├── README.md ├── archive │ ├── convert-lora-to-ggml.py │ ├── convert-lora-to-ggml.sh │ ├── convert-pth-to-ggml.py │ └── migrate-ggml-2023-03-30-pr613.py ├── convert.py ├── inference-4bit.py ├── merge_llama_with_chinese_lora.py ├── requirements.txt ├── serve-cli-4bit.sh └── serve-cli-ggml.py ├── data_sync.sh ├── docs ├── LLMs.md └── notes.md ├── efficient-finetune ├── Chinese-LLaMA-2 │ └── lora │ │ ├── README.md │ │ ├── archive │ │ ├── finetune-ddp-llama2.sh │ │ ├── finetune-ddp.sh │ │ ├── finetune-llama2.sh │ │ ├── finetune.py │ │ ├── finetune.sh │ │ ├── requirements.txt │ │ ├── requirementsv2.txt │ │ ├── serve │ │ │ ├── api.py │ │ │ ├── model.py │ │ │ ├── ui.py │ │ │ └── utils.py │ │ ├── utils.py │ │ └── validation │ │ │ ├── batch_generate.py │ │ │ ├── batch_predict.sh │ │ │ ├── generate.py │ │ │ └── run_exp.py │ │ ├── attn_and_long_ctx_patches.py │ │ ├── build_dataset.py │ │ ├── ds_zero2_no_offload.json │ │ ├── env_installation.sh │ │ ├── gradio_demo-dcardwomzhcn.sh │ │ ├── gradio_demo.py │ │ ├── inference │ │ ├── flash_attn_patch_for_inference.py │ │ ├── gradio_demo.sh │ │ └── speculative_sample.py │ │ ├── inference_hf.py │ │ ├── peftv2 │ │ ├── __init__.py │ │ ├── mapping.py │ │ ├── peft_model.py │ │ ├── tuners │ │ │ ├── __init__.py │ │ │ ├── lora.py │ │ │ ├── p_tuning.py │ │ │ ├── prefix_tuning.py │ │ │ └── prompt_tuning.py │ │ └── utils │ │ │ ├── __init__.py │ │ │ ├── adapters_utils.py │ │ │ ├── config.py │ │ │ ├── other.py │ │ │ └── save_and_load.py │ │ ├── run_clm_pt_with_peft.py │ │ ├── run_clm_sft_with_peft.py │ │ ├── run_pt.sh │ │ ├── run_sft-dcardwomzhcn.sh │ │ ├── run_sft.sh │ │ ├── scripts_record │ │ ├── inference_hf-dcardwomzhcn.sh │ │ ├── inference_hf-dcardwomzhtw.sh │ │ ├── run_sft-dcardwomzhtw.sh │ │ └── run_sft-medicalqazhcn.sh │ │ └── support_models.yaml ├── README.md ├── config │ ├── normal.yaml │ ├── use_deepspeed.yaml │ └── zero_stage3_offload_config.json └── ptuning │ ├── README.md │ └── v2 │ ├── arguments.py │ ├── evaluate.sh │ ├── finetune-chatglm2.sh │ ├── finetune-ddp.sh │ ├── finetune.py │ ├── finetune.sh │ ├── requirements.txt │ ├── run_shell.py │ ├── serve │ ├── cli-cpu.sh │ ├── cli-multi_gpu.sh │ ├── cli-single_gpu.sh │ └── cli.py │ ├── support_models.yaml │ ├── trainer.py │ └── trainer_seq2seq.py ├── get_model.py ├── instruction-datasets └── README.md ├── use_cases └── 中文衛教諮詢問答.md └── utils ├── data_utils.py └── model_utils.py /.gitignore: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/.gitignore -------------------------------------------------------------------------------- /LICENSE: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/LICENSE -------------------------------------------------------------------------------- /README-zhcn.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/README-zhcn.md -------------------------------------------------------------------------------- /README-zhtw.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/README-zhtw.md -------------------------------------------------------------------------------- /README.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/README.md -------------------------------------------------------------------------------- /cpp_format_convert/1-chinese-llama-7b-plus-combined.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/cpp_format_convert/1-chinese-llama-7b-plus-combined.sh -------------------------------------------------------------------------------- /cpp_format_convert/2-convert-merged-llama-to-ggml.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/cpp_format_convert/2-convert-merged-llama-to-ggml.sh -------------------------------------------------------------------------------- /cpp_format_convert/3-convert-fp16-to-4bit.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/cpp_format_convert/3-convert-fp16-to-4bit.sh -------------------------------------------------------------------------------- /cpp_format_convert/README.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/cpp_format_convert/README.md -------------------------------------------------------------------------------- /cpp_format_convert/archive/convert-lora-to-ggml.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/cpp_format_convert/archive/convert-lora-to-ggml.py -------------------------------------------------------------------------------- /cpp_format_convert/archive/convert-lora-to-ggml.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/cpp_format_convert/archive/convert-lora-to-ggml.sh -------------------------------------------------------------------------------- /cpp_format_convert/archive/convert-pth-to-ggml.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/cpp_format_convert/archive/convert-pth-to-ggml.py -------------------------------------------------------------------------------- /cpp_format_convert/archive/migrate-ggml-2023-03-30-pr613.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/cpp_format_convert/archive/migrate-ggml-2023-03-30-pr613.py -------------------------------------------------------------------------------- /cpp_format_convert/convert.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/cpp_format_convert/convert.py -------------------------------------------------------------------------------- /cpp_format_convert/inference-4bit.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/cpp_format_convert/inference-4bit.py -------------------------------------------------------------------------------- /cpp_format_convert/merge_llama_with_chinese_lora.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/cpp_format_convert/merge_llama_with_chinese_lora.py -------------------------------------------------------------------------------- /cpp_format_convert/requirements.txt: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/cpp_format_convert/requirements.txt -------------------------------------------------------------------------------- /cpp_format_convert/serve-cli-4bit.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/cpp_format_convert/serve-cli-4bit.sh -------------------------------------------------------------------------------- /cpp_format_convert/serve-cli-ggml.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/cpp_format_convert/serve-cli-ggml.py -------------------------------------------------------------------------------- /data_sync.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/data_sync.sh -------------------------------------------------------------------------------- /docs/LLMs.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/docs/LLMs.md -------------------------------------------------------------------------------- /docs/notes.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/docs/notes.md -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/README.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/README.md -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/archive/finetune-ddp-llama2.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/archive/finetune-ddp-llama2.sh -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/archive/finetune-ddp.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/archive/finetune-ddp.sh -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/archive/finetune-llama2.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/archive/finetune-llama2.sh -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/archive/finetune.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/archive/finetune.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/archive/finetune.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/archive/finetune.sh -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/archive/requirements.txt: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/archive/requirements.txt -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/archive/requirementsv2.txt: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/archive/requirementsv2.txt -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/archive/serve/api.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/archive/serve/api.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/archive/serve/model.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/archive/serve/model.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/archive/serve/ui.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/archive/serve/ui.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/archive/serve/utils.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/archive/serve/utils.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/archive/utils.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/archive/utils.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/archive/validation/batch_generate.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/archive/validation/batch_generate.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/archive/validation/batch_predict.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/archive/validation/batch_predict.sh -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/archive/validation/generate.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/archive/validation/generate.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/archive/validation/run_exp.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/archive/validation/run_exp.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/attn_and_long_ctx_patches.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/attn_and_long_ctx_patches.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/build_dataset.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/build_dataset.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/ds_zero2_no_offload.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/ds_zero2_no_offload.json -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/env_installation.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/env_installation.sh -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/gradio_demo-dcardwomzhcn.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/gradio_demo-dcardwomzhcn.sh -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/gradio_demo.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/gradio_demo.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/inference/flash_attn_patch_for_inference.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/inference/flash_attn_patch_for_inference.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/inference/gradio_demo.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/inference/gradio_demo.sh -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/inference/speculative_sample.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/inference/speculative_sample.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/inference_hf.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/inference_hf.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/peftv2/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/peftv2/__init__.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/peftv2/mapping.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/peftv2/mapping.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/peftv2/peft_model.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/peftv2/peft_model.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/peftv2/tuners/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/peftv2/tuners/__init__.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/peftv2/tuners/lora.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/peftv2/tuners/lora.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/peftv2/tuners/p_tuning.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/peftv2/tuners/p_tuning.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/peftv2/tuners/prefix_tuning.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/peftv2/tuners/prefix_tuning.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/peftv2/tuners/prompt_tuning.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/peftv2/tuners/prompt_tuning.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/peftv2/utils/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/peftv2/utils/__init__.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/peftv2/utils/adapters_utils.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/peftv2/utils/adapters_utils.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/peftv2/utils/config.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/peftv2/utils/config.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/peftv2/utils/other.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/peftv2/utils/other.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/peftv2/utils/save_and_load.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/peftv2/utils/save_and_load.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/run_clm_pt_with_peft.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/run_clm_pt_with_peft.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/run_clm_sft_with_peft.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/run_clm_sft_with_peft.py -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/run_pt.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/run_pt.sh -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/run_sft-dcardwomzhcn.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/run_sft-dcardwomzhcn.sh -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/run_sft.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/run_sft.sh -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/scripts_record/inference_hf-dcardwomzhcn.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/scripts_record/inference_hf-dcardwomzhcn.sh -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/scripts_record/inference_hf-dcardwomzhtw.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/scripts_record/inference_hf-dcardwomzhtw.sh -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/scripts_record/run_sft-dcardwomzhtw.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/scripts_record/run_sft-dcardwomzhtw.sh -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/scripts_record/run_sft-medicalqazhcn.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/scripts_record/run_sft-medicalqazhcn.sh -------------------------------------------------------------------------------- /efficient-finetune/Chinese-LLaMA-2/lora/support_models.yaml: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/Chinese-LLaMA-2/lora/support_models.yaml -------------------------------------------------------------------------------- /efficient-finetune/README.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/README.md -------------------------------------------------------------------------------- /efficient-finetune/config/normal.yaml: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/config/normal.yaml -------------------------------------------------------------------------------- /efficient-finetune/config/use_deepspeed.yaml: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/config/use_deepspeed.yaml -------------------------------------------------------------------------------- /efficient-finetune/config/zero_stage3_offload_config.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/config/zero_stage3_offload_config.json -------------------------------------------------------------------------------- /efficient-finetune/ptuning/README.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/ptuning/README.md -------------------------------------------------------------------------------- /efficient-finetune/ptuning/v2/arguments.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/ptuning/v2/arguments.py -------------------------------------------------------------------------------- /efficient-finetune/ptuning/v2/evaluate.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/ptuning/v2/evaluate.sh -------------------------------------------------------------------------------- /efficient-finetune/ptuning/v2/finetune-chatglm2.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/ptuning/v2/finetune-chatglm2.sh -------------------------------------------------------------------------------- /efficient-finetune/ptuning/v2/finetune-ddp.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/ptuning/v2/finetune-ddp.sh -------------------------------------------------------------------------------- /efficient-finetune/ptuning/v2/finetune.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/ptuning/v2/finetune.py -------------------------------------------------------------------------------- /efficient-finetune/ptuning/v2/finetune.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/ptuning/v2/finetune.sh -------------------------------------------------------------------------------- /efficient-finetune/ptuning/v2/requirements.txt: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/ptuning/v2/requirements.txt -------------------------------------------------------------------------------- /efficient-finetune/ptuning/v2/run_shell.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/ptuning/v2/run_shell.py -------------------------------------------------------------------------------- /efficient-finetune/ptuning/v2/serve/cli-cpu.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/ptuning/v2/serve/cli-cpu.sh -------------------------------------------------------------------------------- /efficient-finetune/ptuning/v2/serve/cli-multi_gpu.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/ptuning/v2/serve/cli-multi_gpu.sh -------------------------------------------------------------------------------- /efficient-finetune/ptuning/v2/serve/cli-single_gpu.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/ptuning/v2/serve/cli-single_gpu.sh -------------------------------------------------------------------------------- /efficient-finetune/ptuning/v2/serve/cli.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/ptuning/v2/serve/cli.py -------------------------------------------------------------------------------- /efficient-finetune/ptuning/v2/support_models.yaml: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/ptuning/v2/support_models.yaml -------------------------------------------------------------------------------- /efficient-finetune/ptuning/v2/trainer.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/ptuning/v2/trainer.py -------------------------------------------------------------------------------- /efficient-finetune/ptuning/v2/trainer_seq2seq.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/efficient-finetune/ptuning/v2/trainer_seq2seq.py -------------------------------------------------------------------------------- /get_model.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/get_model.py -------------------------------------------------------------------------------- /instruction-datasets/README.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/instruction-datasets/README.md -------------------------------------------------------------------------------- /use_cases/中文衛教諮詢問答.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/use_cases/中文衛教諮詢問答.md -------------------------------------------------------------------------------- /utils/data_utils.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/utils/data_utils.py -------------------------------------------------------------------------------- /utils/model_utils.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/A-baoYang/LLM-Finetune-Guide/HEAD/utils/model_utils.py --------------------------------------------------------------------------------