├── README.md ├── figs ├── method.png └── teaser.png ├── model_release ├── dataset.py ├── fsdp_train.py └── llava │ ├── README.md │ ├── docs │ ├── Customize_Component.md │ ├── Data.md │ ├── Evaluation.md │ ├── LLaVA_Bench.md │ ├── LLaVA_from_LLaMA2.md │ ├── LoRA.md │ ├── MODEL_ZOO.md │ ├── ScienceQA.md │ └── Windows.md │ ├── images │ ├── demo_cli.gif │ ├── llava_example_cmp.png │ ├── llava_logo.png │ └── llava_v1_5_radar.jpg │ ├── llava.egg-info │ ├── PKG-INFO │ ├── SOURCES.txt │ ├── dependency_links.txt │ ├── requires.txt │ └── top_level.txt │ ├── llava │ ├── __init__.py │ ├── __pycache__ │ │ ├── __init__.cpython-310.pyc │ │ ├── __init__.cpython-39.pyc │ │ ├── constants.cpython-310.pyc │ │ ├── constants.cpython-39.pyc │ │ ├── conversation.cpython-310.pyc │ │ ├── mm_utils.cpython-310.pyc │ │ ├── mm_utils.cpython-39.pyc │ │ └── utils.cpython-310.pyc │ ├── constants.py │ ├── conversation.py │ ├── eval │ │ ├── eval_gpt_review.py │ │ ├── eval_gpt_review_bench.py │ │ ├── eval_gpt_review_visual.py │ │ ├── eval_pope.py │ │ ├── eval_science_qa.py │ │ ├── eval_science_qa_gpt4.py │ │ ├── eval_science_qa_gpt4_requery.py │ │ ├── eval_textvqa.py │ │ ├── generate_webpage_data_from_table.py │ │ ├── m4c_evaluator.py │ │ ├── model_qa.py │ │ ├── model_vqa.py │ │ ├── model_vqa_loader.py │ │ ├── model_vqa_mmbench.py │ │ ├── model_vqa_science.py │ │ ├── qa_baseline_gpt35.py │ │ ├── run_llava.py │ │ ├── summarize_gpt_review.py │ │ ├── table │ │ │ ├── answer │ │ │ │ ├── answer_alpaca-13b.jsonl │ │ │ │ ├── answer_bard.jsonl │ │ │ │ ├── answer_gpt35.jsonl │ │ │ │ ├── answer_llama-13b.jsonl │ │ │ │ └── answer_vicuna-13b.jsonl │ │ │ ├── caps_boxes_coco2014_val_80.jsonl │ │ │ ├── model.jsonl │ │ │ ├── prompt.jsonl │ │ │ ├── question.jsonl │ │ │ ├── review │ │ │ │ ├── review_alpaca-13b_vicuna-13b.jsonl │ │ │ │ ├── review_bard_vicuna-13b.jsonl │ │ │ │ ├── review_gpt35_vicuna-13b.jsonl │ │ │ │ └── review_llama-13b_vicuna-13b.jsonl │ │ │ ├── reviewer.jsonl │ │ │ └── rule.json │ │ └── webpage │ │ │ ├── figures │ │ │ ├── alpaca.png │ │ │ ├── bard.jpg │ │ │ ├── chatgpt.svg │ │ │ ├── llama.jpg │ │ │ ├── swords_FILL0_wght300_GRAD0_opsz48.svg │ │ │ └── vicuna.jpeg │ │ │ ├── index.html │ │ │ ├── script.js │ │ │ └── styles.css │ ├── mm_utils.py │ ├── model │ │ ├── __init__.py │ │ ├── __pycache__ │ │ │ ├── __init__.cpython-310.pyc │ │ │ ├── __init__.cpython-39.pyc │ │ │ ├── builder.cpython-310.pyc │ │ │ ├── builder.cpython-39.pyc │ │ │ ├── llava_arch.cpython-310.pyc │ │ │ └── llava_arch.cpython-39.pyc │ │ ├── apply_delta.py │ │ ├── builder.py │ │ ├── consolidate.py │ │ ├── language_model │ │ │ ├── __pycache__ │ │ │ │ ├── llava_llama.cpython-310.pyc │ │ │ │ ├── llava_llama.cpython-39.pyc │ │ │ │ ├── llava_mpt.cpython-310.pyc │ │ │ │ └── llava_mpt.cpython-39.pyc │ │ │ ├── llava_llama.py │ │ │ ├── llava_mpt.py │ │ │ └── mpt │ │ │ │ ├── __pycache__ │ │ │ │ ├── adapt_tokenizer.cpython-310.pyc │ │ │ │ ├── adapt_tokenizer.cpython-39.pyc │ │ │ │ ├── attention.cpython-310.pyc │ │ │ │ ├── attention.cpython-39.pyc │ │ │ │ ├── blocks.cpython-310.pyc │ │ │ │ ├── blocks.cpython-39.pyc │ │ │ │ ├── configuration_mpt.cpython-310.pyc │ │ │ │ ├── configuration_mpt.cpython-39.pyc │ │ │ │ ├── custom_embedding.cpython-310.pyc │ │ │ │ ├── custom_embedding.cpython-39.pyc │ │ │ │ ├── flash_attn_triton.cpython-310.pyc │ │ │ │ ├── flash_attn_triton.cpython-39.pyc │ │ │ │ ├── hf_prefixlm_converter.cpython-310.pyc │ │ │ │ ├── hf_prefixlm_converter.cpython-39.pyc │ │ │ │ ├── meta_init_context.cpython-310.pyc │ │ │ │ ├── meta_init_context.cpython-39.pyc │ │ │ │ ├── modeling_mpt.cpython-310.pyc │ │ │ │ ├── modeling_mpt.cpython-39.pyc │ │ │ │ ├── norm.cpython-310.pyc │ │ │ │ ├── norm.cpython-39.pyc │ │ │ │ ├── param_init_fns.cpython-310.pyc │ │ │ │ └── param_init_fns.cpython-39.pyc │ │ │ │ ├── adapt_tokenizer.py │ │ │ │ ├── attention.py │ │ │ │ ├── blocks.py │ │ │ │ ├── configuration_mpt.py │ │ │ │ ├── custom_embedding.py │ │ │ │ ├── flash_attn_triton.py │ │ │ │ ├── hf_prefixlm_converter.py │ │ │ │ ├── meta_init_context.py │ │ │ │ ├── modeling_mpt.py │ │ │ │ ├── norm.py │ │ │ │ └── param_init_fns.py │ │ ├── llava_arch.py │ │ ├── make_delta.py │ │ ├── multimodal_encoder │ │ │ ├── __pycache__ │ │ │ │ ├── builder.cpython-310.pyc │ │ │ │ ├── builder.cpython-39.pyc │ │ │ │ ├── clip_encoder.cpython-310.pyc │ │ │ │ └── clip_encoder.cpython-39.pyc │ │ │ ├── builder.py │ │ │ └── clip_encoder.py │ │ ├── multimodal_projector │ │ │ ├── __pycache__ │ │ │ │ ├── builder.cpython-310.pyc │ │ │ │ └── builder.cpython-39.pyc │ │ │ └── builder.py │ │ └── utils.py │ ├── serve │ │ ├── __init__.py │ │ ├── __pycache__ │ │ │ ├── __init__.cpython-310.pyc │ │ │ └── cli.cpython-310.pyc │ │ ├── cli.py │ │ ├── controller.py │ │ ├── examples │ │ │ ├── extreme_ironing.jpg │ │ │ └── waterview.jpg │ │ ├── gradio_web_server.py │ │ ├── model_worker.py │ │ ├── register_worker.py │ │ └── test_message.py │ ├── train │ │ ├── llama_flash_attn_monkey_patch.py │ │ ├── llava_trainer.py │ │ ├── train.py │ │ └── train_mem.py │ └── utils.py │ ├── pyproject.toml │ └── scripts │ ├── convert_gqa_for_eval.py │ ├── convert_mmbench_for_submission.py │ ├── convert_mmvet_for_eval.py │ ├── convert_seed_for_submission.py │ ├── convert_sqa_to_llava.py │ ├── convert_sqa_to_llava_base_prompt.py │ ├── convert_vizwiz_for_submission.py │ ├── convert_vqav2_for_submission.py │ ├── extract_mm_projector.py │ ├── finetune.sh │ ├── finetune_full_schedule.sh │ ├── finetune_lora.sh │ ├── finetune_qlora.sh │ ├── finetune_sqa.sh │ ├── merge_lora_weights.py │ ├── pretrain.sh │ ├── sqa_eval_batch.sh │ ├── sqa_eval_gather.sh │ ├── v1_5 │ ├── eval │ │ ├── gqa.sh │ │ ├── llavabench.sh │ │ ├── mmbench.sh │ │ ├── mmbench_cn.sh │ │ ├── mme.sh │ │ ├── mmvet.sh │ │ ├── pope.sh │ │ ├── seed.sh │ │ ├── sqa.sh │ │ ├── textvqa.sh │ │ ├── vizwiz.sh │ │ └── vqav2.sh │ ├── finetune.sh │ └── pretrain.sh │ ├── zero2.json │ ├── zero3.json │ └── zero3_offload.json ├── simulator ├── __init__.py ├── build_grid.py ├── feature_extractor.py ├── grape_object.py ├── grid_sampler.py ├── multisensory_simulator.py ├── semantic_extractor.py └── utils │ ├── __init__.py │ ├── config.py │ ├── dataset_interface.py │ └── reconstruct3d.py └── utils ├── __init__.py ├── __pycache__ ├── __init__.cpython-37.pyc ├── __init__.cpython-39.pyc ├── cloud_point_utils.cpython-37.pyc ├── cloud_point_utils.cpython-39.pyc ├── config.cpython-37.pyc ├── config.cpython-39.pyc ├── dataset_interface.cpython-37.pyc └── dataset_interface.cpython-39.pyc ├── cloud_point_utils.py ├── config.py └── dataset_interface.py /README.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/README.md -------------------------------------------------------------------------------- /figs/method.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/figs/method.png -------------------------------------------------------------------------------- /figs/teaser.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/figs/teaser.png -------------------------------------------------------------------------------- /model_release/dataset.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/dataset.py -------------------------------------------------------------------------------- /model_release/fsdp_train.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/fsdp_train.py -------------------------------------------------------------------------------- /model_release/llava/README.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/README.md -------------------------------------------------------------------------------- /model_release/llava/docs/Customize_Component.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/docs/Customize_Component.md -------------------------------------------------------------------------------- /model_release/llava/docs/Data.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/docs/Data.md -------------------------------------------------------------------------------- /model_release/llava/docs/Evaluation.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/docs/Evaluation.md -------------------------------------------------------------------------------- /model_release/llava/docs/LLaVA_Bench.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/docs/LLaVA_Bench.md -------------------------------------------------------------------------------- /model_release/llava/docs/LLaVA_from_LLaMA2.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/docs/LLaVA_from_LLaMA2.md -------------------------------------------------------------------------------- /model_release/llava/docs/LoRA.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/docs/LoRA.md -------------------------------------------------------------------------------- /model_release/llava/docs/MODEL_ZOO.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/docs/MODEL_ZOO.md -------------------------------------------------------------------------------- /model_release/llava/docs/ScienceQA.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/docs/ScienceQA.md -------------------------------------------------------------------------------- /model_release/llava/docs/Windows.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/docs/Windows.md -------------------------------------------------------------------------------- /model_release/llava/images/demo_cli.gif: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/images/demo_cli.gif -------------------------------------------------------------------------------- /model_release/llava/images/llava_example_cmp.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/images/llava_example_cmp.png -------------------------------------------------------------------------------- /model_release/llava/images/llava_logo.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/images/llava_logo.png -------------------------------------------------------------------------------- /model_release/llava/images/llava_v1_5_radar.jpg: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/images/llava_v1_5_radar.jpg -------------------------------------------------------------------------------- /model_release/llava/llava.egg-info/PKG-INFO: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava.egg-info/PKG-INFO -------------------------------------------------------------------------------- /model_release/llava/llava.egg-info/SOURCES.txt: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava.egg-info/SOURCES.txt -------------------------------------------------------------------------------- /model_release/llava/llava.egg-info/dependency_links.txt: -------------------------------------------------------------------------------- 1 | 2 | -------------------------------------------------------------------------------- /model_release/llava/llava.egg-info/requires.txt: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava.egg-info/requires.txt -------------------------------------------------------------------------------- /model_release/llava/llava.egg-info/top_level.txt: -------------------------------------------------------------------------------- 1 | images 2 | llava 3 | -------------------------------------------------------------------------------- /model_release/llava/llava/__init__.py: -------------------------------------------------------------------------------- 1 | from .model import LlavaLlamaForCausalLM 2 | -------------------------------------------------------------------------------- /model_release/llava/llava/__pycache__/__init__.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/__pycache__/__init__.cpython-310.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/__pycache__/__init__.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/__pycache__/__init__.cpython-39.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/__pycache__/constants.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/__pycache__/constants.cpython-310.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/__pycache__/constants.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/__pycache__/constants.cpython-39.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/__pycache__/conversation.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/__pycache__/conversation.cpython-310.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/__pycache__/mm_utils.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/__pycache__/mm_utils.cpython-310.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/__pycache__/mm_utils.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/__pycache__/mm_utils.cpython-39.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/__pycache__/utils.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/__pycache__/utils.cpython-310.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/constants.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/constants.py -------------------------------------------------------------------------------- /model_release/llava/llava/conversation.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/conversation.py -------------------------------------------------------------------------------- /model_release/llava/llava/eval/eval_gpt_review.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/eval_gpt_review.py -------------------------------------------------------------------------------- /model_release/llava/llava/eval/eval_gpt_review_bench.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/eval_gpt_review_bench.py -------------------------------------------------------------------------------- /model_release/llava/llava/eval/eval_gpt_review_visual.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/eval_gpt_review_visual.py -------------------------------------------------------------------------------- /model_release/llava/llava/eval/eval_pope.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/eval_pope.py -------------------------------------------------------------------------------- /model_release/llava/llava/eval/eval_science_qa.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/eval_science_qa.py -------------------------------------------------------------------------------- /model_release/llava/llava/eval/eval_science_qa_gpt4.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/eval_science_qa_gpt4.py -------------------------------------------------------------------------------- /model_release/llava/llava/eval/eval_science_qa_gpt4_requery.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/eval_science_qa_gpt4_requery.py -------------------------------------------------------------------------------- /model_release/llava/llava/eval/eval_textvqa.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/eval_textvqa.py -------------------------------------------------------------------------------- /model_release/llava/llava/eval/generate_webpage_data_from_table.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/generate_webpage_data_from_table.py -------------------------------------------------------------------------------- /model_release/llava/llava/eval/m4c_evaluator.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/m4c_evaluator.py -------------------------------------------------------------------------------- /model_release/llava/llava/eval/model_qa.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/model_qa.py -------------------------------------------------------------------------------- /model_release/llava/llava/eval/model_vqa.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/model_vqa.py -------------------------------------------------------------------------------- /model_release/llava/llava/eval/model_vqa_loader.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/model_vqa_loader.py -------------------------------------------------------------------------------- /model_release/llava/llava/eval/model_vqa_mmbench.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/model_vqa_mmbench.py -------------------------------------------------------------------------------- /model_release/llava/llava/eval/model_vqa_science.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/model_vqa_science.py -------------------------------------------------------------------------------- /model_release/llava/llava/eval/qa_baseline_gpt35.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/qa_baseline_gpt35.py -------------------------------------------------------------------------------- /model_release/llava/llava/eval/run_llava.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/run_llava.py -------------------------------------------------------------------------------- /model_release/llava/llava/eval/summarize_gpt_review.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/summarize_gpt_review.py -------------------------------------------------------------------------------- /model_release/llava/llava/eval/table/answer/answer_alpaca-13b.jsonl: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/table/answer/answer_alpaca-13b.jsonl -------------------------------------------------------------------------------- /model_release/llava/llava/eval/table/answer/answer_bard.jsonl: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/table/answer/answer_bard.jsonl -------------------------------------------------------------------------------- /model_release/llava/llava/eval/table/answer/answer_gpt35.jsonl: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/table/answer/answer_gpt35.jsonl -------------------------------------------------------------------------------- /model_release/llava/llava/eval/table/answer/answer_llama-13b.jsonl: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/table/answer/answer_llama-13b.jsonl -------------------------------------------------------------------------------- /model_release/llava/llava/eval/table/answer/answer_vicuna-13b.jsonl: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/table/answer/answer_vicuna-13b.jsonl -------------------------------------------------------------------------------- /model_release/llava/llava/eval/table/caps_boxes_coco2014_val_80.jsonl: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/table/caps_boxes_coco2014_val_80.jsonl -------------------------------------------------------------------------------- /model_release/llava/llava/eval/table/model.jsonl: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/table/model.jsonl -------------------------------------------------------------------------------- /model_release/llava/llava/eval/table/prompt.jsonl: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/table/prompt.jsonl -------------------------------------------------------------------------------- /model_release/llava/llava/eval/table/question.jsonl: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/table/question.jsonl -------------------------------------------------------------------------------- /model_release/llava/llava/eval/table/review/review_alpaca-13b_vicuna-13b.jsonl: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/table/review/review_alpaca-13b_vicuna-13b.jsonl -------------------------------------------------------------------------------- /model_release/llava/llava/eval/table/review/review_bard_vicuna-13b.jsonl: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/table/review/review_bard_vicuna-13b.jsonl -------------------------------------------------------------------------------- /model_release/llava/llava/eval/table/review/review_gpt35_vicuna-13b.jsonl: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/table/review/review_gpt35_vicuna-13b.jsonl -------------------------------------------------------------------------------- /model_release/llava/llava/eval/table/review/review_llama-13b_vicuna-13b.jsonl: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/table/review/review_llama-13b_vicuna-13b.jsonl -------------------------------------------------------------------------------- /model_release/llava/llava/eval/table/reviewer.jsonl: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/table/reviewer.jsonl -------------------------------------------------------------------------------- /model_release/llava/llava/eval/table/rule.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/table/rule.json -------------------------------------------------------------------------------- /model_release/llava/llava/eval/webpage/figures/alpaca.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/webpage/figures/alpaca.png -------------------------------------------------------------------------------- /model_release/llava/llava/eval/webpage/figures/bard.jpg: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/webpage/figures/bard.jpg -------------------------------------------------------------------------------- /model_release/llava/llava/eval/webpage/figures/chatgpt.svg: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/webpage/figures/chatgpt.svg -------------------------------------------------------------------------------- /model_release/llava/llava/eval/webpage/figures/llama.jpg: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/webpage/figures/llama.jpg -------------------------------------------------------------------------------- /model_release/llava/llava/eval/webpage/figures/swords_FILL0_wght300_GRAD0_opsz48.svg: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/webpage/figures/swords_FILL0_wght300_GRAD0_opsz48.svg -------------------------------------------------------------------------------- /model_release/llava/llava/eval/webpage/figures/vicuna.jpeg: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/webpage/figures/vicuna.jpeg -------------------------------------------------------------------------------- /model_release/llava/llava/eval/webpage/index.html: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/webpage/index.html -------------------------------------------------------------------------------- /model_release/llava/llava/eval/webpage/script.js: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/webpage/script.js -------------------------------------------------------------------------------- /model_release/llava/llava/eval/webpage/styles.css: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/eval/webpage/styles.css -------------------------------------------------------------------------------- /model_release/llava/llava/mm_utils.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/mm_utils.py -------------------------------------------------------------------------------- /model_release/llava/llava/model/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/__init__.py -------------------------------------------------------------------------------- /model_release/llava/llava/model/__pycache__/__init__.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/__pycache__/__init__.cpython-310.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/__pycache__/__init__.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/__pycache__/__init__.cpython-39.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/__pycache__/builder.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/__pycache__/builder.cpython-310.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/__pycache__/builder.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/__pycache__/builder.cpython-39.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/__pycache__/llava_arch.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/__pycache__/llava_arch.cpython-310.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/__pycache__/llava_arch.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/__pycache__/llava_arch.cpython-39.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/apply_delta.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/apply_delta.py -------------------------------------------------------------------------------- /model_release/llava/llava/model/builder.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/builder.py -------------------------------------------------------------------------------- /model_release/llava/llava/model/consolidate.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/consolidate.py -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/__pycache__/llava_llama.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/__pycache__/llava_llama.cpython-310.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/__pycache__/llava_llama.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/__pycache__/llava_llama.cpython-39.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/__pycache__/llava_mpt.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/__pycache__/llava_mpt.cpython-310.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/__pycache__/llava_mpt.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/__pycache__/llava_mpt.cpython-39.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/llava_llama.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/llava_llama.py -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/llava_mpt.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/llava_mpt.py -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/__pycache__/adapt_tokenizer.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/__pycache__/adapt_tokenizer.cpython-310.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/__pycache__/adapt_tokenizer.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/__pycache__/adapt_tokenizer.cpython-39.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/__pycache__/attention.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/__pycache__/attention.cpython-310.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/__pycache__/attention.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/__pycache__/attention.cpython-39.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/__pycache__/blocks.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/__pycache__/blocks.cpython-310.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/__pycache__/blocks.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/__pycache__/blocks.cpython-39.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/__pycache__/configuration_mpt.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/__pycache__/configuration_mpt.cpython-310.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/__pycache__/configuration_mpt.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/__pycache__/configuration_mpt.cpython-39.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/__pycache__/custom_embedding.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/__pycache__/custom_embedding.cpython-310.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/__pycache__/custom_embedding.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/__pycache__/custom_embedding.cpython-39.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/__pycache__/flash_attn_triton.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/__pycache__/flash_attn_triton.cpython-310.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/__pycache__/flash_attn_triton.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/__pycache__/flash_attn_triton.cpython-39.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/__pycache__/hf_prefixlm_converter.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/__pycache__/hf_prefixlm_converter.cpython-310.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/__pycache__/hf_prefixlm_converter.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/__pycache__/hf_prefixlm_converter.cpython-39.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/__pycache__/meta_init_context.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/__pycache__/meta_init_context.cpython-310.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/__pycache__/meta_init_context.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/__pycache__/meta_init_context.cpython-39.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/__pycache__/modeling_mpt.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/__pycache__/modeling_mpt.cpython-310.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/__pycache__/modeling_mpt.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/__pycache__/modeling_mpt.cpython-39.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/__pycache__/norm.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/__pycache__/norm.cpython-310.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/__pycache__/norm.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/__pycache__/norm.cpython-39.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/__pycache__/param_init_fns.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/__pycache__/param_init_fns.cpython-310.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/__pycache__/param_init_fns.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/__pycache__/param_init_fns.cpython-39.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/adapt_tokenizer.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/adapt_tokenizer.py -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/attention.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/attention.py -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/blocks.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/blocks.py -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/configuration_mpt.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/configuration_mpt.py -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/custom_embedding.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/custom_embedding.py -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/flash_attn_triton.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/flash_attn_triton.py -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/hf_prefixlm_converter.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/hf_prefixlm_converter.py -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/meta_init_context.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/meta_init_context.py -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/modeling_mpt.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/modeling_mpt.py -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/norm.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/norm.py -------------------------------------------------------------------------------- /model_release/llava/llava/model/language_model/mpt/param_init_fns.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/language_model/mpt/param_init_fns.py -------------------------------------------------------------------------------- /model_release/llava/llava/model/llava_arch.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/llava_arch.py -------------------------------------------------------------------------------- /model_release/llava/llava/model/make_delta.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/make_delta.py -------------------------------------------------------------------------------- /model_release/llava/llava/model/multimodal_encoder/__pycache__/builder.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/multimodal_encoder/__pycache__/builder.cpython-310.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/multimodal_encoder/__pycache__/builder.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/multimodal_encoder/__pycache__/builder.cpython-39.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/multimodal_encoder/__pycache__/clip_encoder.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/multimodal_encoder/__pycache__/clip_encoder.cpython-310.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/multimodal_encoder/__pycache__/clip_encoder.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/multimodal_encoder/__pycache__/clip_encoder.cpython-39.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/multimodal_encoder/builder.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/multimodal_encoder/builder.py -------------------------------------------------------------------------------- /model_release/llava/llava/model/multimodal_encoder/clip_encoder.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/multimodal_encoder/clip_encoder.py -------------------------------------------------------------------------------- /model_release/llava/llava/model/multimodal_projector/__pycache__/builder.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/multimodal_projector/__pycache__/builder.cpython-310.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/multimodal_projector/__pycache__/builder.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/multimodal_projector/__pycache__/builder.cpython-39.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/model/multimodal_projector/builder.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/multimodal_projector/builder.py -------------------------------------------------------------------------------- /model_release/llava/llava/model/utils.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/model/utils.py -------------------------------------------------------------------------------- /model_release/llava/llava/serve/__init__.py: -------------------------------------------------------------------------------- 1 | -------------------------------------------------------------------------------- /model_release/llava/llava/serve/__pycache__/__init__.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/serve/__pycache__/__init__.cpython-310.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/serve/__pycache__/cli.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/serve/__pycache__/cli.cpython-310.pyc -------------------------------------------------------------------------------- /model_release/llava/llava/serve/cli.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/serve/cli.py -------------------------------------------------------------------------------- /model_release/llava/llava/serve/controller.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/serve/controller.py -------------------------------------------------------------------------------- /model_release/llava/llava/serve/examples/extreme_ironing.jpg: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/serve/examples/extreme_ironing.jpg -------------------------------------------------------------------------------- /model_release/llava/llava/serve/examples/waterview.jpg: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/serve/examples/waterview.jpg -------------------------------------------------------------------------------- /model_release/llava/llava/serve/gradio_web_server.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/serve/gradio_web_server.py -------------------------------------------------------------------------------- /model_release/llava/llava/serve/model_worker.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/serve/model_worker.py -------------------------------------------------------------------------------- /model_release/llava/llava/serve/register_worker.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/serve/register_worker.py -------------------------------------------------------------------------------- /model_release/llava/llava/serve/test_message.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/serve/test_message.py -------------------------------------------------------------------------------- /model_release/llava/llava/train/llama_flash_attn_monkey_patch.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/train/llama_flash_attn_monkey_patch.py -------------------------------------------------------------------------------- /model_release/llava/llava/train/llava_trainer.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/train/llava_trainer.py -------------------------------------------------------------------------------- /model_release/llava/llava/train/train.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/train/train.py -------------------------------------------------------------------------------- /model_release/llava/llava/train/train_mem.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/train/train_mem.py -------------------------------------------------------------------------------- /model_release/llava/llava/utils.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/llava/utils.py -------------------------------------------------------------------------------- /model_release/llava/pyproject.toml: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/pyproject.toml -------------------------------------------------------------------------------- /model_release/llava/scripts/convert_gqa_for_eval.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/convert_gqa_for_eval.py -------------------------------------------------------------------------------- /model_release/llava/scripts/convert_mmbench_for_submission.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/convert_mmbench_for_submission.py -------------------------------------------------------------------------------- /model_release/llava/scripts/convert_mmvet_for_eval.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/convert_mmvet_for_eval.py -------------------------------------------------------------------------------- /model_release/llava/scripts/convert_seed_for_submission.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/convert_seed_for_submission.py -------------------------------------------------------------------------------- /model_release/llava/scripts/convert_sqa_to_llava.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/convert_sqa_to_llava.py -------------------------------------------------------------------------------- /model_release/llava/scripts/convert_sqa_to_llava_base_prompt.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/convert_sqa_to_llava_base_prompt.py -------------------------------------------------------------------------------- /model_release/llava/scripts/convert_vizwiz_for_submission.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/convert_vizwiz_for_submission.py -------------------------------------------------------------------------------- /model_release/llava/scripts/convert_vqav2_for_submission.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/convert_vqav2_for_submission.py -------------------------------------------------------------------------------- /model_release/llava/scripts/extract_mm_projector.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/extract_mm_projector.py -------------------------------------------------------------------------------- /model_release/llava/scripts/finetune.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/finetune.sh -------------------------------------------------------------------------------- /model_release/llava/scripts/finetune_full_schedule.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/finetune_full_schedule.sh -------------------------------------------------------------------------------- /model_release/llava/scripts/finetune_lora.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/finetune_lora.sh -------------------------------------------------------------------------------- /model_release/llava/scripts/finetune_qlora.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/finetune_qlora.sh -------------------------------------------------------------------------------- /model_release/llava/scripts/finetune_sqa.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/finetune_sqa.sh -------------------------------------------------------------------------------- /model_release/llava/scripts/merge_lora_weights.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/merge_lora_weights.py -------------------------------------------------------------------------------- /model_release/llava/scripts/pretrain.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/pretrain.sh -------------------------------------------------------------------------------- /model_release/llava/scripts/sqa_eval_batch.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/sqa_eval_batch.sh -------------------------------------------------------------------------------- /model_release/llava/scripts/sqa_eval_gather.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/sqa_eval_gather.sh -------------------------------------------------------------------------------- /model_release/llava/scripts/v1_5/eval/gqa.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/v1_5/eval/gqa.sh -------------------------------------------------------------------------------- /model_release/llava/scripts/v1_5/eval/llavabench.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/v1_5/eval/llavabench.sh -------------------------------------------------------------------------------- /model_release/llava/scripts/v1_5/eval/mmbench.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/v1_5/eval/mmbench.sh -------------------------------------------------------------------------------- /model_release/llava/scripts/v1_5/eval/mmbench_cn.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/v1_5/eval/mmbench_cn.sh -------------------------------------------------------------------------------- /model_release/llava/scripts/v1_5/eval/mme.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/v1_5/eval/mme.sh -------------------------------------------------------------------------------- /model_release/llava/scripts/v1_5/eval/mmvet.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/v1_5/eval/mmvet.sh -------------------------------------------------------------------------------- /model_release/llava/scripts/v1_5/eval/pope.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/v1_5/eval/pope.sh -------------------------------------------------------------------------------- /model_release/llava/scripts/v1_5/eval/seed.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/v1_5/eval/seed.sh -------------------------------------------------------------------------------- /model_release/llava/scripts/v1_5/eval/sqa.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/v1_5/eval/sqa.sh -------------------------------------------------------------------------------- /model_release/llava/scripts/v1_5/eval/textvqa.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/v1_5/eval/textvqa.sh -------------------------------------------------------------------------------- /model_release/llava/scripts/v1_5/eval/vizwiz.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/v1_5/eval/vizwiz.sh -------------------------------------------------------------------------------- /model_release/llava/scripts/v1_5/eval/vqav2.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/v1_5/eval/vqav2.sh -------------------------------------------------------------------------------- /model_release/llava/scripts/v1_5/finetune.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/v1_5/finetune.sh -------------------------------------------------------------------------------- /model_release/llava/scripts/v1_5/pretrain.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/v1_5/pretrain.sh -------------------------------------------------------------------------------- /model_release/llava/scripts/zero2.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/zero2.json -------------------------------------------------------------------------------- /model_release/llava/scripts/zero3.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/zero3.json -------------------------------------------------------------------------------- /model_release/llava/scripts/zero3_offload.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/model_release/llava/scripts/zero3_offload.json -------------------------------------------------------------------------------- /simulator/__init__.py: -------------------------------------------------------------------------------- 1 | -------------------------------------------------------------------------------- /simulator/build_grid.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/simulator/build_grid.py -------------------------------------------------------------------------------- /simulator/feature_extractor.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/simulator/feature_extractor.py -------------------------------------------------------------------------------- /simulator/grape_object.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/simulator/grape_object.py -------------------------------------------------------------------------------- /simulator/grid_sampler.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/simulator/grid_sampler.py -------------------------------------------------------------------------------- /simulator/multisensory_simulator.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/simulator/multisensory_simulator.py -------------------------------------------------------------------------------- /simulator/semantic_extractor.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/simulator/semantic_extractor.py -------------------------------------------------------------------------------- /simulator/utils/__init__.py: -------------------------------------------------------------------------------- 1 | -------------------------------------------------------------------------------- /simulator/utils/config.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/simulator/utils/config.py -------------------------------------------------------------------------------- /simulator/utils/dataset_interface.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/simulator/utils/dataset_interface.py -------------------------------------------------------------------------------- /simulator/utils/reconstruct3d.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/simulator/utils/reconstruct3d.py -------------------------------------------------------------------------------- /utils/__init__.py: -------------------------------------------------------------------------------- 1 | -------------------------------------------------------------------------------- /utils/__pycache__/__init__.cpython-37.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/utils/__pycache__/__init__.cpython-37.pyc -------------------------------------------------------------------------------- /utils/__pycache__/__init__.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/utils/__pycache__/__init__.cpython-39.pyc -------------------------------------------------------------------------------- /utils/__pycache__/cloud_point_utils.cpython-37.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/utils/__pycache__/cloud_point_utils.cpython-37.pyc -------------------------------------------------------------------------------- /utils/__pycache__/cloud_point_utils.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/utils/__pycache__/cloud_point_utils.cpython-39.pyc -------------------------------------------------------------------------------- /utils/__pycache__/config.cpython-37.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/utils/__pycache__/config.cpython-37.pyc -------------------------------------------------------------------------------- /utils/__pycache__/config.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/utils/__pycache__/config.cpython-39.pyc -------------------------------------------------------------------------------- /utils/__pycache__/dataset_interface.cpython-37.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/utils/__pycache__/dataset_interface.cpython-37.pyc -------------------------------------------------------------------------------- /utils/__pycache__/dataset_interface.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/utils/__pycache__/dataset_interface.cpython-39.pyc -------------------------------------------------------------------------------- /utils/cloud_point_utils.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/utils/cloud_point_utils.py -------------------------------------------------------------------------------- /utils/config.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/utils/config.py -------------------------------------------------------------------------------- /utils/dataset_interface.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/UMass-Embodied-AGI/MultiPLY/HEAD/utils/dataset_interface.py --------------------------------------------------------------------------------