├── .gitignore ├── INSTALL.md ├── LICENSE ├── README.md ├── configs ├── scan_vitB.yaml ├── scan_vitL.yaml └── scan_vitL_demo.yaml ├── datasets ├── DATASETS.md ├── prepare_ade20k_full_sem_seg.py ├── prepare_ade20k_sem_seg.py ├── prepare_coco_stuff_sem_seg.py ├── prepare_pascal_context.py └── prepare_voc_sem_seg.py ├── demo.py ├── imgs ├── cs.png ├── pipeline.png ├── results.png └── visual.png ├── open_clip_training ├── .github │ └── workflows │ │ ├── ci.yml │ │ ├── clear-cache.yml │ │ └── python-publish.yml ├── .gitignore ├── CITATION.cff ├── HISTORY.md ├── LICENSE ├── MANIFEST.in ├── Makefile ├── README.md ├── docs │ ├── Interacting_with_open_clip.ipynb │ ├── Interacting_with_open_coca.ipynb │ ├── LOW_ACC.md │ ├── PRETRAINED.md │ ├── clip_conceptual_captions.md │ ├── clipa.md │ ├── datacomp_models.md │ ├── openclip_results.csv │ └── script_examples │ │ ├── clipa │ │ ├── vit_b16 │ │ │ ├── i50_t16_finetune.sh │ │ │ └── i50_t16_pretrain.sh │ │ └── vit_l16 │ │ │ ├── i17_t16_finetune.sh │ │ │ ├── i17_t16_pretrain.sh │ │ │ ├── i37_t8_finetune.sh │ │ │ └── i37_t8_pretrain.sh │ │ ├── clipav2 │ │ └── vit_h14 │ │ │ ├── i257_t32_finetunex4.sh │ │ │ ├── i50_t8_pretrain.sh │ │ │ └── i577_t32_finetunex1.sh │ │ └── stability_example.sh ├── pytest.ini ├── scripts │ ├── clipav1_vit_l16_i37_t8.sh │ ├── clipav2_vit_h14_i84_224_336_cl32_gap_datacomp1b.sh │ ├── h14_224_32_finetune.sh │ └── h14_84_8_pretrain.sh ├── setup.py ├── src │ ├── clip_adapter │ │ └── clip_adapter.py │ ├── open_clip │ │ ├── __init__.py │ │ ├── big_vision.py │ │ ├── bpe_simple_vocab_16e6.txt.gz │ │ ├── coca_model.py │ │ ├── constants.py │ │ ├── factory.py │ │ ├── generation_utils.py │ │ ├── hf_configs.py │ │ ├── hf_model.py │ │ ├── loss.py │ │ ├── model.py │ │ ├── model_configs │ │ │ ├── EVA01-g-14-plus.json │ │ │ ├── EVA01-g-14.json │ │ │ ├── EVA02-B-16.json │ │ │ ├── EVA02-E-14-plus.json │ │ │ ├── EVA02-E-14.json │ │ │ ├── EVA02-L-14-336.json │ │ │ ├── EVA02-L-14.json │ │ │ ├── RN101-quickgelu.json │ │ │ ├── RN101.json │ │ │ ├── RN50-quickgelu.json │ │ │ ├── RN50.json │ │ │ ├── RN50x16.json │ │ │ ├── RN50x4.json │ │ │ ├── RN50x64.json │ │ │ ├── ViT-B-16-SigLIP-256.json │ │ │ ├── ViT-B-16-SigLIP-384.json │ │ │ ├── ViT-B-16-SigLIP-512.json │ │ │ ├── ViT-B-16-SigLIP-i18n-256.json │ │ │ ├── ViT-B-16-SigLIP.json │ │ │ ├── ViT-B-16-plus-240.json │ │ │ ├── ViT-B-16-plus.json │ │ │ ├── ViT-B-16-quickgelu.json │ │ │ ├── ViT-B-16.json │ │ │ ├── ViT-B-32-256.json │ │ │ ├── ViT-B-32-plus-256.json │ │ │ ├── ViT-B-32-quickgelu.json │ │ │ ├── ViT-B-32.json │ │ │ ├── ViT-H-14-378-quickgelu.json │ │ │ ├── ViT-H-14-CLIPA-336.json │ │ │ ├── ViT-H-14-CLIPA.json │ │ │ ├── ViT-H-14-quickgelu.json │ │ │ ├── ViT-H-14.json │ │ │ ├── ViT-H-16.json │ │ │ ├── ViT-L-14-280.json │ │ │ ├── ViT-L-14-336.json │ │ │ ├── ViT-L-14-CLIPA-336.json │ │ │ ├── ViT-L-14-CLIPA.json │ │ │ ├── ViT-L-14-quickgelu.json │ │ │ ├── ViT-L-14.json │ │ │ ├── ViT-L-16-320.json │ │ │ ├── ViT-L-16-SigLIP-256.json │ │ │ ├── ViT-L-16-SigLIP-384.json │ │ │ ├── ViT-L-16.json │ │ │ ├── ViT-M-16-alt.json │ │ │ ├── ViT-M-16.json │ │ │ ├── ViT-M-32-alt.json │ │ │ ├── ViT-M-32.json │ │ │ ├── ViT-S-16-alt.json │ │ │ ├── ViT-S-16.json │ │ │ ├── ViT-S-32-alt.json │ │ │ ├── ViT-S-32.json │ │ │ ├── ViT-SO400M-14-SigLIP-384.json │ │ │ ├── ViT-SO400M-14-SigLIP.json │ │ │ ├── ViT-bigG-14-CLIPA-336.json │ │ │ ├── ViT-bigG-14-CLIPA.json │ │ │ ├── ViT-bigG-14.json │ │ │ ├── ViT-e-14.json │ │ │ ├── ViT-g-14.json │ │ │ ├── coca_ViT-B-32.json │ │ │ ├── coca_ViT-L-14.json │ │ │ ├── coca_base.json │ │ │ ├── coca_roberta-ViT-B-32.json │ │ │ ├── convnext_base.json │ │ │ ├── convnext_base_w.json │ │ │ ├── convnext_base_w_320.json │ │ │ ├── convnext_large.json │ │ │ ├── convnext_large_d.json │ │ │ ├── convnext_large_d_320.json │ │ │ ├── convnext_small.json │ │ │ ├── convnext_tiny.json │ │ │ ├── convnext_xlarge.json │ │ │ ├── convnext_xxlarge.json │ │ │ ├── convnext_xxlarge_320.json │ │ │ ├── mt5-base-ViT-B-32.json │ │ │ ├── mt5-xl-ViT-H-14.json │ │ │ ├── nllb-clip-base-siglip.json │ │ │ ├── nllb-clip-base.json │ │ │ ├── nllb-clip-large-siglip.json │ │ │ ├── nllb-clip-large.json │ │ │ ├── roberta-ViT-B-32.json │ │ │ ├── swin_base_patch4_window7_224.json │ │ │ ├── vit_medium_patch16_gap_256.json │ │ │ ├── vit_relpos_medium_patch16_cls_224.json │ │ │ ├── xlm-roberta-base-ViT-B-32.json │ │ │ └── xlm-roberta-large-ViT-H-14.json │ │ ├── modified_resnet.py │ │ ├── openai.py │ │ ├── pos_embed.py │ │ ├── pretrained.py │ │ ├── push_to_hf_hub.py │ │ ├── timm_model.py │ │ ├── tokenizer.py │ │ ├── transform.py │ │ ├── transformer.py │ │ ├── utils.py │ │ ├── version.py │ │ ├── zero_shot_classifier.py │ │ └── zero_shot_metadata.py │ ├── scripts │ │ ├── 1cap_finetune_VitL.sh │ │ └── finetune_VitL_with_mask.sh │ └── training │ │ ├── .gitignore │ │ ├── __init__.py │ │ ├── ade150_zeroshot_data.py │ │ ├── data.py │ │ ├── distributed.py │ │ ├── file_utils.py │ │ ├── main.py │ │ ├── params.py │ │ ├── precision.py │ │ ├── profiler.py │ │ ├── scheduler.py │ │ ├── train.py │ │ └── zero_shot.py ├── tests │ ├── test_download_pretrained.py │ ├── test_hf_model.py │ ├── test_inference.py │ ├── test_inference_simple.py │ ├── test_num_shards.py │ ├── test_training_simple.py │ ├── test_wds.py │ └── util_test.py └── tutorials │ └── int8_tutorial.ipynb ├── requirements.txt ├── scan ├── __init__.py ├── config.py ├── data │ ├── __init__.py │ ├── augmentations.py │ ├── build.py │ ├── dataset_mappers │ │ ├── __init__.py │ │ └── mask_former_semantic_dataset_mapper.py │ └── datasets │ │ ├── __init__.py │ │ ├── csv_data.py │ │ ├── register_ade20k_full.py │ │ ├── register_cc3m.py │ │ ├── register_coco_stuff.py │ │ ├── register_pascal_context.py │ │ └── register_voc_seg.py ├── evaluation │ ├── __init__.py │ └── generalized_sem_seg_evaluation.py ├── frequency.py ├── maskformer_model.py ├── modeling │ ├── __init__.py │ ├── backbone │ │ ├── __init__.py │ │ ├── clip_resnet.py │ │ └── swin.py │ ├── clip_adapter │ │ ├── __init__.py │ │ ├── adapter.py │ │ ├── text_template.py │ │ └── utils.py │ ├── criterion.py │ ├── matcher.py │ ├── meta_arch │ │ ├── __init__.py │ │ ├── mask_former_head.py │ │ └── per_pixel_baseline.py │ ├── pixel_decoder │ │ ├── __init__.py │ │ ├── fpn.py │ │ ├── msdeformattn.py │ │ └── ops │ │ │ ├── functions │ │ │ ├── __init__.py │ │ │ └── ms_deform_attn_func.py │ │ │ ├── make.sh │ │ │ ├── modules │ │ │ ├── __init__.py │ │ │ └── ms_deform_attn.py │ │ │ ├── setup.py │ │ │ ├── src │ │ │ ├── cpu │ │ │ │ ├── ms_deform_attn_cpu.cpp │ │ │ │ └── ms_deform_attn_cpu.h │ │ │ ├── cuda │ │ │ │ ├── ms_deform_attn_cuda.cu │ │ │ │ ├── ms_deform_attn_cuda.h │ │ │ │ └── ms_deform_im2col_cuda.cuh │ │ │ ├── ms_deform_attn.h │ │ │ └── vision.cpp │ │ │ └── test.py │ └── transformer_decoder │ │ ├── __init__.py │ │ ├── mask2former_transformer_decoder.py │ │ ├── maskformer_transformer_decoder.py │ │ ├── open_vocab_mask2former_predictor.py │ │ ├── position_encoding.py │ │ └── transformer.py ├── ovseg_model.py ├── test_time_augmentation.py └── utils │ ├── __init__.py │ ├── events.py │ ├── misc.py │ ├── post_process_utils.py │ └── predictor.py ├── tools ├── convert-pretrained-clip-model-to-d2.py ├── convert-pretrained-swin-model-to-d2.py ├── convert-torchvision-to-d2.py └── replace_clip.py └── train_net.py /.gitignore: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/.gitignore -------------------------------------------------------------------------------- /INSTALL.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/INSTALL.md -------------------------------------------------------------------------------- /LICENSE: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/LICENSE -------------------------------------------------------------------------------- /README.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/README.md -------------------------------------------------------------------------------- /configs/scan_vitB.yaml: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/configs/scan_vitB.yaml -------------------------------------------------------------------------------- /configs/scan_vitL.yaml: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/configs/scan_vitL.yaml -------------------------------------------------------------------------------- /configs/scan_vitL_demo.yaml: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/configs/scan_vitL_demo.yaml -------------------------------------------------------------------------------- /datasets/DATASETS.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/datasets/DATASETS.md -------------------------------------------------------------------------------- /datasets/prepare_ade20k_full_sem_seg.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/datasets/prepare_ade20k_full_sem_seg.py -------------------------------------------------------------------------------- /datasets/prepare_ade20k_sem_seg.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/datasets/prepare_ade20k_sem_seg.py -------------------------------------------------------------------------------- /datasets/prepare_coco_stuff_sem_seg.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/datasets/prepare_coco_stuff_sem_seg.py -------------------------------------------------------------------------------- /datasets/prepare_pascal_context.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/datasets/prepare_pascal_context.py -------------------------------------------------------------------------------- /datasets/prepare_voc_sem_seg.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/datasets/prepare_voc_sem_seg.py -------------------------------------------------------------------------------- /demo.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/demo.py -------------------------------------------------------------------------------- /imgs/cs.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/imgs/cs.png -------------------------------------------------------------------------------- /imgs/pipeline.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/imgs/pipeline.png -------------------------------------------------------------------------------- /imgs/results.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/imgs/results.png -------------------------------------------------------------------------------- /imgs/visual.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/imgs/visual.png -------------------------------------------------------------------------------- /open_clip_training/.github/workflows/ci.yml: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/.github/workflows/ci.yml -------------------------------------------------------------------------------- /open_clip_training/.github/workflows/clear-cache.yml: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/.github/workflows/clear-cache.yml -------------------------------------------------------------------------------- /open_clip_training/.github/workflows/python-publish.yml: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/.github/workflows/python-publish.yml -------------------------------------------------------------------------------- /open_clip_training/.gitignore: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/.gitignore -------------------------------------------------------------------------------- /open_clip_training/CITATION.cff: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/CITATION.cff -------------------------------------------------------------------------------- /open_clip_training/HISTORY.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/HISTORY.md -------------------------------------------------------------------------------- /open_clip_training/LICENSE: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/LICENSE -------------------------------------------------------------------------------- /open_clip_training/MANIFEST.in: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/MANIFEST.in -------------------------------------------------------------------------------- /open_clip_training/Makefile: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/Makefile -------------------------------------------------------------------------------- /open_clip_training/README.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/README.md -------------------------------------------------------------------------------- /open_clip_training/docs/Interacting_with_open_clip.ipynb: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/docs/Interacting_with_open_clip.ipynb -------------------------------------------------------------------------------- /open_clip_training/docs/Interacting_with_open_coca.ipynb: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/docs/Interacting_with_open_coca.ipynb -------------------------------------------------------------------------------- /open_clip_training/docs/LOW_ACC.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/docs/LOW_ACC.md -------------------------------------------------------------------------------- /open_clip_training/docs/PRETRAINED.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/docs/PRETRAINED.md -------------------------------------------------------------------------------- /open_clip_training/docs/clip_conceptual_captions.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/docs/clip_conceptual_captions.md -------------------------------------------------------------------------------- /open_clip_training/docs/clipa.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/docs/clipa.md -------------------------------------------------------------------------------- /open_clip_training/docs/datacomp_models.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/docs/datacomp_models.md -------------------------------------------------------------------------------- /open_clip_training/docs/openclip_results.csv: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/docs/openclip_results.csv -------------------------------------------------------------------------------- /open_clip_training/docs/script_examples/clipa/vit_b16/i50_t16_finetune.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/docs/script_examples/clipa/vit_b16/i50_t16_finetune.sh -------------------------------------------------------------------------------- /open_clip_training/docs/script_examples/clipa/vit_b16/i50_t16_pretrain.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/docs/script_examples/clipa/vit_b16/i50_t16_pretrain.sh -------------------------------------------------------------------------------- /open_clip_training/docs/script_examples/clipa/vit_l16/i17_t16_finetune.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/docs/script_examples/clipa/vit_l16/i17_t16_finetune.sh -------------------------------------------------------------------------------- /open_clip_training/docs/script_examples/clipa/vit_l16/i17_t16_pretrain.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/docs/script_examples/clipa/vit_l16/i17_t16_pretrain.sh -------------------------------------------------------------------------------- /open_clip_training/docs/script_examples/clipa/vit_l16/i37_t8_finetune.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/docs/script_examples/clipa/vit_l16/i37_t8_finetune.sh -------------------------------------------------------------------------------- /open_clip_training/docs/script_examples/clipa/vit_l16/i37_t8_pretrain.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/docs/script_examples/clipa/vit_l16/i37_t8_pretrain.sh -------------------------------------------------------------------------------- /open_clip_training/docs/script_examples/clipav2/vit_h14/i257_t32_finetunex4.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/docs/script_examples/clipav2/vit_h14/i257_t32_finetunex4.sh -------------------------------------------------------------------------------- /open_clip_training/docs/script_examples/clipav2/vit_h14/i50_t8_pretrain.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/docs/script_examples/clipav2/vit_h14/i50_t8_pretrain.sh -------------------------------------------------------------------------------- /open_clip_training/docs/script_examples/clipav2/vit_h14/i577_t32_finetunex1.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/docs/script_examples/clipav2/vit_h14/i577_t32_finetunex1.sh -------------------------------------------------------------------------------- /open_clip_training/docs/script_examples/stability_example.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/docs/script_examples/stability_example.sh -------------------------------------------------------------------------------- /open_clip_training/pytest.ini: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/pytest.ini -------------------------------------------------------------------------------- /open_clip_training/scripts/clipav1_vit_l16_i37_t8.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/scripts/clipav1_vit_l16_i37_t8.sh -------------------------------------------------------------------------------- /open_clip_training/scripts/clipav2_vit_h14_i84_224_336_cl32_gap_datacomp1b.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/scripts/clipav2_vit_h14_i84_224_336_cl32_gap_datacomp1b.sh -------------------------------------------------------------------------------- /open_clip_training/scripts/h14_224_32_finetune.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/scripts/h14_224_32_finetune.sh -------------------------------------------------------------------------------- /open_clip_training/scripts/h14_84_8_pretrain.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/scripts/h14_84_8_pretrain.sh -------------------------------------------------------------------------------- /open_clip_training/setup.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/setup.py -------------------------------------------------------------------------------- /open_clip_training/src/clip_adapter/clip_adapter.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/clip_adapter/clip_adapter.py -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/__init__.py -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/big_vision.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/big_vision.py -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/bpe_simple_vocab_16e6.txt.gz: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/bpe_simple_vocab_16e6.txt.gz -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/coca_model.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/coca_model.py -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/constants.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/constants.py -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/factory.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/factory.py -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/generation_utils.py: -------------------------------------------------------------------------------- 1 | -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/hf_configs.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/hf_configs.py -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/hf_model.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/hf_model.py -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/loss.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/loss.py -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model.py -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/EVA01-g-14-plus.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/EVA01-g-14-plus.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/EVA01-g-14.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/EVA01-g-14.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/EVA02-B-16.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/EVA02-B-16.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/EVA02-E-14-plus.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/EVA02-E-14-plus.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/EVA02-E-14.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/EVA02-E-14.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/EVA02-L-14-336.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/EVA02-L-14-336.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/EVA02-L-14.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/EVA02-L-14.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/RN101-quickgelu.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/RN101-quickgelu.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/RN101.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/RN101.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/RN50-quickgelu.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/RN50-quickgelu.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/RN50.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/RN50.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/RN50x16.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/RN50x16.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/RN50x4.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/RN50x4.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/RN50x64.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/RN50x64.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-B-16-SigLIP-256.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-B-16-SigLIP-256.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-B-16-SigLIP-384.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-B-16-SigLIP-384.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-B-16-SigLIP-512.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-B-16-SigLIP-512.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-B-16-SigLIP-i18n-256.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-B-16-SigLIP-i18n-256.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-B-16-SigLIP.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-B-16-SigLIP.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-B-16-plus-240.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-B-16-plus-240.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-B-16-plus.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-B-16-plus.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-B-16-quickgelu.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-B-16-quickgelu.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-B-16.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-B-16.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-B-32-256.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-B-32-256.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-B-32-plus-256.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-B-32-plus-256.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-B-32-quickgelu.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-B-32-quickgelu.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-B-32.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-B-32.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-H-14-378-quickgelu.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-H-14-378-quickgelu.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-H-14-CLIPA-336.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-H-14-CLIPA-336.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-H-14-CLIPA.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-H-14-CLIPA.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-H-14-quickgelu.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-H-14-quickgelu.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-H-14.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-H-14.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-H-16.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-H-16.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-L-14-280.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-L-14-280.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-L-14-336.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-L-14-336.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-L-14-CLIPA-336.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-L-14-CLIPA-336.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-L-14-CLIPA.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-L-14-CLIPA.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-L-14-quickgelu.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-L-14-quickgelu.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-L-14.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-L-14.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-L-16-320.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-L-16-320.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-L-16-SigLIP-256.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-L-16-SigLIP-256.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-L-16-SigLIP-384.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-L-16-SigLIP-384.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-L-16.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-L-16.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-M-16-alt.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-M-16-alt.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-M-16.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-M-16.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-M-32-alt.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-M-32-alt.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-M-32.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-M-32.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-S-16-alt.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-S-16-alt.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-S-16.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-S-16.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-S-32-alt.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-S-32-alt.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-S-32.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-S-32.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-SO400M-14-SigLIP-384.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-SO400M-14-SigLIP-384.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-SO400M-14-SigLIP.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-SO400M-14-SigLIP.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-bigG-14-CLIPA-336.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-bigG-14-CLIPA-336.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-bigG-14-CLIPA.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-bigG-14-CLIPA.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-bigG-14.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-bigG-14.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-e-14.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-e-14.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/ViT-g-14.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/ViT-g-14.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/coca_ViT-B-32.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/coca_ViT-B-32.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/coca_ViT-L-14.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/coca_ViT-L-14.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/coca_base.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/coca_base.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/coca_roberta-ViT-B-32.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/coca_roberta-ViT-B-32.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/convnext_base.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/convnext_base.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/convnext_base_w.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/convnext_base_w.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/convnext_base_w_320.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/convnext_base_w_320.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/convnext_large.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/convnext_large.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/convnext_large_d.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/convnext_large_d.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/convnext_large_d_320.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/convnext_large_d_320.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/convnext_small.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/convnext_small.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/convnext_tiny.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/convnext_tiny.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/convnext_xlarge.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/convnext_xlarge.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/convnext_xxlarge.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/convnext_xxlarge.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/convnext_xxlarge_320.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/convnext_xxlarge_320.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/mt5-base-ViT-B-32.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/mt5-base-ViT-B-32.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/mt5-xl-ViT-H-14.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/mt5-xl-ViT-H-14.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/nllb-clip-base-siglip.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/nllb-clip-base-siglip.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/nllb-clip-base.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/nllb-clip-base.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/nllb-clip-large-siglip.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/nllb-clip-large-siglip.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/nllb-clip-large.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/nllb-clip-large.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/roberta-ViT-B-32.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/roberta-ViT-B-32.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/swin_base_patch4_window7_224.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/swin_base_patch4_window7_224.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/vit_medium_patch16_gap_256.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/vit_medium_patch16_gap_256.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/vit_relpos_medium_patch16_cls_224.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/vit_relpos_medium_patch16_cls_224.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/xlm-roberta-base-ViT-B-32.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/xlm-roberta-base-ViT-B-32.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/model_configs/xlm-roberta-large-ViT-H-14.json: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/model_configs/xlm-roberta-large-ViT-H-14.json -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/modified_resnet.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/modified_resnet.py -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/openai.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/openai.py -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/pos_embed.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/pos_embed.py -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/pretrained.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/pretrained.py -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/push_to_hf_hub.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/push_to_hf_hub.py -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/timm_model.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/timm_model.py -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/tokenizer.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/tokenizer.py -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/transform.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/transform.py -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/transformer.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/transformer.py -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/utils.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/utils.py -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/version.py: -------------------------------------------------------------------------------- 1 | __version__ = '2.22.0' 2 | -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/zero_shot_classifier.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/zero_shot_classifier.py -------------------------------------------------------------------------------- /open_clip_training/src/open_clip/zero_shot_metadata.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/open_clip/zero_shot_metadata.py -------------------------------------------------------------------------------- /open_clip_training/src/scripts/1cap_finetune_VitL.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/scripts/1cap_finetune_VitL.sh -------------------------------------------------------------------------------- /open_clip_training/src/scripts/finetune_VitL_with_mask.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/scripts/finetune_VitL_with_mask.sh -------------------------------------------------------------------------------- /open_clip_training/src/training/.gitignore: -------------------------------------------------------------------------------- 1 | logs/ 2 | -------------------------------------------------------------------------------- /open_clip_training/src/training/__init__.py: -------------------------------------------------------------------------------- 1 | -------------------------------------------------------------------------------- /open_clip_training/src/training/ade150_zeroshot_data.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/training/ade150_zeroshot_data.py -------------------------------------------------------------------------------- /open_clip_training/src/training/data.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/training/data.py -------------------------------------------------------------------------------- /open_clip_training/src/training/distributed.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/training/distributed.py -------------------------------------------------------------------------------- /open_clip_training/src/training/file_utils.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/training/file_utils.py -------------------------------------------------------------------------------- /open_clip_training/src/training/main.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/training/main.py -------------------------------------------------------------------------------- /open_clip_training/src/training/params.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/training/params.py -------------------------------------------------------------------------------- /open_clip_training/src/training/precision.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/training/precision.py -------------------------------------------------------------------------------- /open_clip_training/src/training/profiler.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/training/profiler.py -------------------------------------------------------------------------------- /open_clip_training/src/training/scheduler.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/training/scheduler.py -------------------------------------------------------------------------------- /open_clip_training/src/training/train.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/training/train.py -------------------------------------------------------------------------------- /open_clip_training/src/training/zero_shot.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/src/training/zero_shot.py -------------------------------------------------------------------------------- /open_clip_training/tests/test_download_pretrained.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/tests/test_download_pretrained.py -------------------------------------------------------------------------------- /open_clip_training/tests/test_hf_model.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/tests/test_hf_model.py -------------------------------------------------------------------------------- /open_clip_training/tests/test_inference.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/tests/test_inference.py -------------------------------------------------------------------------------- /open_clip_training/tests/test_inference_simple.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/tests/test_inference_simple.py -------------------------------------------------------------------------------- /open_clip_training/tests/test_num_shards.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/tests/test_num_shards.py -------------------------------------------------------------------------------- /open_clip_training/tests/test_training_simple.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/tests/test_training_simple.py -------------------------------------------------------------------------------- /open_clip_training/tests/test_wds.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/tests/test_wds.py -------------------------------------------------------------------------------- /open_clip_training/tests/util_test.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/tests/util_test.py -------------------------------------------------------------------------------- /open_clip_training/tutorials/int8_tutorial.ipynb: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/open_clip_training/tutorials/int8_tutorial.ipynb -------------------------------------------------------------------------------- /requirements.txt: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/requirements.txt -------------------------------------------------------------------------------- /scan/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/__init__.py -------------------------------------------------------------------------------- /scan/config.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/config.py -------------------------------------------------------------------------------- /scan/data/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/data/__init__.py -------------------------------------------------------------------------------- /scan/data/augmentations.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/data/augmentations.py -------------------------------------------------------------------------------- /scan/data/build.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/data/build.py -------------------------------------------------------------------------------- /scan/data/dataset_mappers/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/data/dataset_mappers/__init__.py -------------------------------------------------------------------------------- /scan/data/dataset_mappers/mask_former_semantic_dataset_mapper.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/data/dataset_mappers/mask_former_semantic_dataset_mapper.py -------------------------------------------------------------------------------- /scan/data/datasets/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/data/datasets/__init__.py -------------------------------------------------------------------------------- /scan/data/datasets/csv_data.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/data/datasets/csv_data.py -------------------------------------------------------------------------------- /scan/data/datasets/register_ade20k_full.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/data/datasets/register_ade20k_full.py -------------------------------------------------------------------------------- /scan/data/datasets/register_cc3m.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/data/datasets/register_cc3m.py -------------------------------------------------------------------------------- /scan/data/datasets/register_coco_stuff.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/data/datasets/register_coco_stuff.py -------------------------------------------------------------------------------- /scan/data/datasets/register_pascal_context.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/data/datasets/register_pascal_context.py -------------------------------------------------------------------------------- /scan/data/datasets/register_voc_seg.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/data/datasets/register_voc_seg.py -------------------------------------------------------------------------------- /scan/evaluation/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/evaluation/__init__.py -------------------------------------------------------------------------------- /scan/evaluation/generalized_sem_seg_evaluation.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/evaluation/generalized_sem_seg_evaluation.py -------------------------------------------------------------------------------- /scan/frequency.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/frequency.py -------------------------------------------------------------------------------- /scan/maskformer_model.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/maskformer_model.py -------------------------------------------------------------------------------- /scan/modeling/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/__init__.py -------------------------------------------------------------------------------- /scan/modeling/backbone/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/backbone/__init__.py -------------------------------------------------------------------------------- /scan/modeling/backbone/clip_resnet.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/backbone/clip_resnet.py -------------------------------------------------------------------------------- /scan/modeling/backbone/swin.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/backbone/swin.py -------------------------------------------------------------------------------- /scan/modeling/clip_adapter/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/clip_adapter/__init__.py -------------------------------------------------------------------------------- /scan/modeling/clip_adapter/adapter.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/clip_adapter/adapter.py -------------------------------------------------------------------------------- /scan/modeling/clip_adapter/text_template.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/clip_adapter/text_template.py -------------------------------------------------------------------------------- /scan/modeling/clip_adapter/utils.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/clip_adapter/utils.py -------------------------------------------------------------------------------- /scan/modeling/criterion.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/criterion.py -------------------------------------------------------------------------------- /scan/modeling/matcher.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/matcher.py -------------------------------------------------------------------------------- /scan/modeling/meta_arch/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright (c) Facebook, Inc. and its affiliates. 2 | -------------------------------------------------------------------------------- /scan/modeling/meta_arch/mask_former_head.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/meta_arch/mask_former_head.py -------------------------------------------------------------------------------- /scan/modeling/meta_arch/per_pixel_baseline.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/meta_arch/per_pixel_baseline.py -------------------------------------------------------------------------------- /scan/modeling/pixel_decoder/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright (c) Facebook, Inc. and its affiliates. 2 | -------------------------------------------------------------------------------- /scan/modeling/pixel_decoder/fpn.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/pixel_decoder/fpn.py -------------------------------------------------------------------------------- /scan/modeling/pixel_decoder/msdeformattn.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/pixel_decoder/msdeformattn.py -------------------------------------------------------------------------------- /scan/modeling/pixel_decoder/ops/functions/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/pixel_decoder/ops/functions/__init__.py -------------------------------------------------------------------------------- /scan/modeling/pixel_decoder/ops/functions/ms_deform_attn_func.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/pixel_decoder/ops/functions/ms_deform_attn_func.py -------------------------------------------------------------------------------- /scan/modeling/pixel_decoder/ops/make.sh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/pixel_decoder/ops/make.sh -------------------------------------------------------------------------------- /scan/modeling/pixel_decoder/ops/modules/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/pixel_decoder/ops/modules/__init__.py -------------------------------------------------------------------------------- /scan/modeling/pixel_decoder/ops/modules/ms_deform_attn.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/pixel_decoder/ops/modules/ms_deform_attn.py -------------------------------------------------------------------------------- /scan/modeling/pixel_decoder/ops/setup.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/pixel_decoder/ops/setup.py -------------------------------------------------------------------------------- /scan/modeling/pixel_decoder/ops/src/cpu/ms_deform_attn_cpu.cpp: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/pixel_decoder/ops/src/cpu/ms_deform_attn_cpu.cpp -------------------------------------------------------------------------------- /scan/modeling/pixel_decoder/ops/src/cpu/ms_deform_attn_cpu.h: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/pixel_decoder/ops/src/cpu/ms_deform_attn_cpu.h -------------------------------------------------------------------------------- /scan/modeling/pixel_decoder/ops/src/cuda/ms_deform_attn_cuda.cu: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/pixel_decoder/ops/src/cuda/ms_deform_attn_cuda.cu -------------------------------------------------------------------------------- /scan/modeling/pixel_decoder/ops/src/cuda/ms_deform_attn_cuda.h: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/pixel_decoder/ops/src/cuda/ms_deform_attn_cuda.h -------------------------------------------------------------------------------- /scan/modeling/pixel_decoder/ops/src/cuda/ms_deform_im2col_cuda.cuh: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/pixel_decoder/ops/src/cuda/ms_deform_im2col_cuda.cuh -------------------------------------------------------------------------------- /scan/modeling/pixel_decoder/ops/src/ms_deform_attn.h: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/pixel_decoder/ops/src/ms_deform_attn.h -------------------------------------------------------------------------------- /scan/modeling/pixel_decoder/ops/src/vision.cpp: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/pixel_decoder/ops/src/vision.cpp -------------------------------------------------------------------------------- /scan/modeling/pixel_decoder/ops/test.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/pixel_decoder/ops/test.py -------------------------------------------------------------------------------- /scan/modeling/transformer_decoder/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/transformer_decoder/__init__.py -------------------------------------------------------------------------------- /scan/modeling/transformer_decoder/mask2former_transformer_decoder.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/transformer_decoder/mask2former_transformer_decoder.py -------------------------------------------------------------------------------- /scan/modeling/transformer_decoder/maskformer_transformer_decoder.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/transformer_decoder/maskformer_transformer_decoder.py -------------------------------------------------------------------------------- /scan/modeling/transformer_decoder/open_vocab_mask2former_predictor.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/transformer_decoder/open_vocab_mask2former_predictor.py -------------------------------------------------------------------------------- /scan/modeling/transformer_decoder/position_encoding.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/transformer_decoder/position_encoding.py -------------------------------------------------------------------------------- /scan/modeling/transformer_decoder/transformer.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/modeling/transformer_decoder/transformer.py -------------------------------------------------------------------------------- /scan/ovseg_model.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/ovseg_model.py -------------------------------------------------------------------------------- /scan/test_time_augmentation.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/test_time_augmentation.py -------------------------------------------------------------------------------- /scan/utils/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/utils/__init__.py -------------------------------------------------------------------------------- /scan/utils/events.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/utils/events.py -------------------------------------------------------------------------------- /scan/utils/misc.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/utils/misc.py -------------------------------------------------------------------------------- /scan/utils/post_process_utils.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/utils/post_process_utils.py -------------------------------------------------------------------------------- /scan/utils/predictor.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/scan/utils/predictor.py -------------------------------------------------------------------------------- /tools/convert-pretrained-clip-model-to-d2.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/tools/convert-pretrained-clip-model-to-d2.py -------------------------------------------------------------------------------- /tools/convert-pretrained-swin-model-to-d2.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/tools/convert-pretrained-swin-model-to-d2.py -------------------------------------------------------------------------------- /tools/convert-torchvision-to-d2.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/tools/convert-torchvision-to-d2.py -------------------------------------------------------------------------------- /tools/replace_clip.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/tools/replace_clip.py -------------------------------------------------------------------------------- /train_net.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/yongliu20/SCAN/HEAD/train_net.py --------------------------------------------------------------------------------