├── LICENSE ├── MOFA-Video-Hybrid ├── README.md ├── aniportrait │ ├── audio2ldmk.py │ ├── configs │ │ ├── config.yaml │ │ ├── inference_audio.yaml │ │ └── inference_v2.yaml │ └── src │ │ ├── audio_models │ │ ├── mish.py │ │ ├── model.py │ │ ├── pose_model.py │ │ ├── torch_utils.py │ │ └── wav2vec2.py │ │ └── utils │ │ ├── audio_util.py │ │ ├── draw_util.py │ │ ├── face_landmark.py │ │ ├── frame_interpolation.py │ │ ├── mp_models │ │ ├── blaze_face_short_range.tflite │ │ ├── face_landmarker_v2_with_blendshapes.task │ │ └── pose_landmarker_heavy.task │ │ ├── mp_utils.py │ │ ├── pose_util.py │ │ └── util.py ├── ckpt_tree.md ├── expression.mat ├── models │ ├── cmp │ │ ├── experiments │ │ │ ├── rep_learning │ │ │ │ ├── alexnet_yfcc+youtube_voc_16gpu_140k │ │ │ │ │ ├── config.yaml │ │ │ │ │ ├── resume.sh │ │ │ │ │ ├── resume_slurm.sh │ │ │ │ │ ├── train.sh │ │ │ │ │ ├── train_slurm.sh │ │ │ │ │ ├── validate.sh │ │ │ │ │ └── validate_slurm.sh │ │ │ │ ├── alexnet_yfcc_voc_16gpu_70k │ │ │ │ │ ├── config.yaml │ │ │ │ │ ├── resume.sh │ │ │ │ │ ├── resume_slurm.sh │ │ │ │ │ ├── train.sh │ │ │ │ │ ├── train_slurm.sh │ │ │ │ │ ├── validate.sh │ │ │ │ │ └── validate_slurm.sh │ │ │ │ ├── alexnet_yfcc_voc_8gpu_140k │ │ │ │ │ ├── config.yaml │ │ │ │ │ ├── resume.sh │ │ │ │ │ ├── resume_slurm.sh │ │ │ │ │ ├── train.sh │ │ │ │ │ ├── train_slurm.sh │ │ │ │ │ ├── validate.sh │ │ │ │ │ └── validate_slurm.sh │ │ │ │ ├── resnet50_yfcc+youtube+vip+mpii_lip_16gpu_70k │ │ │ │ │ ├── config.yaml │ │ │ │ │ ├── resume.sh │ │ │ │ │ ├── resume_slurm.sh │ │ │ │ │ ├── train.sh │ │ │ │ │ ├── train_slurm.sh │ │ │ │ │ ├── validate.sh │ │ │ │ │ └── validate_slurm.sh │ │ │ │ ├── resnet50_yfcc_coco_16gpu_42k │ │ │ │ │ ├── config.yaml │ │ │ │ │ ├── resume.sh │ │ │ │ │ ├── resume_slurm.sh │ │ │ │ │ ├── train.sh │ │ │ │ │ ├── train_slurm.sh │ │ │ │ │ ├── validate.sh │ │ │ │ │ └── validate_slurm.sh │ │ │ │ └── resnet50_yfcc_voc_16gpu_42k │ │ │ │ │ ├── config.yaml │ │ │ │ │ ├── resume.sh │ │ │ │ │ ├── resume_slurm.sh │ │ │ │ │ ├── train.sh │ │ │ │ │ ├── train_slurm.sh │ │ │ │ │ ├── validate.sh │ │ │ │ │ └── validate_slurm.sh │ │ │ └── semiauto_annot │ │ │ │ └── resnet50_vip+mpii_liteflow │ │ │ │ ├── config.yaml │ │ │ │ ├── resume.sh │ │ │ │ ├── resume_slurm.sh │ │ │ │ ├── train.sh │ │ │ │ ├── train_slurm.sh │ │ │ │ ├── validate.sh │ │ │ │ └── validate_slurm.sh │ │ ├── losses.py │ │ ├── models │ │ │ ├── __init__.py │ │ │ ├── backbone │ │ │ │ ├── __init__.py │ │ │ │ ├── alexnet.py │ │ │ │ └── resnet.py │ │ │ ├── cmp.py │ │ │ ├── modules │ │ │ │ ├── __init__.py │ │ │ │ ├── cmp.py │ │ │ │ ├── decoder.py │ │ │ │ ├── others.py │ │ │ │ ├── shallownet.py │ │ │ │ └── warp.py │ │ │ └── single_stage_model.py │ │ └── utils │ │ │ ├── __init__.py │ │ │ ├── common_utils.py │ │ │ ├── data_utils.py │ │ │ ├── distributed_utils.py │ │ │ ├── flowlib.py │ │ │ ├── scheduler.py │ │ │ └── visualize_utils.py │ ├── controlnet_sdv.py │ ├── ldmk_ctrlnet.py │ ├── occlusion │ │ └── hourglass.py │ ├── softsplat.py │ ├── traj_ctrlnet.py │ └── unet_spatio_temporal_condition_controlnet.py ├── pipeline │ └── pipeline.py ├── requirements.txt ├── run_gradio_audio_driven.py ├── run_gradio_video_driven.py ├── sadtalker_audio2pose │ ├── inference.py │ └── src │ │ ├── audio2exp_models │ │ ├── audio2exp.py │ │ └── networks.py │ │ ├── audio2pose_models │ │ ├── audio2pose.py │ │ ├── audio_encoder.py │ │ ├── cvae.py │ │ ├── discriminator.py │ │ ├── networks.py │ │ └── res_unet.py │ │ ├── config │ │ ├── auido2exp.yaml │ │ ├── auido2pose.yaml │ │ ├── facerender.yaml │ │ ├── facerender_pirender.yaml │ │ ├── facerender_still.yaml │ │ └── similarity_Lm3D_all.mat │ │ ├── face3d │ │ ├── data │ │ │ ├── __init__.py │ │ │ ├── base_dataset.py │ │ │ ├── flist_dataset.py │ │ │ ├── image_folder.py │ │ │ └── template_dataset.py │ │ ├── extract_kp_videos.py │ │ ├── extract_kp_videos_safe.py │ │ ├── models │ │ │ ├── __init__.py │ │ │ ├── arcface_torch │ │ │ │ ├── README.md │ │ │ │ ├── backbones │ │ │ │ │ ├── __init__.py │ │ │ │ │ ├── iresnet.py │ │ │ │ │ ├── iresnet2060.py │ │ │ │ │ └── mobilefacenet.py │ │ │ │ ├── configs │ │ │ │ │ ├── 3millions.py │ │ │ │ │ ├── 3millions_pfc.py │ │ │ │ │ ├── __init__.py │ │ │ │ │ ├── base.py │ │ │ │ │ ├── glint360k_mbf.py │ │ │ │ │ ├── glint360k_r100.py │ │ │ │ │ ├── glint360k_r18.py │ │ │ │ │ ├── glint360k_r34.py │ │ │ │ │ ├── glint360k_r50.py │ │ │ │ │ ├── ms1mv3_mbf.py │ │ │ │ │ ├── ms1mv3_r18.py │ │ │ │ │ ├── ms1mv3_r2060.py │ │ │ │ │ ├── ms1mv3_r34.py │ │ │ │ │ ├── ms1mv3_r50.py │ │ │ │ │ └── speed.py │ │ │ │ ├── dataset.py │ │ │ │ ├── docs │ │ │ │ │ ├── eval.md │ │ │ │ │ ├── install.md │ │ │ │ │ ├── modelzoo.md │ │ │ │ │ └── speed_benchmark.md │ │ │ │ ├── eval │ │ │ │ │ ├── __init__.py │ │ │ │ │ └── verification.py │ │ │ │ ├── eval_ijbc.py │ │ │ │ ├── inference.py │ │ │ │ ├── losses.py │ │ │ │ ├── onnx_helper.py │ │ │ │ ├── onnx_ijbc.py │ │ │ │ ├── partial_fc.py │ │ │ │ ├── requirement.txt │ │ │ │ ├── run.sh │ │ │ │ ├── torch2onnx.py │ │ │ │ ├── train.py │ │ │ │ └── utils │ │ │ │ │ ├── __init__.py │ │ │ │ │ ├── plot.py │ │ │ │ │ ├── utils_amp.py │ │ │ │ │ ├── utils_callbacks.py │ │ │ │ │ ├── utils_config.py │ │ │ │ │ ├── utils_logging.py │ │ │ │ │ └── utils_os.py │ │ │ ├── base_model.py │ │ │ ├── bfm.py │ │ │ ├── facerecon_model.py │ │ │ ├── losses.py │ │ │ ├── networks.py │ │ │ └── template_model.py │ │ ├── options │ │ │ ├── __init__.py │ │ │ ├── base_options.py │ │ │ ├── inference_options.py │ │ │ ├── test_options.py │ │ │ └── train_options.py │ │ ├── util │ │ │ ├── BBRegressorParam_r.mat │ │ │ ├── __init__.py │ │ │ ├── detect_lm68.py │ │ │ ├── generate_list.py │ │ │ ├── html.py │ │ │ ├── load_mats.py │ │ │ ├── nvdiffrast.py │ │ │ ├── preprocess.py │ │ │ ├── skin_mask.py │ │ │ ├── test_mean_face.txt │ │ │ ├── util.py │ │ │ └── visualizer.py │ │ ├── visualize.py │ │ └── visualize_old.py │ │ ├── facerender │ │ ├── animate.py │ │ ├── modules │ │ │ ├── dense_motion.py │ │ │ ├── discriminator.py │ │ │ ├── generator.py │ │ │ ├── keypoint_detector.py │ │ │ ├── make_animation.py │ │ │ ├── mapping.py │ │ │ └── util.py │ │ ├── pirender │ │ │ ├── base_function.py │ │ │ ├── config.py │ │ │ └── face_model.py │ │ ├── pirender_animate.py │ │ ├── pirender_animate_control.py │ │ └── sync_batchnorm │ │ │ ├── __init__.py │ │ │ ├── batchnorm.py │ │ │ ├── comm.py │ │ │ ├── replicate.py │ │ │ └── unittest.py │ │ ├── generate_batch.py │ │ ├── generate_facerender_batch.py │ │ ├── gradio_demo.py │ │ ├── test_audio2coeff.py │ │ └── utils │ │ ├── audio.py │ │ ├── croper.py │ │ ├── face_enhancer.py │ │ ├── flow_util.py │ │ ├── hparams.py │ │ ├── init_path.py │ │ ├── model2safetensor.py │ │ ├── paste_pic.py │ │ ├── preprocess.py │ │ ├── preprocess_fromvideo.py │ │ ├── safetensor_helper.py │ │ ├── text2speech.py │ │ └── videoio.py ├── sadtalker_video2pose │ ├── inference.py │ └── src │ │ ├── audio2exp_models │ │ ├── audio2exp.py │ │ └── networks.py │ │ ├── audio2pose_models │ │ ├── audio2pose.py │ │ ├── audio_encoder.py │ │ ├── cvae.py │ │ ├── discriminator.py │ │ ├── networks.py │ │ └── res_unet.py │ │ ├── config │ │ ├── auido2exp.yaml │ │ ├── auido2pose.yaml │ │ ├── facerender.yaml │ │ ├── facerender_pirender.yaml │ │ ├── facerender_still.yaml │ │ └── similarity_Lm3D_all.mat │ │ ├── face3d │ │ ├── data │ │ │ ├── __init__.py │ │ │ ├── base_dataset.py │ │ │ ├── flist_dataset.py │ │ │ ├── image_folder.py │ │ │ └── template_dataset.py │ │ ├── extract_kp_videos.py │ │ ├── extract_kp_videos_safe.py │ │ ├── models │ │ │ ├── __init__.py │ │ │ ├── arcface_torch │ │ │ │ ├── README.md │ │ │ │ ├── backbones │ │ │ │ │ ├── __init__.py │ │ │ │ │ ├── iresnet.py │ │ │ │ │ ├── iresnet2060.py │ │ │ │ │ └── mobilefacenet.py │ │ │ │ ├── configs │ │ │ │ │ ├── 3millions.py │ │ │ │ │ ├── 3millions_pfc.py │ │ │ │ │ ├── __init__.py │ │ │ │ │ ├── base.py │ │ │ │ │ ├── glint360k_mbf.py │ │ │ │ │ ├── glint360k_r100.py │ │ │ │ │ ├── glint360k_r18.py │ │ │ │ │ ├── glint360k_r34.py │ │ │ │ │ ├── glint360k_r50.py │ │ │ │ │ ├── ms1mv3_mbf.py │ │ │ │ │ ├── ms1mv3_r18.py │ │ │ │ │ ├── ms1mv3_r2060.py │ │ │ │ │ ├── ms1mv3_r34.py │ │ │ │ │ ├── ms1mv3_r50.py │ │ │ │ │ └── speed.py │ │ │ │ ├── dataset.py │ │ │ │ ├── docs │ │ │ │ │ ├── eval.md │ │ │ │ │ ├── install.md │ │ │ │ │ ├── modelzoo.md │ │ │ │ │ └── speed_benchmark.md │ │ │ │ ├── eval │ │ │ │ │ ├── __init__.py │ │ │ │ │ └── verification.py │ │ │ │ ├── eval_ijbc.py │ │ │ │ ├── inference.py │ │ │ │ ├── losses.py │ │ │ │ ├── onnx_helper.py │ │ │ │ ├── onnx_ijbc.py │ │ │ │ ├── partial_fc.py │ │ │ │ ├── requirement.txt │ │ │ │ ├── run.sh │ │ │ │ ├── torch2onnx.py │ │ │ │ ├── train.py │ │ │ │ └── utils │ │ │ │ │ ├── __init__.py │ │ │ │ │ ├── plot.py │ │ │ │ │ ├── utils_amp.py │ │ │ │ │ ├── utils_callbacks.py │ │ │ │ │ ├── utils_config.py │ │ │ │ │ ├── utils_logging.py │ │ │ │ │ └── utils_os.py │ │ │ ├── base_model.py │ │ │ ├── bfm.py │ │ │ ├── facerecon_model.py │ │ │ ├── losses.py │ │ │ ├── networks.py │ │ │ └── template_model.py │ │ ├── options │ │ │ ├── __init__.py │ │ │ ├── base_options.py │ │ │ ├── inference_options.py │ │ │ ├── test_options.py │ │ │ └── train_options.py │ │ ├── util │ │ │ ├── BBRegressorParam_r.mat │ │ │ ├── __init__.py │ │ │ ├── detect_lm68.py │ │ │ ├── generate_list.py │ │ │ ├── html.py │ │ │ ├── load_mats.py │ │ │ ├── nvdiffrast.py │ │ │ ├── preprocess.py │ │ │ ├── skin_mask.py │ │ │ ├── test_mean_face.txt │ │ │ ├── util.py │ │ │ └── visualizer.py │ │ ├── visualize.py │ │ ├── visualize_fromvideo.py │ │ └── visualize_old.py │ │ ├── facerender │ │ ├── animate.py │ │ ├── modules │ │ │ ├── dense_motion.py │ │ │ ├── discriminator.py │ │ │ ├── generator.py │ │ │ ├── keypoint_detector.py │ │ │ ├── make_animation.py │ │ │ ├── mapping.py │ │ │ └── util.py │ │ ├── pirender │ │ │ ├── base_function.py │ │ │ ├── config.py │ │ │ └── face_model.py │ │ ├── pirender_animate.py │ │ ├── pirender_animate_control.py │ │ └── sync_batchnorm │ │ │ ├── __init__.py │ │ │ ├── batchnorm.py │ │ │ ├── comm.py │ │ │ ├── replicate.py │ │ │ └── unittest.py │ │ ├── generate_batch.py │ │ ├── generate_facerender_batch.py │ │ ├── gradio_demo.py │ │ ├── test_audio2coeff.py │ │ └── utils │ │ ├── audio.py │ │ ├── croper.py │ │ ├── face_enhancer.py │ │ ├── flow_util.py │ │ ├── hparams.py │ │ ├── init_path.py │ │ ├── model2safetensor.py │ │ ├── paste_pic.py │ │ ├── preprocess.py │ │ ├── preprocess_fromvideo.py │ │ ├── safetensor_helper.py │ │ ├── text2speech.py │ │ └── videoio.py └── utils │ ├── flow_viz.py │ ├── scheduling_euler_discrete_karras_fix.py │ └── utils.py ├── MOFA-Video-Keypoint ├── .DS_Store ├── README.md ├── aniportrait │ ├── .DS_Store │ ├── audio2ldmk.py │ ├── configs │ │ ├── config.yaml │ │ ├── inference_audio.yaml │ │ └── inference_v2.yaml │ └── src │ │ ├── .DS_Store │ │ ├── audio_models │ │ ├── __pycache__ │ │ │ ├── model.cpython-310.pyc │ │ │ ├── pose_model.cpython-310.pyc │ │ │ ├── torch_utils.cpython-310.pyc │ │ │ └── wav2vec2.cpython-310.pyc │ │ ├── mish.py │ │ ├── model.py │ │ ├── pose_model.py │ │ ├── torch_utils.py │ │ └── wav2vec2.py │ │ └── utils │ │ ├── __pycache__ │ │ ├── audio_util.cpython-310.pyc │ │ ├── face_landmark.cpython-310.pyc │ │ ├── mp_utils.cpython-310.pyc │ │ └── pose_util.cpython-310.pyc │ │ ├── audio_util.py │ │ ├── draw_util.py │ │ ├── face_landmark.py │ │ ├── frame_interpolation.py │ │ ├── mp_models │ │ ├── blaze_face_short_range.tflite │ │ ├── face_landmarker_v2_with_blendshapes.task │ │ └── pose_landmarker_heavy.task │ │ ├── mp_utils.py │ │ ├── pose_util.py │ │ └── util.py ├── demo │ ├── audios │ │ └── 000001.wav │ └── images │ │ └── 000001.jpg ├── expression.mat ├── inference.sh ├── inference_opendomain.py ├── models │ ├── .DS_Store │ ├── __pycache__ │ │ ├── controlnet_sdv.cpython-310.pyc │ │ ├── ldmk_ctrlnet.cpython-310.pyc │ │ ├── softsplat.cpython-310.pyc │ │ └── unet_spatio_temporal_condition_controlnet.cpython-310.pyc │ ├── cmp │ │ ├── .DS_Store │ │ ├── __pycache__ │ │ │ └── losses.cpython-310.pyc │ │ ├── experiments │ │ │ ├── rep_learning │ │ │ │ ├── alexnet_yfcc+youtube_voc_16gpu_140k │ │ │ │ │ ├── config.yaml │ │ │ │ │ ├── resume.sh │ │ │ │ │ ├── resume_slurm.sh │ │ │ │ │ ├── train.sh │ │ │ │ │ ├── train_slurm.sh │ │ │ │ │ ├── validate.sh │ │ │ │ │ └── validate_slurm.sh │ │ │ │ ├── alexnet_yfcc_voc_16gpu_70k │ │ │ │ │ ├── config.yaml │ │ │ │ │ ├── resume.sh │ │ │ │ │ ├── resume_slurm.sh │ │ │ │ │ ├── train.sh │ │ │ │ │ ├── train_slurm.sh │ │ │ │ │ ├── validate.sh │ │ │ │ │ └── validate_slurm.sh │ │ │ │ ├── alexnet_yfcc_voc_8gpu_140k │ │ │ │ │ ├── config.yaml │ │ │ │ │ ├── resume.sh │ │ │ │ │ ├── resume_slurm.sh │ │ │ │ │ ├── train.sh │ │ │ │ │ ├── train_slurm.sh │ │ │ │ │ ├── validate.sh │ │ │ │ │ └── validate_slurm.sh │ │ │ │ ├── resnet50_yfcc+youtube+vip+mpii_lip_16gpu_70k │ │ │ │ │ ├── config.yaml │ │ │ │ │ ├── resume.sh │ │ │ │ │ ├── resume_slurm.sh │ │ │ │ │ ├── train.sh │ │ │ │ │ ├── train_slurm.sh │ │ │ │ │ ├── validate.sh │ │ │ │ │ └── validate_slurm.sh │ │ │ │ ├── resnet50_yfcc_coco_16gpu_42k │ │ │ │ │ ├── config.yaml │ │ │ │ │ ├── resume.sh │ │ │ │ │ ├── resume_slurm.sh │ │ │ │ │ ├── train.sh │ │ │ │ │ ├── train_slurm.sh │ │ │ │ │ ├── validate.sh │ │ │ │ │ └── validate_slurm.sh │ │ │ │ └── resnet50_yfcc_voc_16gpu_42k │ │ │ │ │ ├── config.yaml │ │ │ │ │ ├── resume.sh │ │ │ │ │ ├── resume_slurm.sh │ │ │ │ │ ├── train.sh │ │ │ │ │ ├── train_slurm.sh │ │ │ │ │ ├── validate.sh │ │ │ │ │ └── validate_slurm.sh │ │ │ └── semiauto_annot │ │ │ │ └── resnet50_vip+mpii_liteflow │ │ │ │ ├── config.yaml │ │ │ │ ├── resume.sh │ │ │ │ ├── resume_slurm.sh │ │ │ │ ├── train.sh │ │ │ │ ├── train_slurm.sh │ │ │ │ ├── validate.sh │ │ │ │ └── validate_slurm.sh │ │ ├── losses.py │ │ ├── models │ │ │ ├── __init__.py │ │ │ ├── __pycache__ │ │ │ │ ├── __init__.cpython-310.pyc │ │ │ │ ├── cmp.cpython-310.pyc │ │ │ │ └── single_stage_model.cpython-310.pyc │ │ │ ├── backbone │ │ │ │ ├── __init__.py │ │ │ │ ├── __pycache__ │ │ │ │ │ ├── __init__.cpython-310.pyc │ │ │ │ │ ├── alexnet.cpython-310.pyc │ │ │ │ │ └── resnet.cpython-310.pyc │ │ │ │ ├── alexnet.py │ │ │ │ └── resnet.py │ │ │ ├── cmp.py │ │ │ ├── modules │ │ │ │ ├── __init__.py │ │ │ │ ├── __pycache__ │ │ │ │ │ ├── __init__.cpython-310.pyc │ │ │ │ │ ├── cmp.cpython-310.pyc │ │ │ │ │ ├── decoder.cpython-310.pyc │ │ │ │ │ ├── others.cpython-310.pyc │ │ │ │ │ ├── shallownet.cpython-310.pyc │ │ │ │ │ └── warp.cpython-310.pyc │ │ │ │ ├── cmp.py │ │ │ │ ├── decoder.py │ │ │ │ ├── others.py │ │ │ │ ├── shallownet.py │ │ │ │ └── warp.py │ │ │ └── single_stage_model.py │ │ └── utils │ │ │ ├── __init__.py │ │ │ ├── __pycache__ │ │ │ ├── __init__.cpython-310.pyc │ │ │ ├── common_utils.cpython-310.pyc │ │ │ ├── data_utils.cpython-310.pyc │ │ │ ├── distributed_utils.cpython-310.pyc │ │ │ ├── flowlib.cpython-310.pyc │ │ │ ├── scheduler.cpython-310.pyc │ │ │ └── visualize_utils.cpython-310.pyc │ │ │ ├── common_utils.py │ │ │ ├── data_utils.py │ │ │ ├── distributed_utils.py │ │ │ ├── flowlib.py │ │ │ ├── scheduler.py │ │ │ └── visualize_utils.py │ ├── controlnet_sdv.py │ ├── ldmk_ctrlnet.py │ ├── occlusion │ │ ├── __pycache__ │ │ │ └── hourglass.cpython-310.pyc │ │ └── hourglass.py │ ├── softsplat.py │ ├── traj_ctrlnet.py │ └── unet_spatio_temporal_condition_controlnet.py ├── mofa_keypoint.py ├── pipeline │ ├── __pycache__ │ │ └── svdxt_pipeline_ctrlnet_loop.cpython-310.pyc │ └── svdxt_pipeline_ctrlnet_loop.py ├── requirements.txt ├── sadtalker_audio2pose │ ├── .DS_Store │ ├── inference.py │ └── src │ │ ├── .DS_Store │ │ ├── __pycache__ │ │ ├── generate_batch.cpython-310.pyc │ │ ├── generate_facerender_batch.cpython-310.pyc │ │ └── test_audio2coeff.cpython-310.pyc │ │ ├── audio2exp_models │ │ ├── __pycache__ │ │ │ ├── audio2exp.cpython-310.pyc │ │ │ └── networks.cpython-310.pyc │ │ ├── audio2exp.py │ │ └── networks.py │ │ ├── audio2pose_models │ │ ├── __pycache__ │ │ │ ├── audio2pose.cpython-310.pyc │ │ │ ├── audio_encoder.cpython-310.pyc │ │ │ ├── cvae.cpython-310.pyc │ │ │ ├── discriminator.cpython-310.pyc │ │ │ ├── networks.cpython-310.pyc │ │ │ └── res_unet.cpython-310.pyc │ │ ├── audio2pose.py │ │ ├── audio_encoder.py │ │ ├── cvae.py │ │ ├── discriminator.py │ │ ├── networks.py │ │ └── res_unet.py │ │ ├── config │ │ ├── auido2exp.yaml │ │ ├── auido2pose.yaml │ │ ├── facerender.yaml │ │ ├── facerender_pirender.yaml │ │ ├── facerender_still.yaml │ │ └── similarity_Lm3D_all.mat │ │ ├── face3d │ │ ├── __pycache__ │ │ │ ├── extract_kp_videos_safe.cpython-310.pyc │ │ │ └── visualize.cpython-310.pyc │ │ ├── data │ │ │ ├── __init__.py │ │ │ ├── base_dataset.py │ │ │ ├── flist_dataset.py │ │ │ ├── image_folder.py │ │ │ └── template_dataset.py │ │ ├── extract_kp_videos.py │ │ ├── extract_kp_videos_safe.py │ │ ├── models │ │ │ ├── __init__.py │ │ │ ├── __pycache__ │ │ │ │ ├── __init__.cpython-310.pyc │ │ │ │ ├── base_model.cpython-310.pyc │ │ │ │ ├── bfm.cpython-310.pyc │ │ │ │ ├── facerecon_model.cpython-310.pyc │ │ │ │ ├── losses.cpython-310.pyc │ │ │ │ └── networks.cpython-310.pyc │ │ │ ├── arcface_torch │ │ │ │ ├── README.md │ │ │ │ ├── backbones │ │ │ │ │ ├── __init__.py │ │ │ │ │ ├── __pycache__ │ │ │ │ │ │ ├── __init__.cpython-310.pyc │ │ │ │ │ │ ├── iresnet.cpython-310.pyc │ │ │ │ │ │ └── mobilefacenet.cpython-310.pyc │ │ │ │ │ ├── iresnet.py │ │ │ │ │ ├── iresnet2060.py │ │ │ │ │ └── mobilefacenet.py │ │ │ │ ├── configs │ │ │ │ │ ├── 3millions.py │ │ │ │ │ ├── 3millions_pfc.py │ │ │ │ │ ├── __init__.py │ │ │ │ │ ├── base.py │ │ │ │ │ ├── glint360k_mbf.py │ │ │ │ │ ├── glint360k_r100.py │ │ │ │ │ ├── glint360k_r18.py │ │ │ │ │ ├── glint360k_r34.py │ │ │ │ │ ├── glint360k_r50.py │ │ │ │ │ ├── ms1mv3_mbf.py │ │ │ │ │ ├── ms1mv3_r18.py │ │ │ │ │ ├── ms1mv3_r2060.py │ │ │ │ │ ├── ms1mv3_r34.py │ │ │ │ │ ├── ms1mv3_r50.py │ │ │ │ │ └── speed.py │ │ │ │ ├── dataset.py │ │ │ │ ├── docs │ │ │ │ │ ├── eval.md │ │ │ │ │ ├── install.md │ │ │ │ │ ├── modelzoo.md │ │ │ │ │ └── speed_benchmark.md │ │ │ │ ├── eval │ │ │ │ │ ├── __init__.py │ │ │ │ │ └── verification.py │ │ │ │ ├── eval_ijbc.py │ │ │ │ ├── inference.py │ │ │ │ ├── losses.py │ │ │ │ ├── onnx_helper.py │ │ │ │ ├── onnx_ijbc.py │ │ │ │ ├── partial_fc.py │ │ │ │ ├── requirement.txt │ │ │ │ ├── run.sh │ │ │ │ ├── torch2onnx.py │ │ │ │ ├── train.py │ │ │ │ └── utils │ │ │ │ │ ├── __init__.py │ │ │ │ │ ├── plot.py │ │ │ │ │ ├── utils_amp.py │ │ │ │ │ ├── utils_callbacks.py │ │ │ │ │ ├── utils_config.py │ │ │ │ │ ├── utils_logging.py │ │ │ │ │ └── utils_os.py │ │ │ ├── base_model.py │ │ │ ├── bfm.py │ │ │ ├── facerecon_model.py │ │ │ ├── losses.py │ │ │ ├── networks.py │ │ │ └── template_model.py │ │ ├── options │ │ │ ├── __init__.py │ │ │ ├── base_options.py │ │ │ ├── inference_options.py │ │ │ ├── test_options.py │ │ │ └── train_options.py │ │ ├── util │ │ │ ├── BBRegressorParam_r.mat │ │ │ ├── __init__.py │ │ │ ├── __pycache__ │ │ │ │ ├── __init__.cpython-310.pyc │ │ │ │ ├── load_mats.cpython-310.pyc │ │ │ │ ├── nvdiffrast.cpython-310.pyc │ │ │ │ ├── preprocess.cpython-310.pyc │ │ │ │ └── util.cpython-310.pyc │ │ │ ├── detect_lm68.py │ │ │ ├── generate_list.py │ │ │ ├── html.py │ │ │ ├── load_mats.py │ │ │ ├── nvdiffrast.py │ │ │ ├── preprocess.py │ │ │ ├── skin_mask.py │ │ │ ├── test_mean_face.txt │ │ │ ├── util.py │ │ │ └── visualizer.py │ │ ├── visualize.py │ │ └── visualize_old.py │ │ ├── facerender │ │ ├── __pycache__ │ │ │ ├── animate.cpython-310.pyc │ │ │ └── pirender_animate.cpython-310.pyc │ │ ├── animate.py │ │ ├── modules │ │ │ ├── __pycache__ │ │ │ │ ├── dense_motion.cpython-310.pyc │ │ │ │ ├── generator.cpython-310.pyc │ │ │ │ ├── keypoint_detector.cpython-310.pyc │ │ │ │ ├── make_animation.cpython-310.pyc │ │ │ │ ├── mapping.cpython-310.pyc │ │ │ │ └── util.cpython-310.pyc │ │ │ ├── dense_motion.py │ │ │ ├── discriminator.py │ │ │ ├── generator.py │ │ │ ├── keypoint_detector.py │ │ │ ├── make_animation.py │ │ │ ├── mapping.py │ │ │ └── util.py │ │ ├── pirender │ │ │ ├── __pycache__ │ │ │ │ ├── base_function.cpython-310.pyc │ │ │ │ ├── config.cpython-310.pyc │ │ │ │ └── face_model.cpython-310.pyc │ │ │ ├── base_function.py │ │ │ ├── config.py │ │ │ └── face_model.py │ │ ├── pirender_animate.py │ │ ├── pirender_animate_control.py │ │ └── sync_batchnorm │ │ │ ├── __init__.py │ │ │ ├── __pycache__ │ │ │ ├── __init__.cpython-310.pyc │ │ │ ├── batchnorm.cpython-310.pyc │ │ │ ├── comm.cpython-310.pyc │ │ │ └── replicate.cpython-310.pyc │ │ │ ├── batchnorm.py │ │ │ ├── comm.py │ │ │ ├── replicate.py │ │ │ └── unittest.py │ │ ├── generate_batch.py │ │ ├── generate_facerender_batch.py │ │ ├── gradio_demo.py │ │ ├── test_audio2coeff.py │ │ └── utils │ │ ├── __pycache__ │ │ ├── audio.cpython-310.pyc │ │ ├── croper.cpython-310.pyc │ │ ├── face_enhancer.cpython-310.pyc │ │ ├── flow_util.cpython-310.pyc │ │ ├── hparams.cpython-310.pyc │ │ ├── init_path.cpython-310.pyc │ │ ├── paste_pic.cpython-310.pyc │ │ ├── preprocess.cpython-310.pyc │ │ ├── safetensor_helper.cpython-310.pyc │ │ └── videoio.cpython-310.pyc │ │ ├── audio.py │ │ ├── croper.py │ │ ├── face_enhancer.py │ │ ├── flow_util.py │ │ ├── hparams.py │ │ ├── init_path.py │ │ ├── model2safetensor.py │ │ ├── paste_pic.py │ │ ├── preprocess.py │ │ ├── preprocess_fromvideo.py │ │ ├── safetensor_helper.py │ │ ├── text2speech.py │ │ └── videoio.py └── utils │ ├── __pycache__ │ ├── flow_viz.cpython-310.pyc │ ├── scheduling_euler_discrete_karras_fix.cpython-310.pyc │ └── utils.cpython-310.pyc │ ├── flow_viz.py │ ├── scheduling_euler_discrete_karras_fix.py │ └── utils.py ├── MOFA-Video-Traj ├── README.md ├── models │ ├── cmp │ │ ├── experiments │ │ │ ├── rep_learning │ │ │ │ ├── alexnet_yfcc+youtube_voc_16gpu_140k │ │ │ │ │ ├── config.yaml │ │ │ │ │ ├── resume.sh │ │ │ │ │ ├── resume_slurm.sh │ │ │ │ │ ├── train.sh │ │ │ │ │ ├── train_slurm.sh │ │ │ │ │ ├── validate.sh │ │ │ │ │ └── validate_slurm.sh │ │ │ │ ├── alexnet_yfcc_voc_16gpu_70k │ │ │ │ │ ├── config.yaml │ │ │ │ │ ├── resume.sh │ │ │ │ │ ├── resume_slurm.sh │ │ │ │ │ ├── train.sh │ │ │ │ │ ├── train_slurm.sh │ │ │ │ │ ├── validate.sh │ │ │ │ │ └── validate_slurm.sh │ │ │ │ ├── alexnet_yfcc_voc_8gpu_140k │ │ │ │ │ ├── config.yaml │ │ │ │ │ ├── resume.sh │ │ │ │ │ ├── resume_slurm.sh │ │ │ │ │ ├── train.sh │ │ │ │ │ ├── train_slurm.sh │ │ │ │ │ ├── validate.sh │ │ │ │ │ └── validate_slurm.sh │ │ │ │ ├── resnet50_yfcc+youtube+vip+mpii_lip_16gpu_70k │ │ │ │ │ ├── config.yaml │ │ │ │ │ ├── resume.sh │ │ │ │ │ ├── resume_slurm.sh │ │ │ │ │ ├── train.sh │ │ │ │ │ ├── train_slurm.sh │ │ │ │ │ ├── validate.sh │ │ │ │ │ └── validate_slurm.sh │ │ │ │ ├── resnet50_yfcc_coco_16gpu_42k │ │ │ │ │ ├── config.yaml │ │ │ │ │ ├── resume.sh │ │ │ │ │ ├── resume_slurm.sh │ │ │ │ │ ├── train.sh │ │ │ │ │ ├── train_slurm.sh │ │ │ │ │ ├── validate.sh │ │ │ │ │ └── validate_slurm.sh │ │ │ │ └── resnet50_yfcc_voc_16gpu_42k │ │ │ │ │ ├── config.yaml │ │ │ │ │ ├── resume.sh │ │ │ │ │ ├── resume_slurm.sh │ │ │ │ │ ├── train.sh │ │ │ │ │ ├── train_slurm.sh │ │ │ │ │ ├── validate.sh │ │ │ │ │ └── validate_slurm.sh │ │ │ └── semiauto_annot │ │ │ │ └── resnet50_vip+mpii_liteflow │ │ │ │ ├── config.yaml │ │ │ │ ├── resume.sh │ │ │ │ ├── resume_slurm.sh │ │ │ │ ├── train.sh │ │ │ │ ├── train_slurm.sh │ │ │ │ ├── validate.sh │ │ │ │ └── validate_slurm.sh │ │ ├── losses.py │ │ ├── models │ │ │ ├── __init__.py │ │ │ ├── backbone │ │ │ │ ├── __init__.py │ │ │ │ ├── alexnet.py │ │ │ │ └── resnet.py │ │ │ ├── cmp.py │ │ │ ├── modules │ │ │ │ ├── __init__.py │ │ │ │ ├── cmp.py │ │ │ │ ├── decoder.py │ │ │ │ ├── others.py │ │ │ │ ├── shallownet.py │ │ │ │ └── warp.py │ │ │ └── single_stage_model.py │ │ └── utils │ │ │ ├── __init__.py │ │ │ ├── __pycache__ │ │ │ ├── __init__.cpython-310.pyc │ │ │ ├── __init__.cpython-38.pyc │ │ │ ├── __init__.cpython-39.pyc │ │ │ ├── common_utils.cpython-310.pyc │ │ │ ├── common_utils.cpython-38.pyc │ │ │ ├── common_utils.cpython-39.pyc │ │ │ ├── data_utils.cpython-310.pyc │ │ │ ├── data_utils.cpython-38.pyc │ │ │ ├── data_utils.cpython-39.pyc │ │ │ ├── distributed_utils.cpython-310.pyc │ │ │ ├── distributed_utils.cpython-38.pyc │ │ │ ├── distributed_utils.cpython-39.pyc │ │ │ ├── flowlib.cpython-310.pyc │ │ │ ├── flowlib.cpython-38.pyc │ │ │ ├── flowlib.cpython-39.pyc │ │ │ ├── scheduler.cpython-310.pyc │ │ │ ├── scheduler.cpython-38.pyc │ │ │ ├── scheduler.cpython-39.pyc │ │ │ ├── visualize_utils.cpython-310.pyc │ │ │ ├── visualize_utils.cpython-38.pyc │ │ │ └── visualize_utils.cpython-39.pyc │ │ │ ├── common_utils.py │ │ │ ├── data_utils.py │ │ │ ├── distributed_utils.py │ │ │ ├── flowlib.py │ │ │ ├── scheduler.py │ │ │ └── visualize_utils.py │ ├── controlnet_sdv.py │ ├── softsplat.py │ ├── svdxt_featureflow_forward_controlnet_s2d_fixcmp_norefine.py │ └── unet_spatio_temporal_condition_controlnet.py ├── pipeline │ └── pipeline.py ├── requirements.txt ├── run_gradio.py └── utils │ ├── flow_viz.py │ ├── scheduling_euler_discrete_karras_fix.py │ └── utils.py ├── README.md ├── Training ├── README.md ├── models │ ├── .DS_Store │ ├── cmp │ │ ├── .DS_Store │ │ ├── __pycache__ │ │ │ ├── losses.cpython-310.pyc │ │ │ └── losses.cpython-39.pyc │ │ ├── experiments │ │ │ ├── .DS_Store │ │ │ ├── rep_learning │ │ │ │ ├── alexnet_yfcc+youtube_voc_16gpu_140k │ │ │ │ │ ├── config.yaml │ │ │ │ │ ├── resume.sh │ │ │ │ │ ├── resume_slurm.sh │ │ │ │ │ ├── train.sh │ │ │ │ │ ├── train_slurm.sh │ │ │ │ │ ├── validate.sh │ │ │ │ │ └── validate_slurm.sh │ │ │ │ ├── alexnet_yfcc_voc_16gpu_70k │ │ │ │ │ ├── config.yaml │ │ │ │ │ ├── resume.sh │ │ │ │ │ ├── resume_slurm.sh │ │ │ │ │ ├── train.sh │ │ │ │ │ ├── train_slurm.sh │ │ │ │ │ ├── validate.sh │ │ │ │ │ └── validate_slurm.sh │ │ │ │ ├── alexnet_yfcc_voc_8gpu_140k │ │ │ │ │ ├── config.yaml │ │ │ │ │ ├── resume.sh │ │ │ │ │ ├── resume_slurm.sh │ │ │ │ │ ├── train.sh │ │ │ │ │ ├── train_slurm.sh │ │ │ │ │ ├── validate.sh │ │ │ │ │ └── validate_slurm.sh │ │ │ │ ├── resnet50_yfcc+youtube+vip+mpii_lip_16gpu_70k │ │ │ │ │ ├── config.yaml │ │ │ │ │ ├── resume.sh │ │ │ │ │ ├── resume_slurm.sh │ │ │ │ │ ├── train.sh │ │ │ │ │ ├── train_slurm.sh │ │ │ │ │ ├── validate.sh │ │ │ │ │ └── validate_slurm.sh │ │ │ │ ├── resnet50_yfcc_coco_16gpu_42k │ │ │ │ │ ├── config.yaml │ │ │ │ │ ├── resume.sh │ │ │ │ │ ├── resume_slurm.sh │ │ │ │ │ ├── train.sh │ │ │ │ │ ├── train_slurm.sh │ │ │ │ │ ├── validate.sh │ │ │ │ │ └── validate_slurm.sh │ │ │ │ └── resnet50_yfcc_voc_16gpu_42k │ │ │ │ │ ├── config.yaml │ │ │ │ │ ├── resume.sh │ │ │ │ │ ├── resume_slurm.sh │ │ │ │ │ ├── train.sh │ │ │ │ │ ├── train_slurm.sh │ │ │ │ │ ├── validate.sh │ │ │ │ │ └── validate_slurm.sh │ │ │ └── semiauto_annot │ │ │ │ └── resnet50_vip+mpii_liteflow │ │ │ │ ├── checkpoints │ │ │ │ └── .DS_Store │ │ │ │ ├── config.yaml │ │ │ │ ├── resume.sh │ │ │ │ ├── resume_slurm.sh │ │ │ │ ├── train.sh │ │ │ │ ├── train_slurm.sh │ │ │ │ ├── validate.sh │ │ │ │ └── validate_slurm.sh │ │ ├── losses.py │ │ ├── models │ │ │ ├── __init__.py │ │ │ ├── __pycache__ │ │ │ │ ├── __init__.cpython-310.pyc │ │ │ │ ├── __init__.cpython-39.pyc │ │ │ │ ├── cmp.cpython-310.pyc │ │ │ │ ├── cmp.cpython-39.pyc │ │ │ │ ├── single_stage_model.cpython-310.pyc │ │ │ │ └── single_stage_model.cpython-39.pyc │ │ │ ├── backbone │ │ │ │ ├── __init__.py │ │ │ │ ├── __pycache__ │ │ │ │ │ ├── __init__.cpython-310.pyc │ │ │ │ │ ├── __init__.cpython-39.pyc │ │ │ │ │ ├── alexnet.cpython-310.pyc │ │ │ │ │ ├── alexnet.cpython-39.pyc │ │ │ │ │ ├── resnet.cpython-310.pyc │ │ │ │ │ └── resnet.cpython-39.pyc │ │ │ │ ├── alexnet.py │ │ │ │ └── resnet.py │ │ │ ├── cmp.py │ │ │ ├── modules │ │ │ │ ├── __init__.py │ │ │ │ ├── __pycache__ │ │ │ │ │ ├── __init__.cpython-310.pyc │ │ │ │ │ ├── __init__.cpython-39.pyc │ │ │ │ │ ├── cmp.cpython-310.pyc │ │ │ │ │ ├── cmp.cpython-39.pyc │ │ │ │ │ ├── decoder.cpython-310.pyc │ │ │ │ │ ├── decoder.cpython-39.pyc │ │ │ │ │ ├── others.cpython-310.pyc │ │ │ │ │ ├── others.cpython-39.pyc │ │ │ │ │ ├── shallownet.cpython-310.pyc │ │ │ │ │ ├── shallownet.cpython-39.pyc │ │ │ │ │ ├── warp.cpython-310.pyc │ │ │ │ │ └── warp.cpython-39.pyc │ │ │ │ ├── cmp.py │ │ │ │ ├── decoder.py │ │ │ │ ├── others.py │ │ │ │ ├── shallownet.py │ │ │ │ └── warp.py │ │ │ └── single_stage_model.py │ │ └── utils │ │ │ ├── __init__.py │ │ │ ├── __pycache__ │ │ │ ├── __init__.cpython-310.pyc │ │ │ ├── __init__.cpython-38.pyc │ │ │ ├── __init__.cpython-39.pyc │ │ │ ├── common_utils.cpython-310.pyc │ │ │ ├── common_utils.cpython-38.pyc │ │ │ ├── common_utils.cpython-39.pyc │ │ │ ├── data_utils.cpython-310.pyc │ │ │ ├── data_utils.cpython-38.pyc │ │ │ ├── data_utils.cpython-39.pyc │ │ │ ├── distributed_utils.cpython-310.pyc │ │ │ ├── distributed_utils.cpython-38.pyc │ │ │ ├── distributed_utils.cpython-39.pyc │ │ │ ├── flowlib.cpython-310.pyc │ │ │ ├── flowlib.cpython-38.pyc │ │ │ ├── flowlib.cpython-39.pyc │ │ │ ├── scheduler.cpython-310.pyc │ │ │ ├── scheduler.cpython-38.pyc │ │ │ ├── scheduler.cpython-39.pyc │ │ │ ├── visualize_utils.cpython-310.pyc │ │ │ ├── visualize_utils.cpython-38.pyc │ │ │ └── visualize_utils.cpython-39.pyc │ │ │ ├── common_utils.py │ │ │ ├── data_utils.py │ │ │ ├── distributed_utils.py │ │ │ ├── flowlib.py │ │ │ ├── scheduler.py │ │ │ └── visualize_utils.py │ ├── controlnet_sdv.py │ ├── softsplat.py │ ├── svdxt_featureflow_forward_controlnet_s2d_fixcmp_norefine.py │ └── unet_spatio_temporal_condition_controlnet.py ├── pipeline │ ├── __pycache__ │ │ ├── pipeline.cpython-310.pyc │ │ └── pipeline.cpython-39.pyc │ └── pipeline.py ├── rec_para.txt ├── rec_para_train.txt ├── requirements.txt ├── run_gradio.py ├── train_stage1.py ├── train_stage1.sh ├── train_stage2.py ├── train_stage2.sh ├── train_utils │ ├── .DS_Store │ ├── __pycache__ │ │ ├── dataset.cpython-310.pyc │ │ ├── dataset.cpython-39.pyc │ │ └── sample_flow_utils.cpython-39.pyc │ ├── dataset.py │ ├── sample_flow_utils.py │ └── unimatch │ │ ├── .DS_Store │ │ ├── DATASETS.md │ │ ├── LICENSE │ │ ├── MODEL_ZOO.md │ │ ├── README.md │ │ ├── conda_environment.yml │ │ ├── dataloader │ │ ├── .DS_Store │ │ ├── __init__.py │ │ ├── __pycache__ │ │ │ └── __init__.cpython-310.pyc │ │ ├── flow │ │ │ ├── __init__.py │ │ │ ├── __pycache__ │ │ │ │ ├── __init__.cpython-310.pyc │ │ │ │ ├── datasets.cpython-310.pyc │ │ │ │ └── transforms.cpython-310.pyc │ │ │ ├── chairs_split.txt │ │ │ ├── datasets.py │ │ │ └── transforms.py │ │ └── stereo │ │ │ ├── __init__.py │ │ │ ├── datasets.py │ │ │ └── transforms.py │ │ ├── evaluate_depth.py │ │ ├── evaluate_flow.py │ │ ├── evaluate_stereo.py │ │ ├── loss │ │ ├── __init__.py │ │ ├── __pycache__ │ │ │ ├── __init__.cpython-310.pyc │ │ │ └── flow_loss.cpython-310.pyc │ │ ├── depth_loss.py │ │ ├── flow_loss.py │ │ └── stereo_metric.py │ │ ├── main_depth.py │ │ ├── main_flow.py │ │ ├── main_stereo.py │ │ ├── output │ │ └── gmflow-scale2-regrefine6-davis │ │ │ ├── 00000_flow.png │ │ │ └── 00001_flow.png │ │ ├── pip_install.sh │ │ ├── pretrained │ │ └── .DS_Store │ │ ├── scripts │ │ ├── gmdepth_demo.sh │ │ ├── gmdepth_evaluate.sh │ │ ├── gmdepth_scale1_regrefine1_train.sh │ │ ├── gmdepth_scale1_train.sh │ │ ├── gmflow_demo.sh │ │ ├── gmflow_evaluate.sh │ │ ├── gmflow_scale1_train.sh │ │ ├── gmflow_scale2_regrefine6_train.sh │ │ ├── gmflow_scale2_train.sh │ │ ├── gmflow_submission.sh │ │ ├── gmstereo_demo.sh │ │ ├── gmstereo_evaluate.sh │ │ ├── gmstereo_scale1_train.sh │ │ ├── gmstereo_scale2_regrefine3_train.sh │ │ ├── gmstereo_scale2_train.sh │ │ └── gmstereo_submission.sh │ │ ├── unimatch │ │ ├── __init__.py │ │ ├── __pycache__ │ │ │ ├── __init__.cpython-310.pyc │ │ │ ├── __init__.cpython-311.pyc │ │ │ ├── __init__.cpython-38.pyc │ │ │ ├── __init__.cpython-39.pyc │ │ │ ├── attention.cpython-310.pyc │ │ │ ├── attention.cpython-311.pyc │ │ │ ├── attention.cpython-38.pyc │ │ │ ├── attention.cpython-39.pyc │ │ │ ├── backbone.cpython-310.pyc │ │ │ ├── backbone.cpython-311.pyc │ │ │ ├── backbone.cpython-38.pyc │ │ │ ├── backbone.cpython-39.pyc │ │ │ ├── geometry.cpython-310.pyc │ │ │ ├── geometry.cpython-311.pyc │ │ │ ├── geometry.cpython-38.pyc │ │ │ ├── geometry.cpython-39.pyc │ │ │ ├── matching.cpython-310.pyc │ │ │ ├── matching.cpython-311.pyc │ │ │ ├── matching.cpython-38.pyc │ │ │ ├── matching.cpython-39.pyc │ │ │ ├── position.cpython-310.pyc │ │ │ ├── position.cpython-311.pyc │ │ │ ├── position.cpython-38.pyc │ │ │ ├── position.cpython-39.pyc │ │ │ ├── reg_refine.cpython-310.pyc │ │ │ ├── reg_refine.cpython-311.pyc │ │ │ ├── reg_refine.cpython-38.pyc │ │ │ ├── reg_refine.cpython-39.pyc │ │ │ ├── transformer.cpython-310.pyc │ │ │ ├── transformer.cpython-311.pyc │ │ │ ├── transformer.cpython-38.pyc │ │ │ ├── transformer.cpython-39.pyc │ │ │ ├── trident_conv.cpython-310.pyc │ │ │ ├── trident_conv.cpython-311.pyc │ │ │ ├── trident_conv.cpython-38.pyc │ │ │ ├── trident_conv.cpython-39.pyc │ │ │ ├── unimatch.cpython-310.pyc │ │ │ ├── unimatch.cpython-311.pyc │ │ │ ├── unimatch.cpython-38.pyc │ │ │ ├── unimatch.cpython-39.pyc │ │ │ ├── utils.cpython-310.pyc │ │ │ ├── utils.cpython-311.pyc │ │ │ ├── utils.cpython-38.pyc │ │ │ └── utils.cpython-39.pyc │ │ ├── attention.py │ │ ├── backbone.py │ │ ├── geometry.py │ │ ├── matching.py │ │ ├── position.py │ │ ├── reg_refine.py │ │ ├── transformer.py │ │ ├── trident_conv.py │ │ ├── unimatch.py │ │ └── utils.py │ │ └── utils │ │ ├── __pycache__ │ │ ├── dist_utils.cpython-310.pyc │ │ ├── file_io.cpython-310.pyc │ │ ├── flow_viz.cpython-310.pyc │ │ ├── flow_viz.cpython-311.pyc │ │ ├── flow_viz.cpython-38.pyc │ │ ├── flow_viz.cpython-39.pyc │ │ ├── frame_utils.cpython-310.pyc │ │ ├── logger.cpython-310.pyc │ │ ├── misc.cpython-310.pyc │ │ ├── utils.cpython-310.pyc │ │ └── visualization.cpython-310.pyc │ │ ├── dist_utils.py │ │ ├── file_io.py │ │ ├── flow_viz.py │ │ ├── frame_utils.py │ │ ├── logger.py │ │ ├── misc.py │ │ ├── utils.py │ │ └── visualization.py └── utils │ ├── __pycache__ │ ├── scheduling_euler_discrete_karras_fix.cpython-310.pyc │ └── scheduling_euler_discrete_karras_fix.cpython-39.pyc │ ├── flow_viz.py │ ├── scheduling_euler_discrete_karras_fix.py │ └── utils.py ├── assets ├── images │ ├── README.md │ ├── hybrid_audio.png │ ├── hybrid_video.png │ ├── pipeline.png │ ├── project-mofa.png │ └── traj.png └── teasers │ ├── hybrid │ ├── input.gif │ ├── landmark.gif │ └── output.gif │ ├── ldmk │ ├── flow.gif │ ├── input.gif │ ├── landmark.gif │ ├── output.gif │ └── output_ldmk.gif │ └── traj │ ├── flow.gif │ ├── hint.gif │ ├── input.gif │ └── output.gif └── static └── css └── my_style.css /MOFA-Video-Hybrid/aniportrait/configs/config.yaml: -------------------------------------------------------------------------------- 1 | pretrained_base_model_path: 'ckpts/aniportrait/stable-diffusion-v1-5' 2 | pretrained_vae_path: 'ckpts/aniportrait/sd-vae-ft-mse' 3 | image_encoder_path: 'ckpts/aniportrait/image_encoder' 4 | 5 | denoising_unet_path: "ckpts/aniportrait/denoising_unet.pth" 6 | reference_unet_path: "ckpts/aniportrait/reference_unet.pth" 7 | pose_guider_path: "ckpts/aniportrait/pose_guider.pth" 8 | motion_module_path: "ckpts/aniportrait/motion_module.pth" 9 | 10 | audio_inference_config: "aniportrait/configs/inference_audio.yaml" 11 | inference_config: "aniportrait/configs/inference_v2.yaml" 12 | weight_dtype: 'fp16' 13 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/aniportrait/configs/inference_audio.yaml: -------------------------------------------------------------------------------- 1 | a2m_model: 2 | out_dim: 1404 3 | latent_dim: 512 4 | model_path: ckpts/aniportrait/wav2vec2-base-960h 5 | only_last_fetures: True 6 | from_pretrained: True 7 | 8 | a2p_model: 9 | out_dim: 6 10 | latent_dim: 512 11 | model_path: ckpts/aniportrait/wav2vec2-base-960h 12 | only_last_fetures: True 13 | from_pretrained: True 14 | 15 | pretrained_model: 16 | a2m_ckpt: ckpts/aniportrait/audio2mesh.pt 17 | a2p_ckpt: ckpts/aniportrait/audio2pose.pt 18 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/aniportrait/src/utils/mp_models/blaze_face_short_range.tflite: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Hybrid/aniportrait/src/utils/mp_models/blaze_face_short_range.tflite -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/aniportrait/src/utils/mp_models/face_landmarker_v2_with_blendshapes.task: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Hybrid/aniportrait/src/utils/mp_models/face_landmarker_v2_with_blendshapes.task -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/aniportrait/src/utils/mp_models/pose_landmarker_heavy.task: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Hybrid/aniportrait/src/utils/mp_models/pose_landmarker_heavy.task -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/expression.mat: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Hybrid/expression.mat -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/alexnet_yfcc+youtube_voc_16gpu_140k/resume.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch \ 7 | --load-iter 10000 \ 8 | --resume 9 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/alexnet_yfcc+youtube_voc_16gpu_140k/resume_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm \ 8 | --load-iter 10000 \ 9 | --resume 10 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/alexnet_yfcc+youtube_voc_16gpu_140k/train.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch 7 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/alexnet_yfcc+youtube_voc_16gpu_140k/train_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm 8 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/alexnet_yfcc+youtube_voc_16gpu_140k/validate.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 70000 \ 6 | --validate 7 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/alexnet_yfcc+youtube_voc_16gpu_140k/validate_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py --config $work_path/config.yaml --launcher slurm \ 7 | --load-iter 70000 \ 8 | --validate 9 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_16gpu_70k/resume.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch \ 7 | --load-iter 10000 \ 8 | --resume 9 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_16gpu_70k/resume_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm \ 8 | --load-iter 10000 \ 9 | --resume 10 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_16gpu_70k/train.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch 7 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_16gpu_70k/train_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm 8 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_16gpu_70k/validate.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 70000 \ 6 | --validate 7 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_16gpu_70k/validate_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py --config $work_path/config.yaml --launcher slurm \ 7 | --load-iter 70000 \ 8 | --validate 9 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_8gpu_140k/resume.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 10000 \ 6 | --resume 7 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_8gpu_140k/resume_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm \ 8 | --load-iter 10000 \ 9 | --resume 10 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_8gpu_140k/train.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch 5 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_8gpu_140k/train_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm 8 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_8gpu_140k/validate.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 70000 \ 6 | --validate 7 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_8gpu_140k/validate_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py --config $work_path/config.yaml --launcher slurm \ 7 | --load-iter 70000 \ 8 | --validate 9 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/resnet50_yfcc+youtube+vip+mpii_lip_16gpu_70k/resume.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch \ 7 | --load-iter 10000 \ 8 | --resume 9 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/resnet50_yfcc+youtube+vip+mpii_lip_16gpu_70k/resume_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm \ 8 | --load-iter 10000 \ 9 | --resume 10 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/resnet50_yfcc+youtube+vip+mpii_lip_16gpu_70k/train.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch 7 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/resnet50_yfcc+youtube+vip+mpii_lip_16gpu_70k/train_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm 8 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/resnet50_yfcc+youtube+vip+mpii_lip_16gpu_70k/validate.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 70000 \ 6 | --validate 7 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/resnet50_yfcc+youtube+vip+mpii_lip_16gpu_70k/validate_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py --config $work_path/config.yaml --launcher slurm \ 7 | --load-iter 70000 \ 8 | --validate 9 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/resnet50_yfcc_coco_16gpu_42k/resume.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch \ 7 | --load-iter 10000 \ 8 | --resume 9 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/resnet50_yfcc_coco_16gpu_42k/resume_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm \ 8 | --load-iter 10000 \ 9 | --resume 10 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/resnet50_yfcc_coco_16gpu_42k/train.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch 7 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/resnet50_yfcc_coco_16gpu_42k/train_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm 8 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/resnet50_yfcc_coco_16gpu_42k/validate.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 70000 \ 6 | --validate 7 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/resnet50_yfcc_coco_16gpu_42k/validate_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py --config $work_path/config.yaml --launcher slurm \ 7 | --load-iter 70000 \ 8 | --validate 9 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/resnet50_yfcc_voc_16gpu_42k/resume.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch \ 7 | --load-iter 10000 \ 8 | --resume 9 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/resnet50_yfcc_voc_16gpu_42k/resume_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm \ 8 | --load-iter 10000 \ 9 | --resume 10 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/resnet50_yfcc_voc_16gpu_42k/train.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch 7 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/resnet50_yfcc_voc_16gpu_42k/train_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm 8 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/resnet50_yfcc_voc_16gpu_42k/validate.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 70000 \ 6 | --validate 7 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/rep_learning/resnet50_yfcc_voc_16gpu_42k/validate_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py --config $work_path/config.yaml --launcher slurm \ 7 | --load-iter 70000 \ 8 | --validate 9 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/semiauto_annot/resnet50_vip+mpii_liteflow/resume.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 10000 \ 6 | --resume 7 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/semiauto_annot/resnet50_vip+mpii_liteflow/resume_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm \ 8 | --load-iter 10000 \ 9 | --resume 10 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/semiauto_annot/resnet50_vip+mpii_liteflow/train.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch 5 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/semiauto_annot/resnet50_vip+mpii_liteflow/train_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm 8 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/semiauto_annot/resnet50_vip+mpii_liteflow/validate.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 70000 \ 6 | --validate 7 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/experiments/semiauto_annot/resnet50_vip+mpii_liteflow/validate_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition \ 5 | -n8 --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py --config $work_path/config.yaml --launcher slurm \ 7 | --load-iter 70000 \ 8 | --validate 9 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/models/__init__.py: -------------------------------------------------------------------------------- 1 | from .single_stage_model import * 2 | from .cmp import * 3 | from . import modules 4 | from . import backbone 5 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/models/backbone/__init__.py: -------------------------------------------------------------------------------- 1 | from .resnet import * 2 | from .alexnet import * 3 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/models/modules/__init__.py: -------------------------------------------------------------------------------- 1 | from .warp import * 2 | from .others import * 3 | from .shallownet import * 4 | from .decoder import * 5 | from .cmp import * 6 | 7 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/models/modules/others.py: -------------------------------------------------------------------------------- 1 | import torch.nn as nn 2 | 3 | class FixModule(nn.Module): 4 | 5 | def __init__(self, m): 6 | super(FixModule, self).__init__() 7 | self.module = m 8 | 9 | def forward(self, *args, **kwargs): 10 | return self.module(*args, **kwargs) 11 | 12 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/models/cmp/utils/__init__.py: -------------------------------------------------------------------------------- 1 | from .common_utils import * 2 | from .data_utils import * 3 | from .distributed_utils import * 4 | from .visualize_utils import * 5 | from .scheduler import * 6 | from . import flowlib 7 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/requirements.txt: -------------------------------------------------------------------------------- 1 | diffusers==0.24.0 2 | gradio==4.5.0 3 | scikit-image 4 | torch==2.0.1 5 | torchvision==0.15.2 6 | einops==0.8.0 7 | accelerate==0.30.1 8 | transformers==4.41.1 9 | colorlog==6.8.2 10 | cupy-cuda117==10.6.0 11 | av==12.1.0 12 | gpustat==1.1.1 13 | trimesh==4.4.1 14 | facexlib==0.3.0 15 | omegaconf==2.3.0 16 | librosa==0.10.2.post1 17 | mediapipe==0.10.14 18 | kornia==0.7.2 19 | yacs==0.1.8 20 | gfpgan==1.3.8 21 | numpy==1.23.0 -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_audio2pose/src/config/similarity_Lm3D_all.mat: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Hybrid/sadtalker_audio2pose/src/config/similarity_Lm3D_all.mat -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_audio2pose/src/face3d/models/arcface_torch/configs/3millions.py: -------------------------------------------------------------------------------- 1 | from easydict import EasyDict as edict 2 | 3 | # configs for test speed 4 | 5 | config = edict() 6 | config.loss = "arcface" 7 | config.network = "r50" 8 | config.resume = False 9 | config.output = None 10 | config.embedding_size = 512 11 | config.sample_rate = 1.0 12 | config.fp16 = True 13 | config.momentum = 0.9 14 | config.weight_decay = 5e-4 15 | config.batch_size = 128 16 | config.lr = 0.1 # batch size is 512 17 | 18 | config.rec = "synthetic" 19 | config.num_classes = 300 * 10000 20 | config.num_epoch = 30 21 | config.warmup_epoch = -1 22 | config.decay_epoch = [10, 16, 22] 23 | config.val_targets = [] 24 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_audio2pose/src/face3d/models/arcface_torch/configs/3millions_pfc.py: -------------------------------------------------------------------------------- 1 | from easydict import EasyDict as edict 2 | 3 | # configs for test speed 4 | 5 | config = edict() 6 | config.loss = "arcface" 7 | config.network = "r50" 8 | config.resume = False 9 | config.output = None 10 | config.embedding_size = 512 11 | config.sample_rate = 0.1 12 | config.fp16 = True 13 | config.momentum = 0.9 14 | config.weight_decay = 5e-4 15 | config.batch_size = 128 16 | config.lr = 0.1 # batch size is 512 17 | 18 | config.rec = "synthetic" 19 | config.num_classes = 300 * 10000 20 | config.num_epoch = 30 21 | config.warmup_epoch = -1 22 | config.decay_epoch = [10, 16, 22] 23 | config.val_targets = [] 24 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_audio2pose/src/face3d/models/arcface_torch/configs/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Hybrid/sadtalker_audio2pose/src/face3d/models/arcface_torch/configs/__init__.py -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_audio2pose/src/face3d/models/arcface_torch/configs/speed.py: -------------------------------------------------------------------------------- 1 | from easydict import EasyDict as edict 2 | 3 | # configs for test speed 4 | 5 | config = edict() 6 | config.loss = "arcface" 7 | config.network = "r50" 8 | config.resume = False 9 | config.output = None 10 | config.embedding_size = 512 11 | config.sample_rate = 1.0 12 | config.fp16 = True 13 | config.momentum = 0.9 14 | config.weight_decay = 5e-4 15 | config.batch_size = 128 16 | config.lr = 0.1 # batch size is 512 17 | 18 | config.rec = "synthetic" 19 | config.num_classes = 100 * 10000 20 | config.num_epoch = 30 21 | config.warmup_epoch = -1 22 | config.decay_epoch = [10, 16, 22] 23 | config.val_targets = [] 24 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_audio2pose/src/face3d/models/arcface_torch/docs/modelzoo.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Hybrid/sadtalker_audio2pose/src/face3d/models/arcface_torch/docs/modelzoo.md -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_audio2pose/src/face3d/models/arcface_torch/eval/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Hybrid/sadtalker_audio2pose/src/face3d/models/arcface_torch/eval/__init__.py -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_audio2pose/src/face3d/models/arcface_torch/requirement.txt: -------------------------------------------------------------------------------- 1 | tensorboard 2 | easydict 3 | mxnet 4 | onnx 5 | sklearn 6 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_audio2pose/src/face3d/models/arcface_torch/run.sh: -------------------------------------------------------------------------------- 1 | CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7 python -m torch.distributed.launch --nproc_per_node=8 --nnodes=1 --node_rank=0 --master_addr="127.0.0.1" --master_port=1234 train.py configs/ms1mv3_r50 2 | ps -ef | grep "train" | grep -v grep | awk '{print "kill -9 "$2}' | sh 3 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_audio2pose/src/face3d/models/arcface_torch/utils/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Hybrid/sadtalker_audio2pose/src/face3d/models/arcface_torch/utils/__init__.py -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_audio2pose/src/face3d/models/arcface_torch/utils/utils_config.py: -------------------------------------------------------------------------------- 1 | import importlib 2 | import os.path as osp 3 | 4 | 5 | def get_config(config_file): 6 | assert config_file.startswith('configs/'), 'config file setting must start with configs/' 7 | temp_config_name = osp.basename(config_file) 8 | temp_module_name = osp.splitext(temp_config_name)[0] 9 | config = importlib.import_module("configs.base") 10 | cfg = config.config 11 | config = importlib.import_module("configs.%s" % temp_module_name) 12 | job_cfg = config.config 13 | cfg.update(job_cfg) 14 | if cfg.output is None: 15 | cfg.output = osp.join('work_dirs', temp_module_name) 16 | return cfg -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_audio2pose/src/face3d/models/arcface_torch/utils/utils_os.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Hybrid/sadtalker_audio2pose/src/face3d/models/arcface_torch/utils/utils_os.py -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_audio2pose/src/face3d/options/__init__.py: -------------------------------------------------------------------------------- 1 | """This package options includes option modules: training options, test options, and basic options (used in both training and test).""" 2 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_audio2pose/src/face3d/util/BBRegressorParam_r.mat: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Hybrid/sadtalker_audio2pose/src/face3d/util/BBRegressorParam_r.mat -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_audio2pose/src/face3d/util/__init__.py: -------------------------------------------------------------------------------- 1 | """This package includes a miscellaneous collection of useful helper functions.""" 2 | from src.face3d.util import * 3 | 4 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_audio2pose/src/facerender/sync_batchnorm/__init__.py: -------------------------------------------------------------------------------- 1 | # -*- coding: utf-8 -*- 2 | # File : __init__.py 3 | # Author : Jiayuan Mao 4 | # Email : maojiayuan@gmail.com 5 | # Date : 27/01/2018 6 | # 7 | # This file is part of Synchronized-BatchNorm-PyTorch. 8 | # https://github.com/vacancy/Synchronized-BatchNorm-PyTorch 9 | # Distributed under MIT License. 10 | 11 | from .batchnorm import SynchronizedBatchNorm1d, SynchronizedBatchNorm2d, SynchronizedBatchNorm3d 12 | from .replicate import DataParallelWithCallback, patch_replication_callback 13 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_audio2pose/src/utils/safetensor_helper.py: -------------------------------------------------------------------------------- 1 | 2 | 3 | def load_x_from_safetensor(checkpoint, key): 4 | x_generator = {} 5 | for k,v in checkpoint.items(): 6 | if key in k: 7 | x_generator[k.replace(key+'.', '')] = v 8 | return x_generator -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_audio2pose/src/utils/text2speech.py: -------------------------------------------------------------------------------- 1 | import os 2 | import tempfile 3 | from TTS.api import TTS 4 | 5 | 6 | class TTSTalker(): 7 | def __init__(self) -> None: 8 | model_name = TTS.list_models()[0] 9 | self.tts = TTS(model_name) 10 | 11 | def test(self, text, language='en'): 12 | 13 | tempf = tempfile.NamedTemporaryFile( 14 | delete = False, 15 | suffix = ('.'+'wav'), 16 | ) 17 | 18 | self.tts.tts_to_file(text, speaker=self.tts.speakers[0], language=language, file_path=tempf.name) 19 | 20 | return tempf.name -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_video2pose/src/config/similarity_Lm3D_all.mat: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Hybrid/sadtalker_video2pose/src/config/similarity_Lm3D_all.mat -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_video2pose/src/face3d/models/arcface_torch/configs/3millions.py: -------------------------------------------------------------------------------- 1 | from easydict import EasyDict as edict 2 | 3 | # configs for test speed 4 | 5 | config = edict() 6 | config.loss = "arcface" 7 | config.network = "r50" 8 | config.resume = False 9 | config.output = None 10 | config.embedding_size = 512 11 | config.sample_rate = 1.0 12 | config.fp16 = True 13 | config.momentum = 0.9 14 | config.weight_decay = 5e-4 15 | config.batch_size = 128 16 | config.lr = 0.1 # batch size is 512 17 | 18 | config.rec = "synthetic" 19 | config.num_classes = 300 * 10000 20 | config.num_epoch = 30 21 | config.warmup_epoch = -1 22 | config.decay_epoch = [10, 16, 22] 23 | config.val_targets = [] 24 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_video2pose/src/face3d/models/arcface_torch/configs/3millions_pfc.py: -------------------------------------------------------------------------------- 1 | from easydict import EasyDict as edict 2 | 3 | # configs for test speed 4 | 5 | config = edict() 6 | config.loss = "arcface" 7 | config.network = "r50" 8 | config.resume = False 9 | config.output = None 10 | config.embedding_size = 512 11 | config.sample_rate = 0.1 12 | config.fp16 = True 13 | config.momentum = 0.9 14 | config.weight_decay = 5e-4 15 | config.batch_size = 128 16 | config.lr = 0.1 # batch size is 512 17 | 18 | config.rec = "synthetic" 19 | config.num_classes = 300 * 10000 20 | config.num_epoch = 30 21 | config.warmup_epoch = -1 22 | config.decay_epoch = [10, 16, 22] 23 | config.val_targets = [] 24 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_video2pose/src/face3d/models/arcface_torch/configs/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Hybrid/sadtalker_video2pose/src/face3d/models/arcface_torch/configs/__init__.py -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_video2pose/src/face3d/models/arcface_torch/configs/speed.py: -------------------------------------------------------------------------------- 1 | from easydict import EasyDict as edict 2 | 3 | # configs for test speed 4 | 5 | config = edict() 6 | config.loss = "arcface" 7 | config.network = "r50" 8 | config.resume = False 9 | config.output = None 10 | config.embedding_size = 512 11 | config.sample_rate = 1.0 12 | config.fp16 = True 13 | config.momentum = 0.9 14 | config.weight_decay = 5e-4 15 | config.batch_size = 128 16 | config.lr = 0.1 # batch size is 512 17 | 18 | config.rec = "synthetic" 19 | config.num_classes = 100 * 10000 20 | config.num_epoch = 30 21 | config.warmup_epoch = -1 22 | config.decay_epoch = [10, 16, 22] 23 | config.val_targets = [] 24 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_video2pose/src/face3d/models/arcface_torch/docs/modelzoo.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Hybrid/sadtalker_video2pose/src/face3d/models/arcface_torch/docs/modelzoo.md -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_video2pose/src/face3d/models/arcface_torch/eval/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Hybrid/sadtalker_video2pose/src/face3d/models/arcface_torch/eval/__init__.py -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_video2pose/src/face3d/models/arcface_torch/requirement.txt: -------------------------------------------------------------------------------- 1 | tensorboard 2 | easydict 3 | mxnet 4 | onnx 5 | sklearn 6 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_video2pose/src/face3d/models/arcface_torch/run.sh: -------------------------------------------------------------------------------- 1 | CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7 python -m torch.distributed.launch --nproc_per_node=8 --nnodes=1 --node_rank=0 --master_addr="127.0.0.1" --master_port=1234 train.py configs/ms1mv3_r50 2 | ps -ef | grep "train" | grep -v grep | awk '{print "kill -9 "$2}' | sh 3 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_video2pose/src/face3d/models/arcface_torch/utils/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Hybrid/sadtalker_video2pose/src/face3d/models/arcface_torch/utils/__init__.py -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_video2pose/src/face3d/models/arcface_torch/utils/utils_config.py: -------------------------------------------------------------------------------- 1 | import importlib 2 | import os.path as osp 3 | 4 | 5 | def get_config(config_file): 6 | assert config_file.startswith('configs/'), 'config file setting must start with configs/' 7 | temp_config_name = osp.basename(config_file) 8 | temp_module_name = osp.splitext(temp_config_name)[0] 9 | config = importlib.import_module("configs.base") 10 | cfg = config.config 11 | config = importlib.import_module("configs.%s" % temp_module_name) 12 | job_cfg = config.config 13 | cfg.update(job_cfg) 14 | if cfg.output is None: 15 | cfg.output = osp.join('work_dirs', temp_module_name) 16 | return cfg -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_video2pose/src/face3d/models/arcface_torch/utils/utils_os.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Hybrid/sadtalker_video2pose/src/face3d/models/arcface_torch/utils/utils_os.py -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_video2pose/src/face3d/options/__init__.py: -------------------------------------------------------------------------------- 1 | """This package options includes option modules: training options, test options, and basic options (used in both training and test).""" 2 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_video2pose/src/face3d/util/BBRegressorParam_r.mat: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Hybrid/sadtalker_video2pose/src/face3d/util/BBRegressorParam_r.mat -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_video2pose/src/face3d/util/__init__.py: -------------------------------------------------------------------------------- 1 | """This package includes a miscellaneous collection of useful helper functions.""" 2 | from src.face3d.util import * 3 | 4 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_video2pose/src/facerender/sync_batchnorm/__init__.py: -------------------------------------------------------------------------------- 1 | # -*- coding: utf-8 -*- 2 | # File : __init__.py 3 | # Author : Jiayuan Mao 4 | # Email : maojiayuan@gmail.com 5 | # Date : 27/01/2018 6 | # 7 | # This file is part of Synchronized-BatchNorm-PyTorch. 8 | # https://github.com/vacancy/Synchronized-BatchNorm-PyTorch 9 | # Distributed under MIT License. 10 | 11 | from .batchnorm import SynchronizedBatchNorm1d, SynchronizedBatchNorm2d, SynchronizedBatchNorm3d 12 | from .replicate import DataParallelWithCallback, patch_replication_callback 13 | -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_video2pose/src/utils/safetensor_helper.py: -------------------------------------------------------------------------------- 1 | 2 | 3 | def load_x_from_safetensor(checkpoint, key): 4 | x_generator = {} 5 | for k,v in checkpoint.items(): 6 | if key in k: 7 | x_generator[k.replace(key+'.', '')] = v 8 | return x_generator -------------------------------------------------------------------------------- /MOFA-Video-Hybrid/sadtalker_video2pose/src/utils/text2speech.py: -------------------------------------------------------------------------------- 1 | import os 2 | import tempfile 3 | from TTS.api import TTS 4 | 5 | 6 | class TTSTalker(): 7 | def __init__(self) -> None: 8 | model_name = TTS.list_models()[0] 9 | self.tts = TTS(model_name) 10 | 11 | def test(self, text, language='en'): 12 | 13 | tempf = tempfile.NamedTemporaryFile( 14 | delete = False, 15 | suffix = ('.'+'wav'), 16 | ) 17 | 18 | self.tts.tts_to_file(text, speaker=self.tts.speakers[0], language=language, file_path=tempf.name) 19 | 20 | return tempf.name -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/.DS_Store: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/.DS_Store -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/aniportrait/.DS_Store: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/aniportrait/.DS_Store -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/aniportrait/configs/config.yaml: -------------------------------------------------------------------------------- 1 | pretrained_base_model_path: 'ckpts/aniportrait/stable-diffusion-v1-5' 2 | pretrained_vae_path: 'ckpts/aniportrait/sd-vae-ft-mse' 3 | image_encoder_path: 'ckpts/aniportrait/image_encoder' 4 | 5 | denoising_unet_path: "ckpts/aniportrait/denoising_unet.pth" 6 | reference_unet_path: "ckpts/aniportrait/reference_unet.pth" 7 | pose_guider_path: "ckpts/aniportrait/pose_guider.pth" 8 | motion_module_path: "ckpts/aniportrait/motion_module.pth" 9 | 10 | audio_inference_config: "aniportrait/configs/inference_audio.yaml" 11 | inference_config: "aniportrait/configs/inference_v2.yaml" 12 | weight_dtype: 'fp16' 13 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/aniportrait/configs/inference_audio.yaml: -------------------------------------------------------------------------------- 1 | a2m_model: 2 | out_dim: 1404 3 | latent_dim: 512 4 | model_path: ckpts/aniportrait/wav2vec2-base-960h 5 | only_last_fetures: True 6 | from_pretrained: True 7 | 8 | a2p_model: 9 | out_dim: 6 10 | latent_dim: 512 11 | model_path: ckpts/aniportrait/wav2vec2-base-960h 12 | only_last_fetures: True 13 | from_pretrained: True 14 | 15 | pretrained_model: 16 | a2m_ckpt: ckpts/aniportrait/audio2mesh.pt 17 | a2p_ckpt: ckpts/aniportrait/audio2pose.pt 18 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/aniportrait/src/.DS_Store: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/aniportrait/src/.DS_Store -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/aniportrait/src/audio_models/__pycache__/model.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/aniportrait/src/audio_models/__pycache__/model.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/aniportrait/src/audio_models/__pycache__/pose_model.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/aniportrait/src/audio_models/__pycache__/pose_model.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/aniportrait/src/audio_models/__pycache__/torch_utils.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/aniportrait/src/audio_models/__pycache__/torch_utils.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/aniportrait/src/audio_models/__pycache__/wav2vec2.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/aniportrait/src/audio_models/__pycache__/wav2vec2.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/aniportrait/src/utils/__pycache__/audio_util.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/aniportrait/src/utils/__pycache__/audio_util.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/aniportrait/src/utils/__pycache__/face_landmark.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/aniportrait/src/utils/__pycache__/face_landmark.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/aniportrait/src/utils/__pycache__/mp_utils.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/aniportrait/src/utils/__pycache__/mp_utils.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/aniportrait/src/utils/__pycache__/pose_util.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/aniportrait/src/utils/__pycache__/pose_util.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/aniportrait/src/utils/mp_models/blaze_face_short_range.tflite: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/aniportrait/src/utils/mp_models/blaze_face_short_range.tflite -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/aniportrait/src/utils/mp_models/face_landmarker_v2_with_blendshapes.task: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/aniportrait/src/utils/mp_models/face_landmarker_v2_with_blendshapes.task -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/aniportrait/src/utils/mp_models/pose_landmarker_heavy.task: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/aniportrait/src/utils/mp_models/pose_landmarker_heavy.task -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/demo/audios/000001.wav: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/demo/audios/000001.wav -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/demo/images/000001.jpg: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/demo/images/000001.jpg -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/expression.mat: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/expression.mat -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/inference.sh: -------------------------------------------------------------------------------- 1 | python inference_opendomain.py \ 2 | --audio_path="demo/audios/000001.wav" \ 3 | --img_path="demo/images/000001.jpg" \ 4 | --ckpt_dir="ckpts/mofa/ldmk_controlnet" \ 5 | --save_root="results" \ 6 | --max_frame_len=125 \ 7 | --ldmk_render="aniportrait" -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/.DS_Store: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/models/.DS_Store -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/__pycache__/controlnet_sdv.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/models/__pycache__/controlnet_sdv.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/__pycache__/ldmk_ctrlnet.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/models/__pycache__/ldmk_ctrlnet.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/__pycache__/softsplat.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/models/__pycache__/softsplat.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/__pycache__/unet_spatio_temporal_condition_controlnet.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/models/__pycache__/unet_spatio_temporal_condition_controlnet.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/.DS_Store: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/models/cmp/.DS_Store -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/__pycache__/losses.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/models/cmp/__pycache__/losses.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/alexnet_yfcc+youtube_voc_16gpu_140k/resume.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch \ 7 | --load-iter 10000 \ 8 | --resume 9 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/alexnet_yfcc+youtube_voc_16gpu_140k/resume_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm \ 8 | --load-iter 10000 \ 9 | --resume 10 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/alexnet_yfcc+youtube_voc_16gpu_140k/train.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch 7 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/alexnet_yfcc+youtube_voc_16gpu_140k/train_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm 8 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/alexnet_yfcc+youtube_voc_16gpu_140k/validate.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 70000 \ 6 | --validate 7 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/alexnet_yfcc+youtube_voc_16gpu_140k/validate_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py --config $work_path/config.yaml --launcher slurm \ 7 | --load-iter 70000 \ 8 | --validate 9 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_16gpu_70k/resume.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch \ 7 | --load-iter 10000 \ 8 | --resume 9 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_16gpu_70k/resume_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm \ 8 | --load-iter 10000 \ 9 | --resume 10 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_16gpu_70k/train.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch 7 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_16gpu_70k/train_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm 8 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_16gpu_70k/validate.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 70000 \ 6 | --validate 7 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_16gpu_70k/validate_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py --config $work_path/config.yaml --launcher slurm \ 7 | --load-iter 70000 \ 8 | --validate 9 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_8gpu_140k/resume.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 10000 \ 6 | --resume 7 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_8gpu_140k/resume_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm \ 8 | --load-iter 10000 \ 9 | --resume 10 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_8gpu_140k/train.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch 5 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_8gpu_140k/train_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm 8 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_8gpu_140k/validate.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 70000 \ 6 | --validate 7 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_8gpu_140k/validate_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py --config $work_path/config.yaml --launcher slurm \ 7 | --load-iter 70000 \ 8 | --validate 9 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/resnet50_yfcc+youtube+vip+mpii_lip_16gpu_70k/resume.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch \ 7 | --load-iter 10000 \ 8 | --resume 9 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/resnet50_yfcc+youtube+vip+mpii_lip_16gpu_70k/resume_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm \ 8 | --load-iter 10000 \ 9 | --resume 10 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/resnet50_yfcc+youtube+vip+mpii_lip_16gpu_70k/train.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch 7 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/resnet50_yfcc+youtube+vip+mpii_lip_16gpu_70k/train_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm 8 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/resnet50_yfcc+youtube+vip+mpii_lip_16gpu_70k/validate.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 70000 \ 6 | --validate 7 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/resnet50_yfcc+youtube+vip+mpii_lip_16gpu_70k/validate_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py --config $work_path/config.yaml --launcher slurm \ 7 | --load-iter 70000 \ 8 | --validate 9 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/resnet50_yfcc_coco_16gpu_42k/resume.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch \ 7 | --load-iter 10000 \ 8 | --resume 9 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/resnet50_yfcc_coco_16gpu_42k/resume_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm \ 8 | --load-iter 10000 \ 9 | --resume 10 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/resnet50_yfcc_coco_16gpu_42k/train.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch 7 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/resnet50_yfcc_coco_16gpu_42k/train_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm 8 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/resnet50_yfcc_coco_16gpu_42k/validate.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 70000 \ 6 | --validate 7 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/resnet50_yfcc_coco_16gpu_42k/validate_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py --config $work_path/config.yaml --launcher slurm \ 7 | --load-iter 70000 \ 8 | --validate 9 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/resnet50_yfcc_voc_16gpu_42k/resume.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch \ 7 | --load-iter 10000 \ 8 | --resume 9 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/resnet50_yfcc_voc_16gpu_42k/resume_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm \ 8 | --load-iter 10000 \ 9 | --resume 10 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/resnet50_yfcc_voc_16gpu_42k/train.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch 7 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/resnet50_yfcc_voc_16gpu_42k/train_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm 8 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/resnet50_yfcc_voc_16gpu_42k/validate.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 70000 \ 6 | --validate 7 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/rep_learning/resnet50_yfcc_voc_16gpu_42k/validate_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py --config $work_path/config.yaml --launcher slurm \ 7 | --load-iter 70000 \ 8 | --validate 9 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/semiauto_annot/resnet50_vip+mpii_liteflow/resume.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 10000 \ 6 | --resume 7 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/semiauto_annot/resnet50_vip+mpii_liteflow/resume_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm \ 8 | --load-iter 10000 \ 9 | --resume 10 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/semiauto_annot/resnet50_vip+mpii_liteflow/train.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch 5 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/semiauto_annot/resnet50_vip+mpii_liteflow/train_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm 8 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/semiauto_annot/resnet50_vip+mpii_liteflow/validate.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 70000 \ 6 | --validate 7 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/experiments/semiauto_annot/resnet50_vip+mpii_liteflow/validate_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition \ 5 | -n8 --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py --config $work_path/config.yaml --launcher slurm \ 7 | --load-iter 70000 \ 8 | --validate 9 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/models/__init__.py: -------------------------------------------------------------------------------- 1 | from .single_stage_model import * 2 | from .cmp import * 3 | from . import modules 4 | from . import backbone 5 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/models/__pycache__/__init__.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/models/cmp/models/__pycache__/__init__.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/models/__pycache__/cmp.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/models/cmp/models/__pycache__/cmp.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/models/__pycache__/single_stage_model.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/models/cmp/models/__pycache__/single_stage_model.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/models/backbone/__init__.py: -------------------------------------------------------------------------------- 1 | from .resnet import * 2 | from .alexnet import * 3 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/models/backbone/__pycache__/__init__.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/models/cmp/models/backbone/__pycache__/__init__.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/models/backbone/__pycache__/alexnet.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/models/cmp/models/backbone/__pycache__/alexnet.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/models/backbone/__pycache__/resnet.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/models/cmp/models/backbone/__pycache__/resnet.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/models/modules/__init__.py: -------------------------------------------------------------------------------- 1 | from .warp import * 2 | from .others import * 3 | from .shallownet import * 4 | from .decoder import * 5 | from .cmp import * 6 | 7 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/models/modules/__pycache__/__init__.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/models/cmp/models/modules/__pycache__/__init__.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/models/modules/__pycache__/cmp.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/models/cmp/models/modules/__pycache__/cmp.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/models/modules/__pycache__/decoder.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/models/cmp/models/modules/__pycache__/decoder.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/models/modules/__pycache__/others.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/models/cmp/models/modules/__pycache__/others.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/models/modules/__pycache__/shallownet.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/models/cmp/models/modules/__pycache__/shallownet.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/models/modules/__pycache__/warp.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/models/cmp/models/modules/__pycache__/warp.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/models/modules/others.py: -------------------------------------------------------------------------------- 1 | import torch.nn as nn 2 | 3 | class FixModule(nn.Module): 4 | 5 | def __init__(self, m): 6 | super(FixModule, self).__init__() 7 | self.module = m 8 | 9 | def forward(self, *args, **kwargs): 10 | return self.module(*args, **kwargs) 11 | 12 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/utils/__init__.py: -------------------------------------------------------------------------------- 1 | from .common_utils import * 2 | from .data_utils import * 3 | from .distributed_utils import * 4 | from .visualize_utils import * 5 | from .scheduler import * 6 | from . import flowlib 7 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/utils/__pycache__/__init__.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/models/cmp/utils/__pycache__/__init__.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/utils/__pycache__/common_utils.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/models/cmp/utils/__pycache__/common_utils.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/utils/__pycache__/data_utils.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/models/cmp/utils/__pycache__/data_utils.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/utils/__pycache__/distributed_utils.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/models/cmp/utils/__pycache__/distributed_utils.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/utils/__pycache__/flowlib.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/models/cmp/utils/__pycache__/flowlib.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/utils/__pycache__/scheduler.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/models/cmp/utils/__pycache__/scheduler.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/cmp/utils/__pycache__/visualize_utils.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/models/cmp/utils/__pycache__/visualize_utils.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/models/occlusion/__pycache__/hourglass.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/models/occlusion/__pycache__/hourglass.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/pipeline/__pycache__/svdxt_pipeline_ctrlnet_loop.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/pipeline/__pycache__/svdxt_pipeline_ctrlnet_loop.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/requirements.txt: -------------------------------------------------------------------------------- 1 | diffusers==0.24.0 2 | gradio==4.5.0 3 | opencv-python 4 | opencv-python-headless 5 | scikit-image 6 | torch==2.0.1 7 | torchvision==0.15.2 8 | einops 9 | accelerate==0.30.1 10 | transformers==4.41.1 11 | colorlog 12 | cupy-cuda117 13 | av 14 | gpustat 15 | ffmpeg-python 16 | omegaconf 17 | librosa 18 | mediapipe 19 | kornia 20 | facexlib 21 | yacs 22 | gfpgan 23 | trimesh 24 | numpy==1.23.0 25 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/.DS_Store: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/.DS_Store -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/.DS_Store: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/.DS_Store -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/__pycache__/generate_batch.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/__pycache__/generate_batch.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/__pycache__/generate_facerender_batch.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/__pycache__/generate_facerender_batch.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/__pycache__/test_audio2coeff.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/__pycache__/test_audio2coeff.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/audio2exp_models/__pycache__/audio2exp.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/audio2exp_models/__pycache__/audio2exp.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/audio2exp_models/__pycache__/networks.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/audio2exp_models/__pycache__/networks.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/audio2pose_models/__pycache__/audio2pose.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/audio2pose_models/__pycache__/audio2pose.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/audio2pose_models/__pycache__/audio_encoder.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/audio2pose_models/__pycache__/audio_encoder.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/audio2pose_models/__pycache__/cvae.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/audio2pose_models/__pycache__/cvae.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/audio2pose_models/__pycache__/discriminator.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/audio2pose_models/__pycache__/discriminator.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/audio2pose_models/__pycache__/networks.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/audio2pose_models/__pycache__/networks.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/audio2pose_models/__pycache__/res_unet.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/audio2pose_models/__pycache__/res_unet.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/config/similarity_Lm3D_all.mat: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/config/similarity_Lm3D_all.mat -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/__pycache__/extract_kp_videos_safe.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/__pycache__/extract_kp_videos_safe.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/__pycache__/visualize.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/__pycache__/visualize.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/__pycache__/__init__.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/__pycache__/__init__.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/__pycache__/base_model.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/__pycache__/base_model.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/__pycache__/bfm.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/__pycache__/bfm.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/__pycache__/facerecon_model.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/__pycache__/facerecon_model.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/__pycache__/losses.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/__pycache__/losses.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/__pycache__/networks.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/__pycache__/networks.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/arcface_torch/backbones/__pycache__/__init__.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/arcface_torch/backbones/__pycache__/__init__.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/arcface_torch/backbones/__pycache__/iresnet.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/arcface_torch/backbones/__pycache__/iresnet.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/arcface_torch/backbones/__pycache__/mobilefacenet.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/arcface_torch/backbones/__pycache__/mobilefacenet.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/arcface_torch/configs/3millions.py: -------------------------------------------------------------------------------- 1 | from easydict import EasyDict as edict 2 | 3 | # configs for test speed 4 | 5 | config = edict() 6 | config.loss = "arcface" 7 | config.network = "r50" 8 | config.resume = False 9 | config.output = None 10 | config.embedding_size = 512 11 | config.sample_rate = 1.0 12 | config.fp16 = True 13 | config.momentum = 0.9 14 | config.weight_decay = 5e-4 15 | config.batch_size = 128 16 | config.lr = 0.1 # batch size is 512 17 | 18 | config.rec = "synthetic" 19 | config.num_classes = 300 * 10000 20 | config.num_epoch = 30 21 | config.warmup_epoch = -1 22 | config.decay_epoch = [10, 16, 22] 23 | config.val_targets = [] 24 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/arcface_torch/configs/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/arcface_torch/configs/__init__.py -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/arcface_torch/configs/speed.py: -------------------------------------------------------------------------------- 1 | from easydict import EasyDict as edict 2 | 3 | # configs for test speed 4 | 5 | config = edict() 6 | config.loss = "arcface" 7 | config.network = "r50" 8 | config.resume = False 9 | config.output = None 10 | config.embedding_size = 512 11 | config.sample_rate = 1.0 12 | config.fp16 = True 13 | config.momentum = 0.9 14 | config.weight_decay = 5e-4 15 | config.batch_size = 128 16 | config.lr = 0.1 # batch size is 512 17 | 18 | config.rec = "synthetic" 19 | config.num_classes = 100 * 10000 20 | config.num_epoch = 30 21 | config.warmup_epoch = -1 22 | config.decay_epoch = [10, 16, 22] 23 | config.val_targets = [] 24 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/arcface_torch/docs/modelzoo.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/arcface_torch/docs/modelzoo.md -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/arcface_torch/eval/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/arcface_torch/eval/__init__.py -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/arcface_torch/requirement.txt: -------------------------------------------------------------------------------- 1 | tensorboard 2 | easydict 3 | mxnet 4 | onnx 5 | sklearn 6 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/arcface_torch/run.sh: -------------------------------------------------------------------------------- 1 | CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7 python -m torch.distributed.launch --nproc_per_node=8 --nnodes=1 --node_rank=0 --master_addr="127.0.0.1" --master_port=1234 train.py configs/ms1mv3_r50 2 | ps -ef | grep "train" | grep -v grep | awk '{print "kill -9 "$2}' | sh 3 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/arcface_torch/utils/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/arcface_torch/utils/__init__.py -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/arcface_torch/utils/utils_os.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/models/arcface_torch/utils/utils_os.py -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/options/__init__.py: -------------------------------------------------------------------------------- 1 | """This package options includes option modules: training options, test options, and basic options (used in both training and test).""" 2 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/util/BBRegressorParam_r.mat: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/util/BBRegressorParam_r.mat -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/util/__init__.py: -------------------------------------------------------------------------------- 1 | """This package includes a miscellaneous collection of useful helper functions.""" 2 | from src.face3d.util import * 3 | 4 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/util/__pycache__/__init__.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/util/__pycache__/__init__.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/util/__pycache__/load_mats.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/util/__pycache__/load_mats.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/util/__pycache__/nvdiffrast.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/util/__pycache__/nvdiffrast.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/util/__pycache__/preprocess.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/util/__pycache__/preprocess.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/util/__pycache__/util.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/face3d/util/__pycache__/util.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/__pycache__/animate.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/__pycache__/animate.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/__pycache__/pirender_animate.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/__pycache__/pirender_animate.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/modules/__pycache__/dense_motion.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/modules/__pycache__/dense_motion.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/modules/__pycache__/generator.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/modules/__pycache__/generator.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/modules/__pycache__/keypoint_detector.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/modules/__pycache__/keypoint_detector.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/modules/__pycache__/make_animation.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/modules/__pycache__/make_animation.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/modules/__pycache__/mapping.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/modules/__pycache__/mapping.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/modules/__pycache__/util.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/modules/__pycache__/util.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/pirender/__pycache__/base_function.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/pirender/__pycache__/base_function.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/pirender/__pycache__/config.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/pirender/__pycache__/config.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/pirender/__pycache__/face_model.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/pirender/__pycache__/face_model.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/sync_batchnorm/__init__.py: -------------------------------------------------------------------------------- 1 | # -*- coding: utf-8 -*- 2 | # File : __init__.py 3 | # Author : Jiayuan Mao 4 | # Email : maojiayuan@gmail.com 5 | # Date : 27/01/2018 6 | # 7 | # This file is part of Synchronized-BatchNorm-PyTorch. 8 | # https://github.com/vacancy/Synchronized-BatchNorm-PyTorch 9 | # Distributed under MIT License. 10 | 11 | from .batchnorm import SynchronizedBatchNorm1d, SynchronizedBatchNorm2d, SynchronizedBatchNorm3d 12 | from .replicate import DataParallelWithCallback, patch_replication_callback 13 | -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/sync_batchnorm/__pycache__/__init__.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/sync_batchnorm/__pycache__/__init__.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/sync_batchnorm/__pycache__/batchnorm.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/sync_batchnorm/__pycache__/batchnorm.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/sync_batchnorm/__pycache__/comm.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/sync_batchnorm/__pycache__/comm.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/sync_batchnorm/__pycache__/replicate.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/facerender/sync_batchnorm/__pycache__/replicate.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/utils/__pycache__/audio.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/utils/__pycache__/audio.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/utils/__pycache__/croper.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/utils/__pycache__/croper.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/utils/__pycache__/face_enhancer.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/utils/__pycache__/face_enhancer.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/utils/__pycache__/flow_util.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/utils/__pycache__/flow_util.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/utils/__pycache__/hparams.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/utils/__pycache__/hparams.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/utils/__pycache__/init_path.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/utils/__pycache__/init_path.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/utils/__pycache__/paste_pic.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/utils/__pycache__/paste_pic.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/utils/__pycache__/preprocess.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/utils/__pycache__/preprocess.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/utils/__pycache__/safetensor_helper.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/utils/__pycache__/safetensor_helper.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/utils/__pycache__/videoio.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/sadtalker_audio2pose/src/utils/__pycache__/videoio.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/utils/safetensor_helper.py: -------------------------------------------------------------------------------- 1 | 2 | 3 | def load_x_from_safetensor(checkpoint, key): 4 | x_generator = {} 5 | for k,v in checkpoint.items(): 6 | if key in k: 7 | x_generator[k.replace(key+'.', '')] = v 8 | return x_generator -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/sadtalker_audio2pose/src/utils/text2speech.py: -------------------------------------------------------------------------------- 1 | import os 2 | import tempfile 3 | from TTS.api import TTS 4 | 5 | 6 | class TTSTalker(): 7 | def __init__(self) -> None: 8 | model_name = TTS.list_models()[0] 9 | self.tts = TTS(model_name) 10 | 11 | def test(self, text, language='en'): 12 | 13 | tempf = tempfile.NamedTemporaryFile( 14 | delete = False, 15 | suffix = ('.'+'wav'), 16 | ) 17 | 18 | self.tts.tts_to_file(text, speaker=self.tts.speakers[0], language=language, file_path=tempf.name) 19 | 20 | return tempf.name -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/utils/__pycache__/flow_viz.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/utils/__pycache__/flow_viz.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/utils/__pycache__/scheduling_euler_discrete_karras_fix.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/utils/__pycache__/scheduling_euler_discrete_karras_fix.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Keypoint/utils/__pycache__/utils.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Keypoint/utils/__pycache__/utils.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/alexnet_yfcc+youtube_voc_16gpu_140k/resume.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch \ 7 | --load-iter 10000 \ 8 | --resume 9 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/alexnet_yfcc+youtube_voc_16gpu_140k/resume_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm \ 8 | --load-iter 10000 \ 9 | --resume 10 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/alexnet_yfcc+youtube_voc_16gpu_140k/train.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch 7 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/alexnet_yfcc+youtube_voc_16gpu_140k/train_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm 8 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/alexnet_yfcc+youtube_voc_16gpu_140k/validate.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 70000 \ 6 | --validate 7 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/alexnet_yfcc+youtube_voc_16gpu_140k/validate_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py --config $work_path/config.yaml --launcher slurm \ 7 | --load-iter 70000 \ 8 | --validate 9 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_16gpu_70k/resume.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch \ 7 | --load-iter 10000 \ 8 | --resume 9 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_16gpu_70k/resume_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm \ 8 | --load-iter 10000 \ 9 | --resume 10 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_16gpu_70k/train.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch 7 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_16gpu_70k/train_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm 8 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_16gpu_70k/validate.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 70000 \ 6 | --validate 7 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_16gpu_70k/validate_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py --config $work_path/config.yaml --launcher slurm \ 7 | --load-iter 70000 \ 8 | --validate 9 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_8gpu_140k/resume.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 10000 \ 6 | --resume 7 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_8gpu_140k/resume_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm \ 8 | --load-iter 10000 \ 9 | --resume 10 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_8gpu_140k/train.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch 5 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_8gpu_140k/train_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm 8 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_8gpu_140k/validate.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 70000 \ 6 | --validate 7 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_8gpu_140k/validate_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py --config $work_path/config.yaml --launcher slurm \ 7 | --load-iter 70000 \ 8 | --validate 9 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/resnet50_yfcc+youtube+vip+mpii_lip_16gpu_70k/resume.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch \ 7 | --load-iter 10000 \ 8 | --resume 9 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/resnet50_yfcc+youtube+vip+mpii_lip_16gpu_70k/resume_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm \ 8 | --load-iter 10000 \ 9 | --resume 10 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/resnet50_yfcc+youtube+vip+mpii_lip_16gpu_70k/train.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch 7 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/resnet50_yfcc+youtube+vip+mpii_lip_16gpu_70k/train_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm 8 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/resnet50_yfcc+youtube+vip+mpii_lip_16gpu_70k/validate.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 70000 \ 6 | --validate 7 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/resnet50_yfcc+youtube+vip+mpii_lip_16gpu_70k/validate_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py --config $work_path/config.yaml --launcher slurm \ 7 | --load-iter 70000 \ 8 | --validate 9 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/resnet50_yfcc_coco_16gpu_42k/resume.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch \ 7 | --load-iter 10000 \ 8 | --resume 9 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/resnet50_yfcc_coco_16gpu_42k/resume_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm \ 8 | --load-iter 10000 \ 9 | --resume 10 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/resnet50_yfcc_coco_16gpu_42k/train.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch 7 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/resnet50_yfcc_coco_16gpu_42k/train_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm 8 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/resnet50_yfcc_coco_16gpu_42k/validate.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 70000 \ 6 | --validate 7 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/resnet50_yfcc_coco_16gpu_42k/validate_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py --config $work_path/config.yaml --launcher slurm \ 7 | --load-iter 70000 \ 8 | --validate 9 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/resnet50_yfcc_voc_16gpu_42k/resume.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch \ 7 | --load-iter 10000 \ 8 | --resume 9 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/resnet50_yfcc_voc_16gpu_42k/resume_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm \ 8 | --load-iter 10000 \ 9 | --resume 10 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/resnet50_yfcc_voc_16gpu_42k/train.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch 7 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/resnet50_yfcc_voc_16gpu_42k/train_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm 8 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/resnet50_yfcc_voc_16gpu_42k/validate.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 70000 \ 6 | --validate 7 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/rep_learning/resnet50_yfcc_voc_16gpu_42k/validate_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py --config $work_path/config.yaml --launcher slurm \ 7 | --load-iter 70000 \ 8 | --validate 9 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/semiauto_annot/resnet50_vip+mpii_liteflow/resume.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 10000 \ 6 | --resume 7 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/semiauto_annot/resnet50_vip+mpii_liteflow/resume_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm \ 8 | --load-iter 10000 \ 9 | --resume 10 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/semiauto_annot/resnet50_vip+mpii_liteflow/train.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch 5 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/semiauto_annot/resnet50_vip+mpii_liteflow/train_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm 8 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/semiauto_annot/resnet50_vip+mpii_liteflow/validate.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 70000 \ 6 | --validate 7 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/experiments/semiauto_annot/resnet50_vip+mpii_liteflow/validate_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition \ 5 | -n8 --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py --config $work_path/config.yaml --launcher slurm \ 7 | --load-iter 70000 \ 8 | --validate 9 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/models/__init__.py: -------------------------------------------------------------------------------- 1 | from .single_stage_model import * 2 | from .cmp import * 3 | from . import modules 4 | from . import backbone 5 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/models/backbone/__init__.py: -------------------------------------------------------------------------------- 1 | from .resnet import * 2 | from .alexnet import * 3 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/models/modules/__init__.py: -------------------------------------------------------------------------------- 1 | from .warp import * 2 | from .others import * 3 | from .shallownet import * 4 | from .decoder import * 5 | from .cmp import * 6 | 7 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/models/modules/others.py: -------------------------------------------------------------------------------- 1 | import torch.nn as nn 2 | 3 | class FixModule(nn.Module): 4 | 5 | def __init__(self, m): 6 | super(FixModule, self).__init__() 7 | self.module = m 8 | 9 | def forward(self, *args, **kwargs): 10 | return self.module(*args, **kwargs) 11 | 12 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/utils/__init__.py: -------------------------------------------------------------------------------- 1 | from .common_utils import * 2 | from .data_utils import * 3 | from .distributed_utils import * 4 | from .visualize_utils import * 5 | from .scheduler import * 6 | from . import flowlib 7 | -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/utils/__pycache__/__init__.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Traj/models/cmp/utils/__pycache__/__init__.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/utils/__pycache__/__init__.cpython-38.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Traj/models/cmp/utils/__pycache__/__init__.cpython-38.pyc -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/utils/__pycache__/__init__.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Traj/models/cmp/utils/__pycache__/__init__.cpython-39.pyc -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/utils/__pycache__/common_utils.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Traj/models/cmp/utils/__pycache__/common_utils.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/utils/__pycache__/common_utils.cpython-38.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Traj/models/cmp/utils/__pycache__/common_utils.cpython-38.pyc -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/utils/__pycache__/common_utils.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Traj/models/cmp/utils/__pycache__/common_utils.cpython-39.pyc -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/utils/__pycache__/data_utils.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Traj/models/cmp/utils/__pycache__/data_utils.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/utils/__pycache__/data_utils.cpython-38.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Traj/models/cmp/utils/__pycache__/data_utils.cpython-38.pyc -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/utils/__pycache__/data_utils.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Traj/models/cmp/utils/__pycache__/data_utils.cpython-39.pyc -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/utils/__pycache__/distributed_utils.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Traj/models/cmp/utils/__pycache__/distributed_utils.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/utils/__pycache__/distributed_utils.cpython-38.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Traj/models/cmp/utils/__pycache__/distributed_utils.cpython-38.pyc -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/utils/__pycache__/distributed_utils.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Traj/models/cmp/utils/__pycache__/distributed_utils.cpython-39.pyc -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/utils/__pycache__/flowlib.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Traj/models/cmp/utils/__pycache__/flowlib.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/utils/__pycache__/flowlib.cpython-38.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Traj/models/cmp/utils/__pycache__/flowlib.cpython-38.pyc -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/utils/__pycache__/flowlib.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Traj/models/cmp/utils/__pycache__/flowlib.cpython-39.pyc -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/utils/__pycache__/scheduler.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Traj/models/cmp/utils/__pycache__/scheduler.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/utils/__pycache__/scheduler.cpython-38.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Traj/models/cmp/utils/__pycache__/scheduler.cpython-38.pyc -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/utils/__pycache__/scheduler.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Traj/models/cmp/utils/__pycache__/scheduler.cpython-39.pyc -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/utils/__pycache__/visualize_utils.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Traj/models/cmp/utils/__pycache__/visualize_utils.cpython-310.pyc -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/utils/__pycache__/visualize_utils.cpython-38.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Traj/models/cmp/utils/__pycache__/visualize_utils.cpython-38.pyc -------------------------------------------------------------------------------- /MOFA-Video-Traj/models/cmp/utils/__pycache__/visualize_utils.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/MOFA-Video-Traj/models/cmp/utils/__pycache__/visualize_utils.cpython-39.pyc -------------------------------------------------------------------------------- /MOFA-Video-Traj/requirements.txt: -------------------------------------------------------------------------------- 1 | diffusers==0.24.0 2 | gradio==4.5.0 3 | opencv-python 4 | opencv-python-headless 5 | scikit-image 6 | torch==2.0.1 7 | torchvision==0.15.2 8 | einops 9 | accelerate==0.30.1 10 | transformers==4.41.1 11 | colorlog 12 | cupy-cuda117 13 | av 14 | gpustat 15 | 16 | 17 | # xformers==0.0.22 18 | # decord 19 | # bitsandbytes 20 | 21 | -------------------------------------------------------------------------------- /Training/models/.DS_Store: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/.DS_Store -------------------------------------------------------------------------------- /Training/models/cmp/.DS_Store: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/.DS_Store -------------------------------------------------------------------------------- /Training/models/cmp/__pycache__/losses.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/__pycache__/losses.cpython-310.pyc -------------------------------------------------------------------------------- /Training/models/cmp/__pycache__/losses.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/__pycache__/losses.cpython-39.pyc -------------------------------------------------------------------------------- /Training/models/cmp/experiments/.DS_Store: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/experiments/.DS_Store -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/alexnet_yfcc+youtube_voc_16gpu_140k/resume.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch \ 7 | --load-iter 10000 \ 8 | --resume 9 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/alexnet_yfcc+youtube_voc_16gpu_140k/resume_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm \ 8 | --load-iter 10000 \ 9 | --resume 10 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/alexnet_yfcc+youtube_voc_16gpu_140k/train.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch 7 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/alexnet_yfcc+youtube_voc_16gpu_140k/train_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm 8 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/alexnet_yfcc+youtube_voc_16gpu_140k/validate.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 70000 \ 6 | --validate 7 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/alexnet_yfcc+youtube_voc_16gpu_140k/validate_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py --config $work_path/config.yaml --launcher slurm \ 7 | --load-iter 70000 \ 8 | --validate 9 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_16gpu_70k/resume.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch \ 7 | --load-iter 10000 \ 8 | --resume 9 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_16gpu_70k/resume_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm \ 8 | --load-iter 10000 \ 9 | --resume 10 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_16gpu_70k/train.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch 7 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_16gpu_70k/train_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm 8 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_16gpu_70k/validate.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 70000 \ 6 | --validate 7 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_16gpu_70k/validate_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py --config $work_path/config.yaml --launcher slurm \ 7 | --load-iter 70000 \ 8 | --validate 9 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_8gpu_140k/resume.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 10000 \ 6 | --resume 7 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_8gpu_140k/resume_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm \ 8 | --load-iter 10000 \ 9 | --resume 10 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_8gpu_140k/train.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch 5 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_8gpu_140k/train_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm 8 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_8gpu_140k/validate.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 70000 \ 6 | --validate 7 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/alexnet_yfcc_voc_8gpu_140k/validate_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py --config $work_path/config.yaml --launcher slurm \ 7 | --load-iter 70000 \ 8 | --validate 9 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/resnet50_yfcc+youtube+vip+mpii_lip_16gpu_70k/resume.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch \ 7 | --load-iter 10000 \ 8 | --resume 9 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/resnet50_yfcc+youtube+vip+mpii_lip_16gpu_70k/resume_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm \ 8 | --load-iter 10000 \ 9 | --resume 10 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/resnet50_yfcc+youtube+vip+mpii_lip_16gpu_70k/train.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch 7 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/resnet50_yfcc+youtube+vip+mpii_lip_16gpu_70k/train_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm 8 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/resnet50_yfcc+youtube+vip+mpii_lip_16gpu_70k/validate.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 70000 \ 6 | --validate 7 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/resnet50_yfcc+youtube+vip+mpii_lip_16gpu_70k/validate_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py --config $work_path/config.yaml --launcher slurm \ 7 | --load-iter 70000 \ 8 | --validate 9 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/resnet50_yfcc_coco_16gpu_42k/resume.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch \ 7 | --load-iter 10000 \ 8 | --resume 9 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/resnet50_yfcc_coco_16gpu_42k/resume_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm \ 8 | --load-iter 10000 \ 9 | --resume 10 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/resnet50_yfcc_coco_16gpu_42k/train.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch 7 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/resnet50_yfcc_coco_16gpu_42k/train_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm 8 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/resnet50_yfcc_coco_16gpu_42k/validate.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 70000 \ 6 | --validate 7 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/resnet50_yfcc_coco_16gpu_42k/validate_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py --config $work_path/config.yaml --launcher slurm \ 7 | --load-iter 70000 \ 8 | --validate 9 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/resnet50_yfcc_voc_16gpu_42k/resume.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch \ 7 | --load-iter 10000 \ 8 | --resume 9 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/resnet50_yfcc_voc_16gpu_42k/resume_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm \ 8 | --load-iter 10000 \ 9 | --resume 10 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/resnet50_yfcc_voc_16gpu_42k/train.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 \ 4 | --nnodes=2 --node_rank=$1 \ 5 | --master_addr="192.168.1.1" main.py \ 6 | --config $work_path/config.yaml --launcher pytorch 7 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/resnet50_yfcc_voc_16gpu_42k/train_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n16 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm 8 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/resnet50_yfcc_voc_16gpu_42k/validate.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 70000 \ 6 | --validate 7 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/rep_learning/resnet50_yfcc_voc_16gpu_42k/validate_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py --config $work_path/config.yaml --launcher slurm \ 7 | --load-iter 70000 \ 8 | --validate 9 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/semiauto_annot/resnet50_vip+mpii_liteflow/checkpoints/.DS_Store: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/experiments/semiauto_annot/resnet50_vip+mpii_liteflow/checkpoints/.DS_Store -------------------------------------------------------------------------------- /Training/models/cmp/experiments/semiauto_annot/resnet50_vip+mpii_liteflow/resume.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 10000 \ 6 | --resume 7 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/semiauto_annot/resnet50_vip+mpii_liteflow/resume_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm \ 8 | --load-iter 10000 \ 9 | --resume 10 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/semiauto_annot/resnet50_vip+mpii_liteflow/train.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch 5 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/semiauto_annot/resnet50_vip+mpii_liteflow/train_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition -n8 \ 5 | --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py \ 7 | --config $work_path/config.yaml --launcher slurm 8 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/semiauto_annot/resnet50_vip+mpii_liteflow/validate.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | python -m torch.distributed.launch --nproc_per_node=8 main.py \ 4 | --config $work_path/config.yaml --launcher pytorch \ 5 | --load-iter 70000 \ 6 | --validate 7 | -------------------------------------------------------------------------------- /Training/models/cmp/experiments/semiauto_annot/resnet50_vip+mpii_liteflow/validate_slurm.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | work_path=$(dirname $0) 3 | partition=$1 4 | GLOG_vmodule=MemcachedClient=-1 srun --mpi=pmi2 -p $partition \ 5 | -n8 --gres=gpu:8 --ntasks-per-node=8 \ 6 | python -u main.py --config $work_path/config.yaml --launcher slurm \ 7 | --load-iter 70000 \ 8 | --validate 9 | -------------------------------------------------------------------------------- /Training/models/cmp/models/__init__.py: -------------------------------------------------------------------------------- 1 | from .single_stage_model import * 2 | from .cmp import * 3 | from . import modules 4 | from . import backbone 5 | -------------------------------------------------------------------------------- /Training/models/cmp/models/__pycache__/__init__.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/models/__pycache__/__init__.cpython-310.pyc -------------------------------------------------------------------------------- /Training/models/cmp/models/__pycache__/__init__.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/models/__pycache__/__init__.cpython-39.pyc -------------------------------------------------------------------------------- /Training/models/cmp/models/__pycache__/cmp.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/models/__pycache__/cmp.cpython-310.pyc -------------------------------------------------------------------------------- /Training/models/cmp/models/__pycache__/cmp.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/models/__pycache__/cmp.cpython-39.pyc -------------------------------------------------------------------------------- /Training/models/cmp/models/__pycache__/single_stage_model.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/models/__pycache__/single_stage_model.cpython-310.pyc -------------------------------------------------------------------------------- /Training/models/cmp/models/__pycache__/single_stage_model.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/models/__pycache__/single_stage_model.cpython-39.pyc -------------------------------------------------------------------------------- /Training/models/cmp/models/backbone/__init__.py: -------------------------------------------------------------------------------- 1 | from .resnet import * 2 | from .alexnet import * 3 | -------------------------------------------------------------------------------- /Training/models/cmp/models/backbone/__pycache__/__init__.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/models/backbone/__pycache__/__init__.cpython-310.pyc -------------------------------------------------------------------------------- /Training/models/cmp/models/backbone/__pycache__/__init__.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/models/backbone/__pycache__/__init__.cpython-39.pyc -------------------------------------------------------------------------------- /Training/models/cmp/models/backbone/__pycache__/alexnet.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/models/backbone/__pycache__/alexnet.cpython-310.pyc -------------------------------------------------------------------------------- /Training/models/cmp/models/backbone/__pycache__/alexnet.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/models/backbone/__pycache__/alexnet.cpython-39.pyc -------------------------------------------------------------------------------- /Training/models/cmp/models/backbone/__pycache__/resnet.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/models/backbone/__pycache__/resnet.cpython-310.pyc -------------------------------------------------------------------------------- /Training/models/cmp/models/backbone/__pycache__/resnet.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/models/backbone/__pycache__/resnet.cpython-39.pyc -------------------------------------------------------------------------------- /Training/models/cmp/models/modules/__init__.py: -------------------------------------------------------------------------------- 1 | from .warp import * 2 | from .others import * 3 | from .shallownet import * 4 | from .decoder import * 5 | from .cmp import * 6 | 7 | -------------------------------------------------------------------------------- /Training/models/cmp/models/modules/__pycache__/__init__.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/models/modules/__pycache__/__init__.cpython-310.pyc -------------------------------------------------------------------------------- /Training/models/cmp/models/modules/__pycache__/__init__.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/models/modules/__pycache__/__init__.cpython-39.pyc -------------------------------------------------------------------------------- /Training/models/cmp/models/modules/__pycache__/cmp.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/models/modules/__pycache__/cmp.cpython-310.pyc -------------------------------------------------------------------------------- /Training/models/cmp/models/modules/__pycache__/cmp.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/models/modules/__pycache__/cmp.cpython-39.pyc -------------------------------------------------------------------------------- /Training/models/cmp/models/modules/__pycache__/decoder.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/models/modules/__pycache__/decoder.cpython-310.pyc -------------------------------------------------------------------------------- /Training/models/cmp/models/modules/__pycache__/decoder.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/models/modules/__pycache__/decoder.cpython-39.pyc -------------------------------------------------------------------------------- /Training/models/cmp/models/modules/__pycache__/others.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/models/modules/__pycache__/others.cpython-310.pyc -------------------------------------------------------------------------------- /Training/models/cmp/models/modules/__pycache__/others.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/models/modules/__pycache__/others.cpython-39.pyc -------------------------------------------------------------------------------- /Training/models/cmp/models/modules/__pycache__/shallownet.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/models/modules/__pycache__/shallownet.cpython-310.pyc -------------------------------------------------------------------------------- /Training/models/cmp/models/modules/__pycache__/shallownet.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/models/modules/__pycache__/shallownet.cpython-39.pyc -------------------------------------------------------------------------------- /Training/models/cmp/models/modules/__pycache__/warp.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/models/modules/__pycache__/warp.cpython-310.pyc -------------------------------------------------------------------------------- /Training/models/cmp/models/modules/__pycache__/warp.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/models/modules/__pycache__/warp.cpython-39.pyc -------------------------------------------------------------------------------- /Training/models/cmp/models/modules/others.py: -------------------------------------------------------------------------------- 1 | import torch.nn as nn 2 | 3 | class FixModule(nn.Module): 4 | 5 | def __init__(self, m): 6 | super(FixModule, self).__init__() 7 | self.module = m 8 | 9 | def forward(self, *args, **kwargs): 10 | return self.module(*args, **kwargs) 11 | 12 | -------------------------------------------------------------------------------- /Training/models/cmp/utils/__init__.py: -------------------------------------------------------------------------------- 1 | from .common_utils import * 2 | from .data_utils import * 3 | from .distributed_utils import * 4 | from .visualize_utils import * 5 | from .scheduler import * 6 | from . import flowlib 7 | -------------------------------------------------------------------------------- /Training/models/cmp/utils/__pycache__/__init__.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/utils/__pycache__/__init__.cpython-310.pyc -------------------------------------------------------------------------------- /Training/models/cmp/utils/__pycache__/__init__.cpython-38.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/utils/__pycache__/__init__.cpython-38.pyc -------------------------------------------------------------------------------- /Training/models/cmp/utils/__pycache__/__init__.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/utils/__pycache__/__init__.cpython-39.pyc -------------------------------------------------------------------------------- /Training/models/cmp/utils/__pycache__/common_utils.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/utils/__pycache__/common_utils.cpython-310.pyc -------------------------------------------------------------------------------- /Training/models/cmp/utils/__pycache__/common_utils.cpython-38.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/utils/__pycache__/common_utils.cpython-38.pyc -------------------------------------------------------------------------------- /Training/models/cmp/utils/__pycache__/common_utils.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/utils/__pycache__/common_utils.cpython-39.pyc -------------------------------------------------------------------------------- /Training/models/cmp/utils/__pycache__/data_utils.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/utils/__pycache__/data_utils.cpython-310.pyc -------------------------------------------------------------------------------- /Training/models/cmp/utils/__pycache__/data_utils.cpython-38.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/utils/__pycache__/data_utils.cpython-38.pyc -------------------------------------------------------------------------------- /Training/models/cmp/utils/__pycache__/data_utils.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/utils/__pycache__/data_utils.cpython-39.pyc -------------------------------------------------------------------------------- /Training/models/cmp/utils/__pycache__/distributed_utils.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/utils/__pycache__/distributed_utils.cpython-310.pyc -------------------------------------------------------------------------------- /Training/models/cmp/utils/__pycache__/distributed_utils.cpython-38.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/utils/__pycache__/distributed_utils.cpython-38.pyc -------------------------------------------------------------------------------- /Training/models/cmp/utils/__pycache__/distributed_utils.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/utils/__pycache__/distributed_utils.cpython-39.pyc -------------------------------------------------------------------------------- /Training/models/cmp/utils/__pycache__/flowlib.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/utils/__pycache__/flowlib.cpython-310.pyc -------------------------------------------------------------------------------- /Training/models/cmp/utils/__pycache__/flowlib.cpython-38.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/utils/__pycache__/flowlib.cpython-38.pyc -------------------------------------------------------------------------------- /Training/models/cmp/utils/__pycache__/flowlib.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/utils/__pycache__/flowlib.cpython-39.pyc -------------------------------------------------------------------------------- /Training/models/cmp/utils/__pycache__/scheduler.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/utils/__pycache__/scheduler.cpython-310.pyc -------------------------------------------------------------------------------- /Training/models/cmp/utils/__pycache__/scheduler.cpython-38.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/utils/__pycache__/scheduler.cpython-38.pyc -------------------------------------------------------------------------------- /Training/models/cmp/utils/__pycache__/scheduler.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/utils/__pycache__/scheduler.cpython-39.pyc -------------------------------------------------------------------------------- /Training/models/cmp/utils/__pycache__/visualize_utils.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/utils/__pycache__/visualize_utils.cpython-310.pyc -------------------------------------------------------------------------------- /Training/models/cmp/utils/__pycache__/visualize_utils.cpython-38.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/utils/__pycache__/visualize_utils.cpython-38.pyc -------------------------------------------------------------------------------- /Training/models/cmp/utils/__pycache__/visualize_utils.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/models/cmp/utils/__pycache__/visualize_utils.cpython-39.pyc -------------------------------------------------------------------------------- /Training/pipeline/__pycache__/pipeline.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/pipeline/__pycache__/pipeline.cpython-310.pyc -------------------------------------------------------------------------------- /Training/pipeline/__pycache__/pipeline.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/pipeline/__pycache__/pipeline.cpython-39.pyc -------------------------------------------------------------------------------- /Training/rec_para.txt: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/rec_para.txt -------------------------------------------------------------------------------- /Training/requirements.txt: -------------------------------------------------------------------------------- 1 | diffusers==0.24.0 2 | gradio==4.5.0 3 | opencv-python 4 | opencv-python-headless 5 | scikit-image 6 | torch==2.0.1 7 | torchvision==0.15.2 8 | einops 9 | accelerate==0.30.1 10 | transformers==4.41.1 11 | colorlog 12 | cupy-cuda117 13 | av 14 | gpustat 15 | decord 16 | bitsandbytes 17 | 18 | # xformers==0.0.22 19 | # decord 20 | # bitsandbytes 21 | 22 | -------------------------------------------------------------------------------- /Training/train_stage1.sh: -------------------------------------------------------------------------------- 1 | EXP_NAME="train_stage1" 2 | 3 | accelerate launch train_stage1.py \ 4 | --pretrained_model_name_or_path="./ckpts/stable-video-diffusion-img2vid-xt-1-1" \ 5 | --output_dir="logs/${EXP_NAME}/" \ 6 | --width=384 \ 7 | --height=384 \ 8 | --seed=42 \ 9 | --learning_rate=2e-5 \ 10 | --per_gpu_batch_size=1 \ 11 | --num_train_epochs=5 \ 12 | --mixed_precision="fp16" \ 13 | --gradient_accumulation_steps=1 \ 14 | --checkpointing_steps=2500 \ 15 | --checkpoints_total_limit=100 \ 16 | --validation_steps=2500 \ 17 | --num_frames=25 \ 18 | --gradient_checkpointing \ 19 | --num_validation_images=4 \ 20 | --sample_stride=4 \ 21 | -------------------------------------------------------------------------------- /Training/train_utils/.DS_Store: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/.DS_Store -------------------------------------------------------------------------------- /Training/train_utils/__pycache__/dataset.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/__pycache__/dataset.cpython-310.pyc -------------------------------------------------------------------------------- /Training/train_utils/__pycache__/dataset.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/__pycache__/dataset.cpython-39.pyc -------------------------------------------------------------------------------- /Training/train_utils/__pycache__/sample_flow_utils.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/__pycache__/sample_flow_utils.cpython-39.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/.DS_Store: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/.DS_Store -------------------------------------------------------------------------------- /Training/train_utils/unimatch/dataloader/.DS_Store: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/dataloader/.DS_Store -------------------------------------------------------------------------------- /Training/train_utils/unimatch/dataloader/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/dataloader/__init__.py -------------------------------------------------------------------------------- /Training/train_utils/unimatch/dataloader/__pycache__/__init__.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/dataloader/__pycache__/__init__.cpython-310.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/dataloader/flow/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/dataloader/flow/__init__.py -------------------------------------------------------------------------------- /Training/train_utils/unimatch/dataloader/flow/__pycache__/__init__.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/dataloader/flow/__pycache__/__init__.cpython-310.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/dataloader/flow/__pycache__/datasets.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/dataloader/flow/__pycache__/datasets.cpython-310.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/dataloader/flow/__pycache__/transforms.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/dataloader/flow/__pycache__/transforms.cpython-310.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/dataloader/stereo/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/dataloader/stereo/__init__.py -------------------------------------------------------------------------------- /Training/train_utils/unimatch/loss/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/loss/__init__.py -------------------------------------------------------------------------------- /Training/train_utils/unimatch/loss/__pycache__/__init__.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/loss/__pycache__/__init__.cpython-310.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/loss/__pycache__/flow_loss.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/loss/__pycache__/flow_loss.cpython-310.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/output/gmflow-scale2-regrefine6-davis/00000_flow.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/output/gmflow-scale2-regrefine6-davis/00000_flow.png -------------------------------------------------------------------------------- /Training/train_utils/unimatch/output/gmflow-scale2-regrefine6-davis/00001_flow.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/output/gmflow-scale2-regrefine6-davis/00001_flow.png -------------------------------------------------------------------------------- /Training/train_utils/unimatch/pip_install.sh: -------------------------------------------------------------------------------- 1 | #!/usr/bin/env bash 2 | 3 | 4 | pip install torch==1.9.0+cu102 torchvision==0.10.0+cu102 torchaudio==0.9.0 -f https://download.pytorch.org/whl/torch_stable.html 5 | 6 | pip install imageio==2.9.0 imageio-ffmpeg matplotlib opencv-python pillow scikit-image scipy tensorboard==2.9.1 setuptools==59.5.0 7 | -------------------------------------------------------------------------------- /Training/train_utils/unimatch/pretrained/.DS_Store: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/pretrained/.DS_Store -------------------------------------------------------------------------------- /Training/train_utils/unimatch/scripts/gmdepth_demo.sh: -------------------------------------------------------------------------------- 1 | #!/usr/bin/env bash 2 | 3 | 4 | # gmdepth-scale1-regrefine1 5 | CUDA_VISIBLE_DEVICES=0 python main_depth.py \ 6 | --inference_dir demo/depth-scannet \ 7 | --output_path output/gmdepth-scale1-regrefine1-scannet \ 8 | --resume pretrained/gmdepth-scale1-regrefine1-resumeflowthings-scannet-90325722.pth \ 9 | --reg_refine \ 10 | --num_reg_refine 1 11 | 12 | # --pred_bidir_depth 13 | 14 | -------------------------------------------------------------------------------- /Training/train_utils/unimatch/scripts/gmdepth_evaluate.sh: -------------------------------------------------------------------------------- 1 | #!/usr/bin/env bash 2 | 3 | 4 | # gmdepth-scale1 5 | CUDA_VISIBLE_DEVICES=0 python main_depth.py \ 6 | --eval \ 7 | --resume pretrained/gmdepth-scale1-resumeflowthings-demon-a2fe127b.pth \ 8 | --val_dataset demon \ 9 | --demon_split scenes11 10 | 11 | 12 | # gmdepth-scale1-regrefine1, this is our final model 13 | CUDA_VISIBLE_DEVICES=0 python main_depth.py \ 14 | --eval \ 15 | --resume pretrained/gmdepth-scale1-regrefine1-resumeflowthings-demon-7c23f230.pth \ 16 | --val_dataset demon \ 17 | --demon_split scenes11 \ 18 | --reg_refine \ 19 | --num_reg_refine 1 20 | 21 | -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__init__.py -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/__init__.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/__init__.cpython-310.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/__init__.cpython-311.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/__init__.cpython-311.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/__init__.cpython-38.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/__init__.cpython-38.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/__init__.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/__init__.cpython-39.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/attention.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/attention.cpython-310.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/attention.cpython-311.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/attention.cpython-311.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/attention.cpython-38.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/attention.cpython-38.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/attention.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/attention.cpython-39.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/backbone.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/backbone.cpython-310.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/backbone.cpython-311.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/backbone.cpython-311.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/backbone.cpython-38.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/backbone.cpython-38.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/backbone.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/backbone.cpython-39.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/geometry.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/geometry.cpython-310.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/geometry.cpython-311.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/geometry.cpython-311.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/geometry.cpython-38.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/geometry.cpython-38.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/geometry.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/geometry.cpython-39.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/matching.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/matching.cpython-310.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/matching.cpython-311.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/matching.cpython-311.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/matching.cpython-38.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/matching.cpython-38.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/matching.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/matching.cpython-39.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/position.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/position.cpython-310.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/position.cpython-311.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/position.cpython-311.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/position.cpython-38.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/position.cpython-38.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/position.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/position.cpython-39.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/reg_refine.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/reg_refine.cpython-310.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/reg_refine.cpython-311.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/reg_refine.cpython-311.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/reg_refine.cpython-38.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/reg_refine.cpython-38.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/reg_refine.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/reg_refine.cpython-39.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/transformer.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/transformer.cpython-310.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/transformer.cpython-311.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/transformer.cpython-311.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/transformer.cpython-38.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/transformer.cpython-38.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/transformer.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/transformer.cpython-39.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/trident_conv.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/trident_conv.cpython-310.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/trident_conv.cpython-311.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/trident_conv.cpython-311.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/trident_conv.cpython-38.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/trident_conv.cpython-38.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/trident_conv.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/trident_conv.cpython-39.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/unimatch.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/unimatch.cpython-310.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/unimatch.cpython-311.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/unimatch.cpython-311.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/unimatch.cpython-38.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/unimatch.cpython-38.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/unimatch.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/unimatch.cpython-39.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/utils.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/utils.cpython-310.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/utils.cpython-311.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/utils.cpython-311.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/utils.cpython-38.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/utils.cpython-38.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/unimatch/__pycache__/utils.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/unimatch/__pycache__/utils.cpython-39.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/utils/__pycache__/dist_utils.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/utils/__pycache__/dist_utils.cpython-310.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/utils/__pycache__/file_io.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/utils/__pycache__/file_io.cpython-310.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/utils/__pycache__/flow_viz.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/utils/__pycache__/flow_viz.cpython-310.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/utils/__pycache__/flow_viz.cpython-311.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/utils/__pycache__/flow_viz.cpython-311.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/utils/__pycache__/flow_viz.cpython-38.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/utils/__pycache__/flow_viz.cpython-38.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/utils/__pycache__/flow_viz.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/utils/__pycache__/flow_viz.cpython-39.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/utils/__pycache__/frame_utils.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/utils/__pycache__/frame_utils.cpython-310.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/utils/__pycache__/logger.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/utils/__pycache__/logger.cpython-310.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/utils/__pycache__/misc.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/utils/__pycache__/misc.cpython-310.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/utils/__pycache__/utils.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/utils/__pycache__/utils.cpython-310.pyc -------------------------------------------------------------------------------- /Training/train_utils/unimatch/utils/__pycache__/visualization.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/train_utils/unimatch/utils/__pycache__/visualization.cpython-310.pyc -------------------------------------------------------------------------------- /Training/utils/__pycache__/scheduling_euler_discrete_karras_fix.cpython-310.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/utils/__pycache__/scheduling_euler_discrete_karras_fix.cpython-310.pyc -------------------------------------------------------------------------------- /Training/utils/__pycache__/scheduling_euler_discrete_karras_fix.cpython-39.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/Training/utils/__pycache__/scheduling_euler_discrete_karras_fix.cpython-39.pyc -------------------------------------------------------------------------------- /assets/images/README.md: -------------------------------------------------------------------------------- 1 | README 2 | -------------------------------------------------------------------------------- /assets/images/hybrid_audio.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/assets/images/hybrid_audio.png -------------------------------------------------------------------------------- /assets/images/hybrid_video.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/assets/images/hybrid_video.png -------------------------------------------------------------------------------- /assets/images/pipeline.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/assets/images/pipeline.png -------------------------------------------------------------------------------- /assets/images/project-mofa.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/assets/images/project-mofa.png -------------------------------------------------------------------------------- /assets/images/traj.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/assets/images/traj.png -------------------------------------------------------------------------------- /assets/teasers/hybrid/input.gif: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/assets/teasers/hybrid/input.gif -------------------------------------------------------------------------------- /assets/teasers/hybrid/landmark.gif: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/assets/teasers/hybrid/landmark.gif -------------------------------------------------------------------------------- /assets/teasers/hybrid/output.gif: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/assets/teasers/hybrid/output.gif -------------------------------------------------------------------------------- /assets/teasers/ldmk/flow.gif: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/assets/teasers/ldmk/flow.gif -------------------------------------------------------------------------------- /assets/teasers/ldmk/input.gif: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/assets/teasers/ldmk/input.gif -------------------------------------------------------------------------------- /assets/teasers/ldmk/landmark.gif: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/assets/teasers/ldmk/landmark.gif -------------------------------------------------------------------------------- /assets/teasers/ldmk/output.gif: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/assets/teasers/ldmk/output.gif -------------------------------------------------------------------------------- /assets/teasers/ldmk/output_ldmk.gif: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/assets/teasers/ldmk/output_ldmk.gif -------------------------------------------------------------------------------- /assets/teasers/traj/flow.gif: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/assets/teasers/traj/flow.gif -------------------------------------------------------------------------------- /assets/teasers/traj/hint.gif: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/assets/teasers/traj/hint.gif -------------------------------------------------------------------------------- /assets/teasers/traj/input.gif: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/assets/teasers/traj/input.gif -------------------------------------------------------------------------------- /assets/teasers/traj/output.gif: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/MyNiuuu/MOFA-Video/3ad3b546da4ce0a2cc8aa72f2f8028fbf788c8cd/assets/teasers/traj/output.gif -------------------------------------------------------------------------------- /static/css/my_style.css: -------------------------------------------------------------------------------- 1 | .video_row { 2 | width: 90%; 3 | margin: auto; 4 | display: block; 5 | display: flex; 6 | align-items: center; 7 | justify-content: space-between; 8 | } 9 | 10 | .video_container { 11 | flex: 1; 12 | text-align: center; 13 | border: none; 14 | height: auto; 15 | margin-left: 2px; 16 | margin-right: 2px; 17 | } 18 | 19 | 20 | .span_captions { 21 | width: 90%; 22 | margin: auto; 23 | display: block; 24 | display: flex; 25 | justify-content: space-between; 26 | } 27 | 28 | .caption { 29 | flex: 1; 30 | text-align: center; 31 | font-size: 12px; 32 | } --------------------------------------------------------------------------------