├── .dockerignore ├── .gitattributes ├── .github ├── ISSUE_TEMPLATE │ ├── bug_report.yaml │ └── feature_request.yaml └── workflows │ ├── assign.yaml │ ├── docker-cd.yaml │ ├── issue.yaml │ ├── python.yaml │ └── release.yaml ├── .gitignore ├── .pre-commit-config.yaml ├── .readthedocs.yaml ├── LICENSE ├── MANIFEST.in ├── README.md ├── README_ja_JP.md ├── README_zh_CN.md ├── assets ├── demo.gif ├── index.jpg ├── screenshot.png ├── stay_ahead.gif ├── xinference-downloading.png └── xorbits-logo.png ├── benchmark ├── README.md ├── benchmark_latency.py ├── benchmark_long.py ├── benchmark_rerank.py ├── benchmark_runner.py ├── benchmark_serving.py └── utils.py ├── doc ├── Makefile ├── source │ ├── _static │ │ ├── actor.svg │ │ ├── distributed_inference.png │ │ ├── favicon.svg │ │ └── switcher.json │ ├── conf.py │ ├── development │ │ ├── contributing_codebase.rst │ │ ├── contributing_environment.rst │ │ ├── index.rst │ │ └── xinference_internals.rst │ ├── examples │ │ ├── ai_podcast.rst │ │ ├── chatbot.rst │ │ ├── gradio_chatinterface.rst │ │ ├── index.rst │ │ ├── langchain_streamlit_doc_chat.rst │ │ └── pdf_chatbot.rst │ ├── gen_docs.py │ ├── getting_started │ │ ├── environments.rst │ │ ├── index.rst │ │ ├── installation.rst │ │ ├── installation_npu.rst │ │ ├── logging.rst │ │ ├── troubleshooting.rst │ │ ├── using_docker_image.rst │ │ ├── using_kubernetes.rst │ │ └── using_xinference.rst │ ├── index.rst │ ├── locale │ │ └── zh_CN │ │ │ └── LC_MESSAGES │ │ │ ├── development │ │ │ ├── contributing_codebase.po │ │ │ ├── contributing_environment.po │ │ │ ├── index.po │ │ │ └── xinference_internals.po │ │ │ ├── examples │ │ │ ├── ai_podcast.po │ │ │ ├── chatbot.po │ │ │ ├── gradio_chatinterface.po │ │ │ ├── index.po │ │ │ ├── langchain_streamlit_doc_chat.po │ │ │ └── pdf_chatbot.po │ │ │ ├── getting_started.po │ │ │ ├── getting_started │ │ │ ├── environments.po │ │ │ ├── index.po │ │ │ ├── installation.po │ │ │ ├── installation_npu.po │ │ │ ├── logging.po │ │ │ ├── troubleshooting.po │ │ │ ├── using_docker_image.po │ │ │ ├── using_kubernetes.po │ │ │ └── using_xinference.po │ │ │ ├── index.po │ │ │ ├── models │ │ │ ├── builtin │ │ │ │ ├── audio │ │ │ │ │ └── index.po │ │ │ │ ├── embedding │ │ │ │ │ ├── bge-base-en-v1.5.po │ │ │ │ │ ├── bge-base-en.po │ │ │ │ │ ├── bge-base-zh-v1.5.po │ │ │ │ │ ├── bge-base-zh.po │ │ │ │ │ ├── bge-large-en-v1.5.po │ │ │ │ │ ├── bge-large-en.po │ │ │ │ │ ├── bge-large-zh-noinstruct.po │ │ │ │ │ ├── bge-large-zh-v1.5.po │ │ │ │ │ ├── bge-large-zh.po │ │ │ │ │ ├── bge-small-en-v1.5.po │ │ │ │ │ ├── bge-small-zh-v1.5.po │ │ │ │ │ ├── bge-small-zh.po │ │ │ │ │ ├── e5-large-v2.po │ │ │ │ │ ├── gte-base.po │ │ │ │ │ ├── gte-large.po │ │ │ │ │ ├── index.po │ │ │ │ │ ├── jina-embeddings-v2-base-en.po │ │ │ │ │ ├── jina-embeddings-v2-small-en.po │ │ │ │ │ └── multilingual-e5-large.po │ │ │ │ ├── image │ │ │ │ │ ├── flux.1-dev.po │ │ │ │ │ ├── flux.1-schnell.po │ │ │ │ │ ├── index.po │ │ │ │ │ ├── kolors.po │ │ │ │ │ ├── sd-turbo.po │ │ │ │ │ ├── sd3-medium.po │ │ │ │ │ ├── sdxl-turbo.po │ │ │ │ │ ├── stable-diffusion-2-inpainting.po │ │ │ │ │ ├── stable-diffusion-inpainting.po │ │ │ │ │ ├── stable-diffusion-v1.5.po │ │ │ │ │ ├── stable-diffusion-xl-base-1.0.po │ │ │ │ │ └── stable-diffusion-xl-inpainting.po │ │ │ │ ├── index.po │ │ │ │ ├── llm │ │ │ │ │ ├── baichuan-2-chat.po │ │ │ │ │ ├── baichuan-2.po │ │ │ │ │ ├── baichuan-chat.po │ │ │ │ │ ├── baichuan.po │ │ │ │ │ ├── chatglm.po │ │ │ │ │ ├── chatglm2-32k.po │ │ │ │ │ ├── chatglm2.po │ │ │ │ │ ├── chatglm3-32k.po │ │ │ │ │ ├── chatglm3.po │ │ │ │ │ ├── code-llama-instruct.po │ │ │ │ │ ├── code-llama-python.po │ │ │ │ │ ├── code-llama.po │ │ │ │ │ ├── deepseek-chat.po │ │ │ │ │ ├── deepseek-coder-instruct.po │ │ │ │ │ ├── falcon-instruct.po │ │ │ │ │ ├── falcon.po │ │ │ │ │ ├── glaive-coder.po │ │ │ │ │ ├── gorilla-openfunctions-v1.po │ │ │ │ │ ├── gpt-2.po │ │ │ │ │ ├── index.po │ │ │ │ │ ├── internlm-20b.po │ │ │ │ │ ├── internlm-7b.po │ │ │ │ │ ├── internlm-chat-20b.po │ │ │ │ │ ├── internlm-chat-7b.po │ │ │ │ │ ├── llama-2-chat.po │ │ │ │ │ ├── llama-2.po │ │ │ │ │ ├── mistral-instruct-v0.1.po │ │ │ │ │ ├── mistral-instruct-v0.2.po │ │ │ │ │ ├── mistral-v0.1.po │ │ │ │ │ ├── mixtral-instruct-v0.1.po │ │ │ │ │ ├── mixtral-v0.1.po │ │ │ │ │ ├── openbuddy.po │ │ │ │ │ ├── openhermes-2.5.po │ │ │ │ │ ├── opt.po │ │ │ │ │ ├── orca.po │ │ │ │ │ ├── qwen-chat.po │ │ │ │ │ ├── starchat-beta.po │ │ │ │ │ ├── starcoder.po │ │ │ │ │ ├── starcoderplus.po │ │ │ │ │ ├── tiny-llama.po │ │ │ │ │ ├── vicuna-v1.3.po │ │ │ │ │ ├── vicuna-v1.5-16k.po │ │ │ │ │ ├── vicuna-v1.5.po │ │ │ │ │ ├── wizardcoder-python-v1.0.po │ │ │ │ │ ├── wizardlm-v1.0.po │ │ │ │ │ ├── wizardmath-v1.0.po │ │ │ │ │ ├── xverse-chat.po │ │ │ │ │ ├── xverse.po │ │ │ │ │ ├── yi-200k.po │ │ │ │ │ ├── yi-chat.po │ │ │ │ │ ├── yi.po │ │ │ │ │ ├── zephyr-7b-alpha.po │ │ │ │ │ └── zephyr-7b-beta.po │ │ │ │ ├── rerank │ │ │ │ │ ├── bge-reranker-base.po │ │ │ │ │ ├── bge-reranker-large.po │ │ │ │ │ └── index.po │ │ │ │ └── video │ │ │ │ │ ├── cogvideox-2b.po │ │ │ │ │ └── index.po │ │ │ ├── custom.po │ │ │ ├── index.po │ │ │ ├── lora.po │ │ │ ├── model_abilities │ │ │ │ ├── audio.po │ │ │ │ ├── chat.po │ │ │ │ ├── embed.po │ │ │ │ ├── image.po │ │ │ │ ├── index.po │ │ │ │ ├── multimodal.po │ │ │ │ ├── rerank.po │ │ │ │ ├── tools.po │ │ │ │ └── video.po │ │ │ ├── model_memory.po │ │ │ ├── source │ │ │ │ └── source.po │ │ │ ├── sources │ │ │ │ └── sources.po │ │ │ └── virtualenv.po │ │ │ ├── reference.po │ │ │ ├── reference │ │ │ └── index.po │ │ │ └── user_guide │ │ │ ├── auth_system.po │ │ │ ├── backends.po │ │ │ ├── cache_management.po │ │ │ ├── client_api.po │ │ │ ├── continuous_batching.po │ │ │ ├── distributed_inference.po │ │ │ ├── index.po │ │ │ └── vllm_enhancement.po │ ├── models │ │ ├── builtin │ │ │ ├── audio │ │ │ │ ├── belle-distilwhisper-large-v2-zh.rst │ │ │ │ ├── belle-whisper-large-v2-zh.rst │ │ │ │ ├── belle-whisper-large-v3-zh.rst │ │ │ │ ├── chattts.rst │ │ │ │ ├── cosyvoice-300m-instruct.rst │ │ │ │ ├── cosyvoice-300m-sft.rst │ │ │ │ ├── cosyvoice-300m.rst │ │ │ │ ├── cosyvoice2-0.5b.rst │ │ │ │ ├── f5-tts-mlx.rst │ │ │ │ ├── f5-tts.rst │ │ │ │ ├── fishspeech-1.5.rst │ │ │ │ ├── index.rst │ │ │ │ ├── kokoro-82m.rst │ │ │ │ ├── megatts3.rst │ │ │ │ ├── melotts-chinese.rst │ │ │ │ ├── melotts-english-v2.rst │ │ │ │ ├── melotts-english-v3.rst │ │ │ │ ├── melotts-english.rst │ │ │ │ ├── melotts-french.rst │ │ │ │ ├── melotts-japanese.rst │ │ │ │ ├── melotts-korean.rst │ │ │ │ ├── melotts-spanish.rst │ │ │ │ ├── paraformer-zh-hotword.rst │ │ │ │ ├── paraformer-zh-long.rst │ │ │ │ ├── paraformer-zh-spk.rst │ │ │ │ ├── paraformer-zh.rst │ │ │ │ ├── sensevoicesmall.rst │ │ │ │ ├── whisper-base-mlx.rst │ │ │ │ ├── whisper-base.en-mlx.rst │ │ │ │ ├── whisper-base.en.rst │ │ │ │ ├── whisper-base.rst │ │ │ │ ├── whisper-large-v3-mlx.rst │ │ │ │ ├── whisper-large-v3-turbo-mlx.rst │ │ │ │ ├── whisper-large-v3-turbo.rst │ │ │ │ ├── whisper-large-v3.rst │ │ │ │ ├── whisper-medium-mlx.rst │ │ │ │ ├── whisper-medium.en-mlx.rst │ │ │ │ ├── whisper-medium.en.rst │ │ │ │ ├── whisper-medium.rst │ │ │ │ ├── whisper-small-mlx.rst │ │ │ │ ├── whisper-small.en-mlx.rst │ │ │ │ ├── whisper-small.en.rst │ │ │ │ ├── whisper-small.rst │ │ │ │ ├── whisper-tiny-mlx.rst │ │ │ │ ├── whisper-tiny.en-mlx.rst │ │ │ │ ├── whisper-tiny.en.rst │ │ │ │ └── whisper-tiny.rst │ │ │ ├── embedding │ │ │ │ ├── bce-embedding-base_v1.rst │ │ │ │ ├── bge-base-en-v1.5.rst │ │ │ │ ├── bge-base-en.rst │ │ │ │ ├── bge-base-zh-v1.5.rst │ │ │ │ ├── bge-base-zh.rst │ │ │ │ ├── bge-large-en-v1.5.rst │ │ │ │ ├── bge-large-en.rst │ │ │ │ ├── bge-large-zh-noinstruct.rst │ │ │ │ ├── bge-large-zh-v1.5.rst │ │ │ │ ├── bge-large-zh.rst │ │ │ │ ├── bge-m3.rst │ │ │ │ ├── bge-small-en-v1.5.rst │ │ │ │ ├── bge-small-zh-v1.5.rst │ │ │ │ ├── bge-small-zh.rst │ │ │ │ ├── e5-large-v2.rst │ │ │ │ ├── gte-base.rst │ │ │ │ ├── gte-large.rst │ │ │ │ ├── gte-qwen2.rst │ │ │ │ ├── index.rst │ │ │ │ ├── jina-clip-v2.rst │ │ │ │ ├── jina-embeddings-v2-base-en.rst │ │ │ │ ├── jina-embeddings-v2-base-zh.rst │ │ │ │ ├── jina-embeddings-v2-small-en.rst │ │ │ │ ├── jina-embeddings-v3.rst │ │ │ │ ├── m3e-base.rst │ │ │ │ ├── m3e-large.rst │ │ │ │ ├── m3e-small.rst │ │ │ │ ├── multilingual-e5-large.rst │ │ │ │ ├── text2vec-base-chinese-paraphrase.rst │ │ │ │ ├── text2vec-base-chinese-sentence.rst │ │ │ │ ├── text2vec-base-chinese.rst │ │ │ │ ├── text2vec-base-multilingual.rst │ │ │ │ └── text2vec-large-chinese.rst │ │ │ ├── image │ │ │ │ ├── cogview4.rst │ │ │ │ ├── flux.1-dev.rst │ │ │ │ ├── flux.1-schnell.rst │ │ │ │ ├── got-ocr2_0.rst │ │ │ │ ├── hunyuandit-v1.2-distilled.rst │ │ │ │ ├── hunyuandit-v1.2.rst │ │ │ │ ├── index.rst │ │ │ │ ├── kolors.rst │ │ │ │ ├── sd-turbo.rst │ │ │ │ ├── sd3-medium.rst │ │ │ │ ├── sd3.5-large-turbo.rst │ │ │ │ ├── sd3.5-large.rst │ │ │ │ ├── sd3.5-medium.rst │ │ │ │ ├── sdxl-turbo.rst │ │ │ │ ├── stable-diffusion-2-inpainting.rst │ │ │ │ ├── stable-diffusion-inpainting.rst │ │ │ │ ├── stable-diffusion-v1.5.rst │ │ │ │ ├── stable-diffusion-xl-base-1.0.rst │ │ │ │ └── stable-diffusion-xl-inpainting.rst │ │ │ ├── index.rst │ │ │ ├── llm │ │ │ │ ├── baichuan-2-chat.rst │ │ │ │ ├── baichuan-2.rst │ │ │ │ ├── code-llama-instruct.rst │ │ │ │ ├── code-llama-python.rst │ │ │ │ ├── code-llama.rst │ │ │ │ ├── codegeex4.rst │ │ │ │ ├── codeqwen1.5-chat.rst │ │ │ │ ├── codeqwen1.5.rst │ │ │ │ ├── codeshell-chat.rst │ │ │ │ ├── codeshell.rst │ │ │ │ ├── codestral-v0.1.rst │ │ │ │ ├── cogagent.rst │ │ │ │ ├── deepseek-chat.rst │ │ │ │ ├── deepseek-coder-instruct.rst │ │ │ │ ├── deepseek-coder.rst │ │ │ │ ├── deepseek-prover-v2.rst │ │ │ │ ├── deepseek-r1-0528-qwen3.rst │ │ │ │ ├── deepseek-r1-0528.rst │ │ │ │ ├── deepseek-r1-distill-llama.rst │ │ │ │ ├── deepseek-r1-distill-qwen.rst │ │ │ │ ├── deepseek-r1.rst │ │ │ │ ├── deepseek-v2-chat-0628.rst │ │ │ │ ├── deepseek-v2-chat.rst │ │ │ │ ├── deepseek-v2.5.rst │ │ │ │ ├── deepseek-v3-0324.rst │ │ │ │ ├── deepseek-v3.rst │ │ │ │ ├── deepseek-vl2.rst │ │ │ │ ├── deepseek.rst │ │ │ │ ├── dianjin-r1.rst │ │ │ │ ├── fin-r1.rst │ │ │ │ ├── gemma-3-1b-it.rst │ │ │ │ ├── gemma-3-it.rst │ │ │ │ ├── glm-4v.rst │ │ │ │ ├── glm-edge-chat.rst │ │ │ │ ├── glm4-0414.rst │ │ │ │ ├── glm4-chat-1m.rst │ │ │ │ ├── glm4-chat.rst │ │ │ │ ├── gorilla-openfunctions-v2.rst │ │ │ │ ├── gpt-2.rst │ │ │ │ ├── huatuogpt-o1-llama-3.1.rst │ │ │ │ ├── huatuogpt-o1-qwen2.5.rst │ │ │ │ ├── index.rst │ │ │ │ ├── internlm3-instruct.rst │ │ │ │ ├── internvl3.rst │ │ │ │ ├── llama-2-chat.rst │ │ │ │ ├── llama-2.rst │ │ │ │ ├── llama-3-instruct.rst │ │ │ │ ├── llama-3.1-instruct.rst │ │ │ │ ├── llama-3.1.rst │ │ │ │ ├── llama-3.2-vision-instruct.rst │ │ │ │ ├── llama-3.2-vision.rst │ │ │ │ ├── llama-3.3-instruct.rst │ │ │ │ ├── llama-3.rst │ │ │ │ ├── marco-o1.rst │ │ │ │ ├── minicpm-2b-dpo-bf16.rst │ │ │ │ ├── minicpm-2b-dpo-fp16.rst │ │ │ │ ├── minicpm-2b-dpo-fp32.rst │ │ │ │ ├── minicpm-2b-sft-bf16.rst │ │ │ │ ├── minicpm-2b-sft-fp32.rst │ │ │ │ ├── minicpm-v-2.6.rst │ │ │ │ ├── minicpm3-4b.rst │ │ │ │ ├── mistral-instruct-v0.1.rst │ │ │ │ ├── mistral-instruct-v0.2.rst │ │ │ │ ├── mistral-instruct-v0.3.rst │ │ │ │ ├── mistral-large-instruct.rst │ │ │ │ ├── mistral-nemo-instruct.rst │ │ │ │ ├── mistral-v0.1.rst │ │ │ │ ├── mixtral-8x22b-instruct-v0.1.rst │ │ │ │ ├── mixtral-instruct-v0.1.rst │ │ │ │ ├── mixtral-v0.1.rst │ │ │ │ ├── moonlight-16b-a3b-instruct.rst │ │ │ │ ├── openhermes-2.5.rst │ │ │ │ ├── opt.rst │ │ │ │ ├── orion-chat.rst │ │ │ │ ├── ovis2.rst │ │ │ │ ├── phi-2.rst │ │ │ │ ├── phi-3-mini-128k-instruct.rst │ │ │ │ ├── phi-3-mini-4k-instruct.rst │ │ │ │ ├── qvq-72b-preview.rst │ │ │ │ ├── qwen-chat.rst │ │ │ │ ├── qwen1.5-chat.rst │ │ │ │ ├── qwen1.5-moe-chat.rst │ │ │ │ ├── qwen2-audio-instruct.rst │ │ │ │ ├── qwen2-audio.rst │ │ │ │ ├── qwen2-instruct.rst │ │ │ │ ├── qwen2-moe-instruct.rst │ │ │ │ ├── qwen2-vl-instruct.rst │ │ │ │ ├── qwen2.5-coder-instruct.rst │ │ │ │ ├── qwen2.5-coder.rst │ │ │ │ ├── qwen2.5-instruct-1m.rst │ │ │ │ ├── qwen2.5-instruct.rst │ │ │ │ ├── qwen2.5-omni.rst │ │ │ │ ├── qwen2.5-vl-instruct.rst │ │ │ │ ├── qwen2.5.rst │ │ │ │ ├── qwen3.rst │ │ │ │ ├── qwq-32b-preview.rst │ │ │ │ ├── qwq-32b.rst │ │ │ │ ├── seallm_v2.5.rst │ │ │ │ ├── seallm_v2.rst │ │ │ │ ├── seallms-v3.rst │ │ │ │ ├── skywork-math.rst │ │ │ │ ├── skywork-or1-preview.rst │ │ │ │ ├── skywork-or1.rst │ │ │ │ ├── skywork.rst │ │ │ │ ├── telechat.rst │ │ │ │ ├── tiny-llama.rst │ │ │ │ ├── wizardcoder-python-v1.0.rst │ │ │ │ ├── wizardmath-v1.0.rst │ │ │ │ ├── xiyansql-qwencoder-2504.rst │ │ │ │ ├── xverse-chat.rst │ │ │ │ ├── xverse.rst │ │ │ │ ├── yi-1.5-chat-16k.rst │ │ │ │ ├── yi-1.5-chat.rst │ │ │ │ ├── yi-1.5.rst │ │ │ │ ├── yi-200k.rst │ │ │ │ ├── yi-chat.rst │ │ │ │ └── yi.rst │ │ │ ├── rerank │ │ │ │ ├── bce-reranker-base_v1.rst │ │ │ │ ├── bge-reranker-base.rst │ │ │ │ ├── bge-reranker-large.rst │ │ │ │ ├── bge-reranker-v2-gemma.rst │ │ │ │ ├── bge-reranker-v2-m3.rst │ │ │ │ ├── bge-reranker-v2-minicpm-layerwise.rst │ │ │ │ ├── index.rst │ │ │ │ ├── jina-reranker-v2.rst │ │ │ │ └── minicpm-reranker.rst │ │ │ └── video │ │ │ │ ├── cogvideox-2b.rst │ │ │ │ ├── cogvideox-5b.rst │ │ │ │ ├── hunyuanvideo.rst │ │ │ │ ├── index.rst │ │ │ │ ├── wan2.1-1.3b.rst │ │ │ │ ├── wan2.1-14b.rst │ │ │ │ ├── wan2.1-flf2v-14b-720p.rst │ │ │ │ ├── wan2.1-i2v-14b-480p.rst │ │ │ │ └── wan2.1-i2v-14b-720p.rst │ │ ├── custom.rst │ │ ├── index.rst │ │ ├── lora.rst │ │ ├── model_abilities │ │ │ ├── audio.rst │ │ │ ├── chat.rst │ │ │ ├── embed.rst │ │ │ ├── image.rst │ │ │ ├── index.rst │ │ │ ├── multimodal.rst │ │ │ ├── rerank.rst │ │ │ ├── tools.rst │ │ │ └── video.rst │ │ ├── model_memory.rst │ │ ├── sources │ │ │ └── sources.rst │ │ └── virtualenv.rst │ ├── norm_zh.py │ ├── reference │ │ └── index.rst │ └── user_guide │ │ ├── auth_system.rst │ │ ├── backends.rst │ │ ├── client_api.rst │ │ ├── continuous_batching.rst │ │ ├── distributed_inference.rst │ │ ├── index.rst │ │ ├── metrics.rst │ │ └── vllm_enhancement.rst └── templates │ ├── audio.rst.jinja │ ├── audio_index.rst.jinja │ ├── embedding.rst.jinja │ ├── embedding_index.rst.jinja │ ├── image.rst.jinja │ ├── image_index.rst.jinja │ ├── llm.rst.jinja │ ├── llm_index.rst.jinja │ ├── metrics.jinja │ ├── rerank.rst.jinja │ ├── rerank_index.rst.jinja │ ├── video.rst.jinja │ └── video_index.rst.jinja ├── examples ├── AI_podcast.py ├── AI_podcast_ZH.py ├── AI_translate.py ├── Custom_StableDiffusion_ControlNet.ipynb ├── FunctionCall.ipynb ├── LangChain_QA.ipynb ├── LangChain_Streamlit_Doc_Chat.py ├── StableDiffusionControlNet.ipynb ├── Xinference_Quick_Start.ipynb ├── audio_to_text.ipynb ├── chat.py ├── chat_vl.ipynb ├── draft.png └── gradio_chatinterface.py ├── pyproject.toml ├── setup.cfg ├── setup.py ├── versioneer.py └── xinference ├── __init__.py ├── _compat.py ├── _version.py ├── api ├── __init__.py ├── oauth2 │ ├── __init__.py │ ├── auth_service.py │ ├── types.py │ └── utils.py └── restful_api.py ├── client ├── __init__.py ├── common.py ├── handlers.py ├── restful │ ├── __init__.py │ └── restful_client.py └── tests │ ├── __init__.py │ ├── test_client.py │ └── test_client_with_auth.py ├── conftest.py ├── constants.py ├── core ├── __init__.py ├── cache_tracker.py ├── chat_interface.py ├── event.py ├── media_interface.py ├── metrics.py ├── model.py ├── progress_tracker.py ├── resource.py ├── scheduler.py ├── status_guard.py ├── supervisor.py ├── tests │ ├── __init__.py │ ├── test_continuous_batching.py │ ├── test_metrics.py │ ├── test_model.py │ ├── test_progressor.py │ ├── test_restful_api.py │ ├── test_types.py │ ├── test_utils.py │ └── test_worker.py ├── utils.py └── worker.py ├── deploy ├── __init__.py ├── cmdline.py ├── docker │ ├── Dockerfile │ ├── cpu.Dockerfile │ ├── docker-compose-distributed.yml │ ├── docker-compose.yml │ ├── requirements-base.txt │ ├── requirements-ml.txt │ ├── requirements-models.txt │ ├── requirements.txt │ ├── requirements_cpu-base.txt │ ├── requirements_cpu-ml.txt │ ├── requirements_cpu-models.txt │ └── requirements_cpu.txt ├── local.py ├── supervisor.py ├── test │ ├── __init__.py │ └── test_cmdline.py ├── utils.py └── worker.py ├── device_utils.py ├── fields.py ├── isolation.py ├── model ├── __init__.py ├── audio │ ├── __init__.py │ ├── chattts.py │ ├── core.py │ ├── cosyvoice.py │ ├── custom.py │ ├── f5tts.py │ ├── f5tts_mlx.py │ ├── fish_speech.py │ ├── funasr.py │ ├── kokoro.py │ ├── megatts.py │ ├── melotts.py │ ├── model_spec.json │ ├── model_spec_modelscope.json │ ├── tests │ │ ├── __init__.py │ │ ├── basic_ref_en.wav │ │ ├── bbc_news.npy │ │ ├── bbc_news.wav │ │ ├── common_voice_zh-CN_38026095.mp3 │ │ ├── cross_lingual_prompt.wav │ │ ├── jfk.flac │ │ ├── test_chattts.py │ │ ├── test_cosyvoice.py │ │ ├── test_f5tts.py │ │ ├── test_f5tts_mlx.py │ │ ├── test_fish_speech.py │ │ ├── test_funasr.py │ │ ├── test_kokoro.py │ │ ├── test_megatts.py │ │ ├── test_melotts.py │ │ ├── test_whisper.py │ │ ├── test_whisper_mlx.py │ │ └── zero_shot_prompt.wav │ ├── utils.py │ ├── whisper.py │ └── whisper_mlx.py ├── core.py ├── embedding │ ├── __init__.py │ ├── core.py │ ├── custom.py │ ├── model_spec.json │ ├── model_spec_modelscope.json │ ├── tests │ │ ├── __init__.py │ │ ├── test_embedding_models.py │ │ └── test_integrated_embedding.py │ └── utils.py ├── flexible │ ├── __init__.py │ ├── core.py │ ├── launchers │ │ ├── __init__.py │ │ ├── image_process_launcher.py │ │ └── transformers_launcher.py │ ├── tests │ │ ├── __init__.py │ │ └── test_flexible_models.py │ └── utils.py ├── image │ ├── __init__.py │ ├── core.py │ ├── custom.py │ ├── model_spec.json │ ├── model_spec_modelscope.json │ ├── ocr │ │ ├── __init__.py │ │ └── got_ocr2.py │ ├── scheduler │ │ ├── __init__.py │ │ └── flux.py │ ├── sdapi.py │ ├── stable_diffusion │ │ ├── __init__.py │ │ ├── core.py │ │ └── mlx.py │ ├── tests │ │ ├── __init__.py │ │ ├── test_got_ocr2.py │ │ └── test_stable_diffusion.py │ └── utils.py ├── llm │ ├── __init__.py │ ├── core.py │ ├── llama_cpp │ │ ├── __init__.py │ │ ├── core.py │ │ ├── memory.py │ │ └── tests │ │ │ ├── __init__.py │ │ │ ├── dummy.gguf │ │ │ └── test_gguf.py │ ├── llm_family.json │ ├── llm_family.py │ ├── llm_family_csghub.json │ ├── llm_family_modelscope.json │ ├── llm_family_openmind_hub.json │ ├── lmdeploy │ │ ├── __init__.py │ │ ├── core.py │ │ └── tests │ │ │ └── __init__.py │ ├── memory.py │ ├── mlx │ │ ├── __init__.py │ │ ├── core.py │ │ └── tests │ │ │ ├── __init__.py │ │ │ ├── fish.png │ │ │ └── test_mlx.py │ ├── reasoning_parser.py │ ├── sglang │ │ ├── __init__.py │ │ └── core.py │ ├── tests │ │ ├── __init__.py │ │ ├── test_llm_family.py │ │ ├── test_llm_model.py │ │ ├── test_memory_estimate.py │ │ ├── test_multimodal.py │ │ ├── test_stream_options.py │ │ └── test_utils.py │ ├── transformers │ │ ├── __init__.py │ │ ├── chatglm.py │ │ ├── core.py │ │ ├── deepseek_v2.py │ │ ├── gemma3.py │ │ ├── multimodal │ │ │ ├── __init__.py │ │ │ ├── cogagent.py │ │ │ ├── core.py │ │ │ ├── deepseek_vl2.py │ │ │ ├── gemma3.py │ │ │ ├── glm4v.py │ │ │ ├── intern_vl.py │ │ │ ├── minicpmv26.py │ │ │ ├── ovis2.py │ │ │ ├── qwen-omni.py │ │ │ ├── qwen2_audio.py │ │ │ └── qwen2_vl.py │ │ ├── opt.py │ │ ├── tensorizer_utils.py │ │ ├── tests │ │ │ ├── __init__.py │ │ │ ├── test_opt.py │ │ │ └── test_tensorizer.py │ │ └── utils.py │ ├── utils.py │ └── vllm │ │ ├── __init__.py │ │ ├── core.py │ │ ├── distributed_executor.py │ │ ├── tests │ │ ├── __init__.py │ │ └── test_distributed_executor.py │ │ ├── utils.py │ │ └── xavier │ │ ├── __init__.py │ │ ├── allocator.py │ │ ├── block.py │ │ ├── block_manager.py │ │ ├── block_tracker.py │ │ ├── collective.py │ │ ├── collective_manager.py │ │ ├── engine.py │ │ ├── executor.py │ │ ├── scheduler.py │ │ ├── test │ │ ├── __init__.py │ │ └── test_xavier.py │ │ └── transfer.py ├── rerank │ ├── __init__.py │ ├── core.py │ ├── custom.py │ ├── model_spec.json │ ├── model_spec_modelscope.json │ ├── tests │ │ ├── __init__.py │ │ └── test_rerank.py │ └── utils.py ├── tests │ ├── __init__.py │ └── test_utils.py ├── utils.py └── video │ ├── __init__.py │ ├── core.py │ ├── diffusers.py │ ├── model_spec.json │ ├── model_spec_modelscope.json │ └── tests │ ├── __init__.py │ └── test_diffusers_video.py ├── thirdparty ├── __init__.py ├── cosyvoice │ ├── __init__.py │ ├── bin │ │ ├── average_model.py │ │ ├── export_jit.py │ │ ├── export_onnx.py │ │ ├── inference.py │ │ ├── spk2info.pt │ │ └── train.py │ ├── cli │ │ ├── __init__.py │ │ ├── cosyvoice.py │ │ ├── frontend.py │ │ └── model.py │ ├── dataset │ │ ├── __init__.py │ │ ├── dataset.py │ │ └── processor.py │ ├── flow │ │ ├── decoder.py │ │ ├── flow.py │ │ ├── flow_matching.py │ │ └── length_regulator.py │ ├── hifigan │ │ ├── discriminator.py │ │ ├── f0_predictor.py │ │ ├── generator.py │ │ └── hifigan.py │ ├── llm │ │ └── llm.py │ ├── tokenizer │ │ ├── assets │ │ │ └── multilingual_zh_ja_yue_char_del.tiktoken │ │ └── tokenizer.py │ ├── transformer │ │ ├── __init__.py │ │ ├── activation.py │ │ ├── attention.py │ │ ├── convolution.py │ │ ├── decoder.py │ │ ├── decoder_layer.py │ │ ├── embedding.py │ │ ├── encoder.py │ │ ├── encoder_layer.py │ │ ├── label_smoothing_loss.py │ │ ├── positionwise_feed_forward.py │ │ ├── subsampling.py │ │ └── upsample_encoder.py │ └── utils │ │ ├── __init__.py │ │ ├── class_utils.py │ │ ├── common.py │ │ ├── executor.py │ │ ├── file_utils.py │ │ ├── frontend_utils.py │ │ ├── losses.py │ │ ├── mask.py │ │ ├── scheduler.py │ │ └── train_utils.py ├── deepseek_vl │ ├── __init__.py │ ├── models │ │ ├── __init__.py │ │ ├── clip_encoder.py │ │ ├── image_processing_vlm.py │ │ ├── modeling_vlm.py │ │ ├── processing_vlm.py │ │ ├── projector.py │ │ ├── sam.py │ │ └── siglip_vit.py │ ├── serve │ │ ├── __init__.py │ │ ├── app_deepseek.py │ │ ├── app_modules │ │ │ ├── __init__.py │ │ │ ├── gradio_utils.py │ │ │ ├── overwrites.py │ │ │ ├── presets.py │ │ │ └── utils.py │ │ ├── assets │ │ │ ├── Kelpy-Codos.js │ │ │ ├── avatar.png │ │ │ ├── custom.css │ │ │ ├── custom.js │ │ │ └── favicon.ico │ │ ├── examples │ │ │ ├── app.png │ │ │ ├── chart.png │ │ │ ├── mirror.png │ │ │ ├── pipeline.png │ │ │ ├── puzzle.png │ │ │ └── rap.jpeg │ │ └── inference.py │ └── utils │ │ ├── __init__.py │ │ ├── conversation.py │ │ └── io.py ├── deepseek_vl2 │ ├── __init__.py │ ├── models │ │ ├── __init__.py │ │ ├── configuration_deepseek.py │ │ ├── conversation.py │ │ ├── modeling_deepseek.py │ │ ├── modeling_deepseek_vl_v2.py │ │ ├── processing_deepseek_vl_v2.py │ │ └── siglip_vit.py │ ├── serve │ │ ├── __init__.py │ │ ├── app_modules │ │ │ ├── __init__.py │ │ │ ├── gradio_utils.py │ │ │ ├── overwrites.py │ │ │ ├── presets.py │ │ │ └── utils.py │ │ ├── assets │ │ │ ├── Kelpy-Codos.js │ │ │ ├── avatar.png │ │ │ ├── custom.css │ │ │ ├── custom.js │ │ │ ├── favicon.ico │ │ │ └── simsun.ttc │ │ └── inference.py │ └── utils │ │ ├── __init__.py │ │ └── io.py ├── f5_tts │ ├── __init__.py │ ├── api.py │ ├── configs │ │ ├── E2TTS_Base_train.yaml │ │ ├── E2TTS_Small_train.yaml │ │ ├── F5TTS_Base_train.yaml │ │ └── F5TTS_Small_train.yaml │ ├── eval │ │ ├── README.md │ │ ├── ecapa_tdnn.py │ │ ├── eval_infer_batch.py │ │ ├── eval_infer_batch.sh │ │ ├── eval_librispeech_test_clean.py │ │ ├── eval_seedtts_testset.py │ │ └── utils_eval.py │ ├── infer │ │ ├── README.md │ │ ├── SHARED.md │ │ ├── examples │ │ │ ├── basic │ │ │ │ ├── basic.toml │ │ │ │ ├── basic_ref_en.wav │ │ │ │ └── basic_ref_zh.wav │ │ │ ├── multi │ │ │ │ ├── country.flac │ │ │ │ ├── main.flac │ │ │ │ ├── story.toml │ │ │ │ ├── story.txt │ │ │ │ └── town.flac │ │ │ └── vocab.txt │ │ ├── infer_cli.py │ │ ├── infer_gradio.py │ │ ├── speech_edit.py │ │ └── utils_infer.py │ ├── model │ │ ├── __init__.py │ │ ├── backbones │ │ │ ├── README.md │ │ │ ├── dit.py │ │ │ ├── mmdit.py │ │ │ └── unett.py │ │ ├── cfm.py │ │ ├── dataset.py │ │ ├── modules.py │ │ ├── trainer.py │ │ └── utils.py │ ├── scripts │ │ ├── count_max_epoch.py │ │ └── count_params_gflops.py │ ├── socket_server.py │ └── train │ │ ├── README.md │ │ ├── datasets │ │ ├── prepare_csv_wavs.py │ │ ├── prepare_emilia.py │ │ ├── prepare_libritts.py │ │ ├── prepare_ljspeech.py │ │ └── prepare_wenetspeech4tts.py │ │ ├── finetune_cli.py │ │ ├── finetune_gradio.py │ │ └── train.py ├── fish_speech │ ├── __init__.py │ ├── fish_speech │ │ ├── __init__.py │ │ ├── callbacks │ │ │ ├── __init__.py │ │ │ └── grad_norm.py │ │ ├── configs │ │ │ ├── base.yaml │ │ │ ├── firefly_gan_vq.yaml │ │ │ ├── lora │ │ │ │ └── r_8_alpha_16.yaml │ │ │ └── text2semantic_finetune.yaml │ │ ├── conversation.py │ │ ├── datasets │ │ │ ├── concat_repeat.py │ │ │ ├── protos │ │ │ │ ├── text-data.proto │ │ │ │ ├── text_data_pb2.py │ │ │ │ └── text_data_stream.py │ │ │ ├── semantic.py │ │ │ └── vqgan.py │ │ ├── i18n │ │ │ ├── README.md │ │ │ ├── __init__.py │ │ │ ├── core.py │ │ │ ├── locale │ │ │ │ ├── en_US.json │ │ │ │ ├── es_ES.json │ │ │ │ ├── ja_JP.json │ │ │ │ ├── ko_KR.json │ │ │ │ ├── pt_BR.json │ │ │ │ └── zh_CN.json │ │ │ └── scan.py │ │ ├── models │ │ │ ├── text2semantic │ │ │ │ ├── __init__.py │ │ │ │ ├── lit_module.py │ │ │ │ ├── llama.py │ │ │ │ └── lora.py │ │ │ └── vqgan │ │ │ │ ├── __init__.py │ │ │ │ ├── modules │ │ │ │ ├── firefly.py │ │ │ │ └── fsq.py │ │ │ │ └── utils.py │ │ ├── scheduler.py │ │ ├── text │ │ │ ├── __init__.py │ │ │ ├── chn_text_norm │ │ │ │ ├── .gitignore │ │ │ │ ├── README.md │ │ │ │ ├── __init__.py │ │ │ │ ├── basic_class.py │ │ │ │ ├── basic_constant.py │ │ │ │ ├── basic_util.py │ │ │ │ ├── cardinal.py │ │ │ │ ├── date.py │ │ │ │ ├── digit.py │ │ │ │ ├── fraction.py │ │ │ │ ├── money.py │ │ │ │ ├── percentage.py │ │ │ │ ├── telephone.py │ │ │ │ └── text.py │ │ │ ├── clean.py │ │ │ └── spliter.py │ │ ├── tokenizer.py │ │ ├── train.py │ │ ├── utils │ │ │ ├── __init__.py │ │ │ ├── braceexpand.py │ │ │ ├── context.py │ │ │ ├── file.py │ │ │ ├── instantiators.py │ │ │ ├── logger.py │ │ │ ├── logging_utils.py │ │ │ ├── rich_utils.py │ │ │ ├── spectrogram.py │ │ │ └── utils.py │ │ └── webui │ │ │ ├── css │ │ │ └── style.css │ │ │ ├── html │ │ │ └── footer.html │ │ │ ├── js │ │ │ └── animate.js │ │ │ ├── launch_utils.py │ │ │ └── manage.py │ └── tools │ │ ├── api_client.py │ │ ├── api_server.py │ │ ├── download_models.py │ │ ├── e2e_webui.py │ │ ├── extract_model.py │ │ ├── file.py │ │ ├── fish_e2e.py │ │ ├── inference_engine │ │ ├── __init__.py │ │ ├── reference_loader.py │ │ ├── utils.py │ │ └── vq_manager.py │ │ ├── llama │ │ ├── build_dataset.py │ │ ├── eval_in_context.py │ │ ├── generate.py │ │ ├── merge_lora.py │ │ ├── quantize.py │ │ └── rebuild_tokenizer.py │ │ ├── run_webui.py │ │ ├── schema.py │ │ ├── sensevoice │ │ ├── README.md │ │ ├── __init__.py │ │ ├── auto_model.py │ │ ├── fun_asr.py │ │ └── vad_utils.py │ │ ├── server │ │ ├── agent │ │ │ ├── __init__.py │ │ │ ├── generate.py │ │ │ ├── generation_utils.py │ │ │ └── pre_generation_utils.py │ │ ├── api_utils.py │ │ ├── exception_handler.py │ │ ├── inference.py │ │ ├── model_manager.py │ │ ├── model_utils.py │ │ └── views.py │ │ ├── smart_pad.py │ │ ├── vqgan │ │ ├── create_train_split.py │ │ ├── extract_vq.py │ │ └── inference.py │ │ ├── webui │ │ ├── __init__.py │ │ ├── inference.py │ │ └── variables.py │ │ └── whisper_asr.py ├── internvl │ ├── __init__.py │ └── conversation.py ├── llava │ ├── __init__.py │ ├── conversation.py │ ├── mm_utils.py │ └── model │ │ ├── __init__.py │ │ ├── clip_encoder │ │ ├── __init__.py │ │ ├── builder.py │ │ └── clip_encoder.py │ │ ├── constants.py │ │ ├── llava_arch.py │ │ ├── llava_llama.py │ │ └── multimodal_projector │ │ ├── __init__.py │ │ └── builder.py ├── matcha │ ├── VERSION │ ├── __init__.py │ ├── app.py │ ├── cli.py │ ├── data │ │ ├── __init__.py │ │ ├── components │ │ │ └── __init__.py │ │ └── text_mel_datamodule.py │ ├── hifigan │ │ ├── LICENSE │ │ ├── README.md │ │ ├── __init__.py │ │ ├── config.py │ │ ├── denoiser.py │ │ ├── env.py │ │ ├── meldataset.py │ │ ├── models.py │ │ └── xutils.py │ ├── models │ │ ├── __init__.py │ │ ├── baselightningmodule.py │ │ ├── components │ │ │ ├── __init__.py │ │ │ ├── decoder.py │ │ │ ├── flow_matching.py │ │ │ ├── text_encoder.py │ │ │ └── transformer.py │ │ └── matcha_tts.py │ ├── onnx │ │ ├── __init__.py │ │ ├── export.py │ │ └── infer.py │ ├── text │ │ ├── __init__.py │ │ ├── cleaners.py │ │ ├── numbers.py │ │ └── symbols.py │ ├── train.py │ └── utils │ │ ├── __init__.py │ │ ├── audio.py │ │ ├── generate_data_statistics.py │ │ ├── get_durations_from_trained_model.py │ │ ├── instantiators.py │ │ ├── logging_utils.py │ │ ├── model.py │ │ ├── monotonic_align │ │ ├── __init__.py │ │ ├── core.pyx │ │ └── setup.py │ │ ├── pylogger.py │ │ ├── rich_utils.py │ │ └── utils.py ├── megatts3 │ ├── __init__.py │ └── tts │ │ ├── frontend_function.py │ │ ├── gradio_api.py │ │ ├── infer_cli.py │ │ ├── modules │ │ ├── aligner │ │ │ └── whisper_small.py │ │ ├── ar_dur │ │ │ ├── ar_dur_predictor.py │ │ │ └── commons │ │ │ │ ├── layers.py │ │ │ │ ├── nar_tts_modules.py │ │ │ │ ├── rel_transformer.py │ │ │ │ ├── rot_transformer.py │ │ │ │ ├── seq_utils.py │ │ │ │ └── transformer.py │ │ ├── llm_dit │ │ │ ├── cfm.py │ │ │ ├── dit.py │ │ │ ├── time_embedding.py │ │ │ └── transformer.py │ │ └── wavvae │ │ │ ├── decoder │ │ │ ├── diag_gaussian.py │ │ │ ├── hifigan_modules.py │ │ │ ├── seanet_encoder.py │ │ │ └── wavvae_v3.py │ │ │ └── encoder │ │ │ └── common_modules │ │ │ ├── conv.py │ │ │ ├── lstm.py │ │ │ └── seanet.py │ │ └── utils │ │ ├── audio_utils │ │ ├── align.py │ │ ├── io.py │ │ └── plot.py │ │ ├── commons │ │ ├── ckpt_utils.py │ │ └── hparams.py │ │ └── text_utils │ │ ├── dict.json │ │ ├── ph_tone_convert.py │ │ ├── split_text.py │ │ └── text_encoder.py ├── melo │ ├── __init__.py │ ├── api.py │ ├── app.py │ ├── attentions.py │ ├── commons.py │ ├── configs │ │ └── config.json │ ├── data │ │ └── example │ │ │ └── metadata.list │ ├── data_utils.py │ ├── download_utils.py │ ├── infer.py │ ├── init_downloads.py │ ├── losses.py │ ├── main.py │ ├── mel_processing.py │ ├── models.py │ ├── modules.py │ ├── monotonic_align │ │ ├── __init__.py │ │ └── core.py │ ├── preprocess_text.py │ ├── split_utils.py │ ├── text │ │ ├── __init__.py │ │ ├── chinese.py │ │ ├── chinese_bert.py │ │ ├── chinese_mix.py │ │ ├── cleaner.py │ │ ├── cleaner_multiling.py │ │ ├── cmudict.rep │ │ ├── cmudict_cache.pickle │ │ ├── english.py │ │ ├── english_bert.py │ │ ├── english_utils │ │ │ ├── __init__.py │ │ │ ├── abbreviations.py │ │ │ ├── number_norm.py │ │ │ └── time_norm.py │ │ ├── es_phonemizer │ │ │ ├── __init__.py │ │ │ ├── base.py │ │ │ ├── cleaner.py │ │ │ ├── es_symbols.json │ │ │ ├── es_symbols.txt │ │ │ ├── es_symbols_v2.json │ │ │ ├── es_to_ipa.py │ │ │ ├── example_ipa.txt │ │ │ ├── gruut_wrapper.py │ │ │ ├── punctuation.py │ │ │ ├── spanish_symbols.txt │ │ │ └── test.ipynb │ │ ├── fr_phonemizer │ │ │ ├── __init__.py │ │ │ ├── base.py │ │ │ ├── cleaner.py │ │ │ ├── en_symbols.json │ │ │ ├── example_ipa.txt │ │ │ ├── fr_symbols.json │ │ │ ├── fr_to_ipa.py │ │ │ ├── french_abbreviations.py │ │ │ ├── french_symbols.txt │ │ │ ├── gruut_wrapper.py │ │ │ └── punctuation.py │ │ ├── french.py │ │ ├── french_bert.py │ │ ├── japanese.py │ │ ├── japanese_bert.py │ │ ├── ko_dictionary.py │ │ ├── korean.py │ │ ├── opencpop-strict.txt │ │ ├── spanish.py │ │ ├── spanish_bert.py │ │ ├── symbols.py │ │ └── tone_sandhi.py │ ├── train.py │ ├── train.sh │ ├── transforms.py │ └── utils.py ├── mlx │ ├── __init__.py │ └── flux │ │ ├── __init__.py │ │ ├── autoencoder.py │ │ ├── clip.py │ │ ├── datasets.py │ │ ├── flux.py │ │ ├── layers.py │ │ ├── lora.py │ │ ├── model.py │ │ ├── sampler.py │ │ ├── t5.py │ │ ├── tokenizers.py │ │ ├── trainer.py │ │ └── utils.py └── whisper │ ├── __init__.py │ ├── __main__.py │ ├── assets │ ├── gpt2.tiktoken │ ├── mel_filters.npz │ └── multilingual.tiktoken │ ├── audio.py │ ├── decoding.py │ ├── model.py │ ├── normalizers │ ├── __init__.py │ ├── basic.py │ ├── english.json │ └── english.py │ ├── timing.py │ ├── tokenizer.py │ ├── transcribe.py │ ├── triton_ops.py │ ├── utils.py │ └── version.py ├── types.py ├── utils.py └── web └── ui ├── .eslintignore ├── .eslintrc.yml ├── .gitignore ├── .prettierignore ├── .prettierrc.yml ├── package-lock.json ├── package.json ├── public ├── favicon.svg └── index.html └── src ├── App.js ├── components ├── MenuSide.js ├── Title.js ├── alertComponent.js ├── apiContext.js ├── authAlertDialog.js ├── copyComponent │ ├── copyComponent.js │ └── style.css ├── deleteDialog.js ├── errorMessageSnackBar.js ├── fetchWrapper.js ├── fetcher.js ├── hotkeyFocusTextField.js ├── successMessageSnackBar.js ├── tableTitle.js ├── themeButton.js ├── themeContext.js ├── titleTypography.js ├── translateButton.js ├── utils.js └── versionLabel.js ├── i18n.js ├── index.css ├── index.js ├── locales ├── en.json ├── ja.json ├── ko.json └── zh.json ├── media └── icon.webp ├── router └── index.js ├── scenes ├── _layout │ └── index.js ├── cluster_info │ ├── index.js │ ├── nodeInfo.js │ └── style.js ├── launch_model │ ├── LaunchModel.js │ ├── components │ │ ├── addPair.js │ │ ├── copyComponent.js │ │ ├── drawer.js │ │ ├── pasteDialog.js │ │ └── progress.js │ ├── data │ │ └── data.js │ ├── index.js │ ├── launchCustom.js │ ├── modelCard.js │ └── styles │ │ └── modelCardStyle.css ├── login │ ├── header.js │ └── login.js ├── register_model │ ├── components │ │ ├── addControlnet.js │ │ ├── addModelSpecs.js │ │ └── addStop.js │ ├── data │ │ └── languages.js │ ├── index.js │ ├── registerModel.js │ └── styles │ │ └── registerModelStyle.css └── running_models │ └── index.js └── theme.js /.dockerignore: -------------------------------------------------------------------------------- 1 | doc/ 2 | .idea/ 3 | .github/ 4 | build/ 5 | xinference.egg-info/ 6 | xinference/web/ui/build/ 7 | xinference/web/ui/node_modules/ 8 | -------------------------------------------------------------------------------- /.gitattributes: -------------------------------------------------------------------------------- 1 | xinference/_version.py export-subst 2 | -------------------------------------------------------------------------------- /.github/ISSUE_TEMPLATE/feature_request.yaml: -------------------------------------------------------------------------------- 1 | name: "Feature request" 2 | description: Submit a request for a new Xinference feature / 提交一个新的 Xinference 的功能建议 3 | labels: [ "feature" ] 4 | body: 5 | - type: textarea 6 | id: feature-request 7 | validations: 8 | required: true 9 | attributes: 10 | label: Feature request / 功能建议 11 | description: | 12 | A brief description of the functional proposal. 13 | 对功能建议的简述。 14 | 15 | - type: textarea 16 | id: motivation 17 | validations: 18 | required: true 19 | attributes: 20 | label: Motivation / 动机 21 | description: | 22 | Your motivation for making the suggestion. If that motivation is related to another GitHub issue, link to it here. 23 | 您提出建议的动机。如果该动机与另一个 GitHub 问题有关,请在此处提供对应的链接。 24 | 25 | - type: textarea 26 | id: contribution 27 | validations: 28 | required: true 29 | attributes: 30 | label: Your contribution / 您的贡献 31 | description: | 32 | 33 | Your PR link or any other link you can help with. 34 | 您的PR链接或者其他您能提供帮助的链接。 -------------------------------------------------------------------------------- /.github/workflows/assign.yaml: -------------------------------------------------------------------------------- 1 | name: Assign 2 | on: 3 | issue_comment: 4 | types: created 5 | 6 | permissions: 7 | contents: read 8 | 9 | jobs: 10 | issue_assign: 11 | permissions: 12 | issues: write 13 | pull-requests: write 14 | runs-on: ubuntu-22.04 15 | steps: 16 | - if: github.event.comment.body == 'take' 17 | run: | 18 | echo "Assigning issue ${{ github.event.issue.number }} to ${{ github.event.comment.user.login }}" 19 | curl -H "Authorization: token ${{ secrets.GITHUB_TOKEN }}" -d '{"assignees": ["${{ github.event.comment.user.login }}"]}' https://api.github.com/repos/${{ github.repository }}/issues/${{ github.event.issue.number }}/assignees -------------------------------------------------------------------------------- /.github/workflows/issue.yaml: -------------------------------------------------------------------------------- 1 | name: Close inactive issues 2 | on: 3 | schedule: 4 | - cron: "0 19 * * *" 5 | workflow_dispatch: 6 | 7 | jobs: 8 | close-issues: 9 | runs-on: ubuntu-latest 10 | permissions: 11 | issues: write 12 | pull-requests: write 13 | steps: 14 | - uses: actions/stale@v9 15 | with: 16 | days-before-issue-stale: 7 17 | days-before-issue-close: 5 18 | stale-issue-label: "stale" 19 | stale-issue-message: "This issue is stale because it has been open for 7 days with no activity." 20 | close-issue-message: "This issue was closed because it has been inactive for 5 days since being marked as stale." 21 | days-before-pr-stale: -1 22 | days-before-pr-close: -1 23 | operations-per-run: 500 24 | repo-token: ${{ secrets.GITHUB_TOKEN }} 25 | -------------------------------------------------------------------------------- /.readthedocs.yaml: -------------------------------------------------------------------------------- 1 | version: 2 2 | 3 | # Build documentation in the docs/ directory with Sphinx 4 | sphinx: 5 | configuration: doc/source/conf.py 6 | 7 | build: 8 | os: ubuntu-20.04 9 | tools: 10 | python: "3.9" 11 | 12 | python: 13 | install: 14 | - method: pip 15 | path: . 16 | extra_requirements: 17 | - doc 18 | 19 | submodules: 20 | include: all 21 | recursive: true 22 | -------------------------------------------------------------------------------- /MANIFEST.in: -------------------------------------------------------------------------------- 1 | global-include *.pyx 2 | global-include *.pxd 3 | global-include xinference/**/*.json 4 | global-exclude *.c 5 | global-exclude *.cpp 6 | include setup.cfg 7 | include pyproject.toml 8 | global-exclude .DS_Store 9 | include versioneer.py 10 | include xinference/_version.py 11 | global-exclude conftest.py 12 | include xinference/locale/*.json 13 | include xinference/model/llm/*.json 14 | include xinference/model/embedding/*.json 15 | graft xinference/thirdparty 16 | global-include xinference/web/ui/build/**/* -------------------------------------------------------------------------------- /assets/demo.gif: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/assets/demo.gif -------------------------------------------------------------------------------- /assets/index.jpg: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/assets/index.jpg -------------------------------------------------------------------------------- /assets/screenshot.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/assets/screenshot.png -------------------------------------------------------------------------------- /assets/stay_ahead.gif: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/assets/stay_ahead.gif -------------------------------------------------------------------------------- /assets/xinference-downloading.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/assets/xinference-downloading.png -------------------------------------------------------------------------------- /assets/xorbits-logo.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/assets/xorbits-logo.png -------------------------------------------------------------------------------- /doc/source/_static/distributed_inference.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/doc/source/_static/distributed_inference.png -------------------------------------------------------------------------------- /doc/source/_static/switcher.json: -------------------------------------------------------------------------------- 1 | [ 2 | { 3 | "name": "简体中文(Chinese)", 4 | "version": "zh-cn", 5 | "url": "https://inference.readthedocs.io/zh-cn/latest/" 6 | }, 7 | { 8 | "name": "English", 9 | "version": "en", 10 | "url": "https://inference.readthedocs.io/en/latest/", 11 | "preferred": true 12 | } 13 | ] -------------------------------------------------------------------------------- /doc/source/development/index.rst: -------------------------------------------------------------------------------- 1 | .. _development_index: 2 | 3 | =========== 4 | Development 5 | =========== 6 | 7 | .. toctree:: 8 | :maxdepth: 2 9 | 10 | contributing_environment 11 | contributing_codebase 12 | xinference_internals 13 | -------------------------------------------------------------------------------- /doc/source/getting_started/index.rst: -------------------------------------------------------------------------------- 1 | .. _getting_started_index: 2 | 3 | =============== 4 | Getting Started 5 | =============== 6 | 7 | 8 | .. toctree:: 9 | :maxdepth: 2 10 | 11 | installation 12 | using_xinference 13 | logging 14 | using_docker_image 15 | using_kubernetes 16 | troubleshooting 17 | environments 18 | -------------------------------------------------------------------------------- /doc/source/locale/zh_CN/LC_MESSAGES/development/index.po: -------------------------------------------------------------------------------- 1 | # SOME DESCRIPTIVE TITLE. 2 | # Copyright (C) 2023, Xorbits Inc. 3 | # This file is distributed under the same license as the Xinference package. 4 | # FIRST AUTHOR , 2024. 5 | # 6 | #, fuzzy 7 | msgid "" 8 | msgstr "" 9 | "Project-Id-Version: Xinference \n" 10 | "Report-Msgid-Bugs-To: \n" 11 | "POT-Creation-Date: 2024-03-06 12:05+0800\n" 12 | "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" 13 | "Last-Translator: FULL NAME \n" 14 | "Language: zh_CN\n" 15 | "Language-Team: zh_CN \n" 16 | "Plural-Forms: nplurals=1; plural=0;\n" 17 | "MIME-Version: 1.0\n" 18 | "Content-Type: text/plain; charset=utf-8\n" 19 | "Content-Transfer-Encoding: 8bit\n" 20 | "Generated-By: Babel 2.14.0\n" 21 | 22 | #: ../../source/development/index.rst:5 23 | msgid "Development" 24 | msgstr "开发指南" 25 | 26 | -------------------------------------------------------------------------------- /doc/source/locale/zh_CN/LC_MESSAGES/getting_started.po: -------------------------------------------------------------------------------- 1 | # SOME DESCRIPTIVE TITLE. 2 | # Copyright (C) 2023, Xorbits Inc. 3 | # This file is distributed under the same license as the Xinference package. 4 | # FIRST AUTHOR , 2023. 5 | # 6 | #, fuzzy 7 | msgid "" 8 | msgstr "" 9 | "Project-Id-Version: Xinference \n" 10 | "Report-Msgid-Bugs-To: \n" 11 | "POT-Creation-Date: 2023-07-18 10:54+0800\n" 12 | "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" 13 | "Last-Translator: FULL NAME \n" 14 | "Language: zh_CN\n" 15 | "Language-Team: zh_CN \n" 16 | "Plural-Forms: nplurals=1; plural=0;\n" 17 | "MIME-Version: 1.0\n" 18 | "Content-Type: text/plain; charset=utf-8\n" 19 | "Content-Transfer-Encoding: 8bit\n" 20 | "Generated-By: Babel 2.12.1\n" 21 | 22 | #: ../../source/getting_started/index.rst:5 23 | msgid "Getting Started" 24 | msgstr "入门指南" 25 | 26 | -------------------------------------------------------------------------------- /doc/source/locale/zh_CN/LC_MESSAGES/getting_started/index.po: -------------------------------------------------------------------------------- 1 | # SOME DESCRIPTIVE TITLE. 2 | # Copyright (C) 2023, Xorbits Inc. 3 | # This file is distributed under the same license as the Xinference package. 4 | # FIRST AUTHOR , 2023. 5 | # 6 | #, fuzzy 7 | msgid "" 8 | msgstr "" 9 | "Project-Id-Version: Xinference \n" 10 | "Report-Msgid-Bugs-To: \n" 11 | "POT-Creation-Date: 2023-10-16 10:33+0800\n" 12 | "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" 13 | "Last-Translator: FULL NAME \n" 14 | "Language: zh_CN\n" 15 | "Language-Team: zh_CN \n" 16 | "Plural-Forms: nplurals=1; plural=0;\n" 17 | "MIME-Version: 1.0\n" 18 | "Content-Type: text/plain; charset=utf-8\n" 19 | "Content-Transfer-Encoding: 8bit\n" 20 | "Generated-By: Babel 2.12.1\n" 21 | 22 | #: ../../source/getting_started/index.rst:5 23 | msgid "Getting Started" 24 | msgstr "入门指南" 25 | 26 | -------------------------------------------------------------------------------- /doc/source/locale/zh_CN/LC_MESSAGES/models/builtin/audio/index.po: -------------------------------------------------------------------------------- 1 | # SOME DESCRIPTIVE TITLE. 2 | # Copyright (C) 2023, Xorbits Inc. 3 | # This file is distributed under the same license as the Xinference package. 4 | # FIRST AUTHOR , 2024. 5 | # 6 | #, fuzzy 7 | msgid "" 8 | msgstr "" 9 | "Project-Id-Version: Xinference \n" 10 | "Report-Msgid-Bugs-To: \n" 11 | "POT-Creation-Date: 2024-02-01 16:47+0800\n" 12 | "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" 13 | "Last-Translator: FULL NAME \n" 14 | "Language: zh_CN\n" 15 | "Language-Team: zh_CN \n" 16 | "Plural-Forms: nplurals=1; plural=0;\n" 17 | "MIME-Version: 1.0\n" 18 | "Content-Type: text/plain; charset=utf-8\n" 19 | "Content-Transfer-Encoding: 8bit\n" 20 | "Generated-By: Babel 2.13.1\n" 21 | 22 | #: ../../source/models/builtin/audio/index.rst:5 23 | msgid "Audio Models" 24 | msgstr "音频模型" 25 | 26 | #: ../../source/models/builtin/audio/index.rst:7 27 | msgid "The following is a list of built-in audio models in Xinference:" 28 | msgstr "以下是 Xinference 中内置的音频模型列表:" 29 | 30 | -------------------------------------------------------------------------------- /doc/source/locale/zh_CN/LC_MESSAGES/models/builtin/embedding/index.po: -------------------------------------------------------------------------------- 1 | # SOME DESCRIPTIVE TITLE. 2 | # Copyright (C) 2023, Xorbits Inc. 3 | # This file is distributed under the same license as the Xinference package. 4 | # FIRST AUTHOR , 2023. 5 | # 6 | #, fuzzy 7 | msgid "" 8 | msgstr "" 9 | "Project-Id-Version: Xinference \n" 10 | "Report-Msgid-Bugs-To: \n" 11 | "POT-Creation-Date: 2023-12-25 17:11+0800\n" 12 | "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" 13 | "Last-Translator: FULL NAME \n" 14 | "Language-Team: LANGUAGE \n" 15 | "MIME-Version: 1.0\n" 16 | "Content-Type: text/plain; charset=utf-8\n" 17 | "Content-Transfer-Encoding: 8bit\n" 18 | "Generated-By: Babel 2.11.0\n" 19 | 20 | #: ../../source/models/builtin/embedding/index.rst:5 21 | msgid "Embedding Models" 22 | msgstr "嵌入模型" 23 | 24 | #: ../../source/models/builtin/embedding/index.rst:7 25 | msgid "The following is a list of built-in embedding models in Xinference:" 26 | msgstr "以下是 Xinference 中内置的嵌入模型列表:" 27 | 28 | -------------------------------------------------------------------------------- /doc/source/locale/zh_CN/LC_MESSAGES/models/builtin/image/index.po: -------------------------------------------------------------------------------- 1 | # SOME DESCRIPTIVE TITLE. 2 | # Copyright (C) 2023, Xorbits Inc. 3 | # This file is distributed under the same license as the Xinference package. 4 | # FIRST AUTHOR , 2024. 5 | # 6 | #, fuzzy 7 | msgid "" 8 | msgstr "" 9 | "Project-Id-Version: Xinference \n" 10 | "Report-Msgid-Bugs-To: \n" 11 | "POT-Creation-Date: 2024-03-11 13:33+0800\n" 12 | "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" 13 | "Last-Translator: FULL NAME \n" 14 | "Language: zh_CN\n" 15 | "Language-Team: zh_CN \n" 16 | "Plural-Forms: nplurals=1; plural=0;\n" 17 | "MIME-Version: 1.0\n" 18 | "Content-Type: text/plain; charset=utf-8\n" 19 | "Content-Transfer-Encoding: 8bit\n" 20 | "Generated-By: Babel 2.14.0\n" 21 | 22 | #: ../../source/models/builtin/image/index.rst:5 23 | msgid "Image Models" 24 | msgstr "图像模型" 25 | 26 | #: ../../source/models/builtin/image/index.rst:7 27 | msgid "The following is a list of built-in image models in Xinference:" 28 | msgstr "以下是 Xinference 中内置的图像模型列表:" 29 | 30 | -------------------------------------------------------------------------------- /doc/source/locale/zh_CN/LC_MESSAGES/models/builtin/index.po: -------------------------------------------------------------------------------- 1 | # SOME DESCRIPTIVE TITLE. 2 | # Copyright (C) 2023, Xorbits Inc. 3 | # This file is distributed under the same license as the Xinference package. 4 | # FIRST AUTHOR , 2023. 5 | # 6 | #, fuzzy 7 | msgid "" 8 | msgstr "" 9 | "Project-Id-Version: Xinference \n" 10 | "Report-Msgid-Bugs-To: \n" 11 | "POT-Creation-Date: 2023-12-25 17:11+0800\n" 12 | "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" 13 | "Last-Translator: FULL NAME \n" 14 | "Language: zh_CN\n" 15 | "Language-Team: zh_CN \n" 16 | "Plural-Forms: nplurals=1; plural=0;\n" 17 | "MIME-Version: 1.0\n" 18 | "Content-Type: text/plain; charset=utf-8\n" 19 | "Content-Transfer-Encoding: 8bit\n" 20 | "Generated-By: Babel 2.11.0\n" 21 | 22 | #: ../../source/models/builtin/index.rst:5 23 | msgid "Builtin Models" 24 | msgstr "内置模型" 25 | -------------------------------------------------------------------------------- /doc/source/locale/zh_CN/LC_MESSAGES/models/builtin/rerank/index.po: -------------------------------------------------------------------------------- 1 | # SOME DESCRIPTIVE TITLE. 2 | # Copyright (C) 2023, Xorbits Inc. 3 | # This file is distributed under the same license as the Xinference package. 4 | # FIRST AUTHOR , 2023. 5 | # 6 | #, fuzzy 7 | msgid "" 8 | msgstr "" 9 | "Project-Id-Version: Xinference \n" 10 | "Report-Msgid-Bugs-To: \n" 11 | "POT-Creation-Date: 2023-12-25 17:11+0800\n" 12 | "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" 13 | "Last-Translator: FULL NAME \n" 14 | "Language-Team: LANGUAGE \n" 15 | "MIME-Version: 1.0\n" 16 | "Content-Type: text/plain; charset=utf-8\n" 17 | "Content-Transfer-Encoding: 8bit\n" 18 | "Generated-By: Babel 2.11.0\n" 19 | 20 | #: ../../source/models/builtin/rerank/index.rst:5 21 | msgid "Rerank Models" 22 | msgstr "重排序模型" 23 | 24 | #: ../../source/models/builtin/rerank/index.rst:7 25 | msgid "The following is a list of built-in rerank models in Xinference:" 26 | msgstr "以下是 Xinference 中内置的重排序模型列表:" 27 | 28 | -------------------------------------------------------------------------------- /doc/source/locale/zh_CN/LC_MESSAGES/models/builtin/video/index.po: -------------------------------------------------------------------------------- 1 | # SOME DESCRIPTIVE TITLE. 2 | # Copyright (C) 2023, Xorbits Inc. 3 | # This file is distributed under the same license as the Xinference package. 4 | # FIRST AUTHOR , 2024. 5 | # 6 | #, fuzzy 7 | msgid "" 8 | msgstr "" 9 | "Project-Id-Version: Xinference \n" 10 | "Report-Msgid-Bugs-To: \n" 11 | "POT-Creation-Date: 2024-08-13 17:44+0800\n" 12 | "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" 13 | "Last-Translator: FULL NAME \n" 14 | "Language: zh_CN\n" 15 | "Language-Team: zh_CN \n" 16 | "Plural-Forms: nplurals=1; plural=0;\n" 17 | "MIME-Version: 1.0\n" 18 | "Content-Type: text/plain; charset=utf-8\n" 19 | "Content-Transfer-Encoding: 8bit\n" 20 | "Generated-By: Babel 2.14.0\n" 21 | 22 | #: ../../source/models/builtin/video/index.rst:5 23 | msgid "Video Models" 24 | msgstr "视频模型" 25 | 26 | #: ../../source/models/builtin/video/index.rst:7 27 | msgid "The following is a list of built-in video models in Xinference:" 28 | msgstr "以下是 Xinference 中内置的视频模型列表:" 29 | 30 | -------------------------------------------------------------------------------- /doc/source/locale/zh_CN/LC_MESSAGES/models/model_abilities/index.po: -------------------------------------------------------------------------------- 1 | # SOME DESCRIPTIVE TITLE. 2 | # Copyright (C) 2023, Xorbits Inc. 3 | # This file is distributed under the same license as the Xinference package. 4 | # FIRST AUTHOR , 2024. 5 | # 6 | #, fuzzy 7 | msgid "" 8 | msgstr "" 9 | "Project-Id-Version: Xinference \n" 10 | "Report-Msgid-Bugs-To: \n" 11 | "POT-Creation-Date: 2024-02-01 16:47+0800\n" 12 | "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" 13 | "Last-Translator: FULL NAME \n" 14 | "Language: zh_CN\n" 15 | "Language-Team: zh_CN \n" 16 | "Plural-Forms: nplurals=1; plural=0;\n" 17 | "MIME-Version: 1.0\n" 18 | "Content-Type: text/plain; charset=utf-8\n" 19 | "Content-Transfer-Encoding: 8bit\n" 20 | "Generated-By: Babel 2.13.1\n" 21 | 22 | #: ../../source/models/model_abilities/index.rst:5 23 | msgid "Model Abilities" 24 | msgstr "模型能力" 25 | 26 | -------------------------------------------------------------------------------- /doc/source/locale/zh_CN/LC_MESSAGES/reference.po: -------------------------------------------------------------------------------- 1 | # SOME DESCRIPTIVE TITLE. 2 | # Copyright (C) 2023, Xorbits Inc. 3 | # This file is distributed under the same license as the Xinference package. 4 | # FIRST AUTHOR , 2023. 5 | # 6 | #, fuzzy 7 | msgid "" 8 | msgstr "" 9 | "Project-Id-Version: Xinference \n" 10 | "Report-Msgid-Bugs-To: \n" 11 | "POT-Creation-Date: 2023-07-18 10:54+0800\n" 12 | "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" 13 | "Last-Translator: FULL NAME \n" 14 | "Language: zh_CN\n" 15 | "Language-Team: zh_CN \n" 16 | "Plural-Forms: nplurals=1; plural=0;\n" 17 | "MIME-Version: 1.0\n" 18 | "Content-Type: text/plain; charset=utf-8\n" 19 | "Content-Transfer-Encoding: 8bit\n" 20 | "Generated-By: Babel 2.12.1\n" 21 | 22 | #: ../../source/reference/index.rst:5 23 | msgid "API Reference" 24 | msgstr "API 指南" 25 | 26 | -------------------------------------------------------------------------------- /doc/source/locale/zh_CN/LC_MESSAGES/user_guide/cache_management.po: -------------------------------------------------------------------------------- 1 | # SOME DESCRIPTIVE TITLE. 2 | # Copyright (C) 2023, Xorbits Inc. 3 | # This file is distributed under the same license as the Xinference package. 4 | # FIRST AUTHOR , 2023. 5 | # 6 | #, fuzzy 7 | msgid "" 8 | msgstr "" 9 | "Project-Id-Version: Xinference \n" 10 | "Report-Msgid-Bugs-To: \n" 11 | "POT-Creation-Date: 2023-10-16 10:33+0800\n" 12 | "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" 13 | "Last-Translator: FULL NAME \n" 14 | "Language: zh_CN\n" 15 | "Language-Team: zh_CN \n" 16 | "Plural-Forms: nplurals=1; plural=0;\n" 17 | "MIME-Version: 1.0\n" 18 | "Content-Type: text/plain; charset=utf-8\n" 19 | "Content-Transfer-Encoding: 8bit\n" 20 | "Generated-By: Babel 2.12.1\n" 21 | 22 | #: ../../source/user_guide/cache_management.rst:5 23 | msgid "Cache Management" 24 | msgstr "缓存管理" 25 | 26 | -------------------------------------------------------------------------------- /doc/source/locale/zh_CN/LC_MESSAGES/user_guide/index.po: -------------------------------------------------------------------------------- 1 | # SOME DESCRIPTIVE TITLE. 2 | # Copyright (C) 2023, Xorbits Inc. 3 | # This file is distributed under the same license as the Xinference package. 4 | # FIRST AUTHOR , 2023. 5 | # 6 | #, fuzzy 7 | msgid "" 8 | msgstr "" 9 | "Project-Id-Version: Xinference \n" 10 | "Report-Msgid-Bugs-To: \n" 11 | "POT-Creation-Date: 2023-10-16 10:33+0800\n" 12 | "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" 13 | "Last-Translator: FULL NAME \n" 14 | "Language: zh_CN\n" 15 | "Language-Team: zh_CN \n" 16 | "Plural-Forms: nplurals=1; plural=0;\n" 17 | "MIME-Version: 1.0\n" 18 | "Content-Type: text/plain; charset=utf-8\n" 19 | "Content-Transfer-Encoding: 8bit\n" 20 | "Generated-By: Babel 2.12.1\n" 21 | 22 | #: ../../source/user_guide/index.rst:5 23 | msgid "User Guide" 24 | msgstr "用户指南" 25 | 26 | -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/belle-distilwhisper-large-v2-zh.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_belle-distilwhisper-large-v2-zh: 2 | 3 | =============================== 4 | Belle-distilwhisper-large-v2-zh 5 | =============================== 6 | 7 | - **Model Name:** Belle-distilwhisper-large-v2-zh 8 | - **Model Family:** whisper 9 | - **Abilities:** ['audio2text'] 10 | - **Multilingual:** False 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** BELLE-2/Belle-distilwhisper-large-v2-zh 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name Belle-distilwhisper-large-v2-zh --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/belle-whisper-large-v2-zh.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_belle-whisper-large-v2-zh: 2 | 3 | ========================= 4 | Belle-whisper-large-v2-zh 5 | ========================= 6 | 7 | - **Model Name:** Belle-whisper-large-v2-zh 8 | - **Model Family:** whisper 9 | - **Abilities:** ['audio2text'] 10 | - **Multilingual:** False 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** BELLE-2/Belle-whisper-large-v2-zh 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name Belle-whisper-large-v2-zh --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/belle-whisper-large-v3-zh.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_belle-whisper-large-v3-zh: 2 | 3 | ========================= 4 | Belle-whisper-large-v3-zh 5 | ========================= 6 | 7 | - **Model Name:** Belle-whisper-large-v3-zh 8 | - **Model Family:** whisper 9 | - **Abilities:** ['audio2text'] 10 | - **Multilingual:** False 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** BELLE-2/Belle-whisper-large-v3-zh 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name Belle-whisper-large-v3-zh --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/chattts.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_chattts: 2 | 3 | ======= 4 | ChatTTS 5 | ======= 6 | 7 | - **Model Name:** ChatTTS 8 | - **Model Family:** ChatTTS 9 | - **Abilities:** ['text2audio'] 10 | - **Multilingual:** True 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** 2Noise/ChatTTS 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name ChatTTS --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/cosyvoice-300m-instruct.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_cosyvoice-300m-instruct: 2 | 3 | ======================= 4 | CosyVoice-300M-Instruct 5 | ======================= 6 | 7 | - **Model Name:** CosyVoice-300M-Instruct 8 | - **Model Family:** CosyVoice 9 | - **Abilities:** ['text2audio'] 10 | - **Multilingual:** True 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** FunAudioLLM/CosyVoice-300M-Instruct 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name CosyVoice-300M-Instruct --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/cosyvoice-300m-sft.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_cosyvoice-300m-sft: 2 | 3 | ================== 4 | CosyVoice-300M-SFT 5 | ================== 6 | 7 | - **Model Name:** CosyVoice-300M-SFT 8 | - **Model Family:** CosyVoice 9 | - **Abilities:** ['text2audio'] 10 | - **Multilingual:** True 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** FunAudioLLM/CosyVoice-300M-SFT 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name CosyVoice-300M-SFT --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/cosyvoice-300m.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_cosyvoice-300m: 2 | 3 | ============== 4 | CosyVoice-300M 5 | ============== 6 | 7 | - **Model Name:** CosyVoice-300M 8 | - **Model Family:** CosyVoice 9 | - **Abilities:** ['text2audio'] 10 | - **Multilingual:** True 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** FunAudioLLM/CosyVoice-300M 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name CosyVoice-300M --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/cosyvoice2-0.5b.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_cosyvoice2-0.5b: 2 | 3 | =============== 4 | CosyVoice2-0.5B 5 | =============== 6 | 7 | - **Model Name:** CosyVoice2-0.5B 8 | - **Model Family:** CosyVoice 9 | - **Abilities:** ['text2audio'] 10 | - **Multilingual:** True 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** JunHowie/CosyVoice2-0.5B 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name CosyVoice2-0.5B --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/f5-tts-mlx.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_f5-tts-mlx: 2 | 3 | ========== 4 | F5-TTS-MLX 5 | ========== 6 | 7 | - **Model Name:** F5-TTS-MLX 8 | - **Model Family:** F5-TTS-MLX 9 | - **Abilities:** ['text2audio'] 10 | - **Multilingual:** True 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** lucasnewman/f5-tts-mlx 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name F5-TTS-MLX --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/f5-tts.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_f5-tts: 2 | 3 | ====== 4 | F5-TTS 5 | ====== 6 | 7 | - **Model Name:** F5-TTS 8 | - **Model Family:** F5-TTS 9 | - **Abilities:** ['text2audio'] 10 | - **Multilingual:** True 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** SWivid/F5-TTS 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name F5-TTS --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/fishspeech-1.5.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_fishspeech-1.5: 2 | 3 | ============== 4 | FishSpeech-1.5 5 | ============== 6 | 7 | - **Model Name:** FishSpeech-1.5 8 | - **Model Family:** FishAudio 9 | - **Abilities:** ['text2audio'] 10 | - **Multilingual:** True 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** fishaudio/fish-speech-1.5 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name FishSpeech-1.5 --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/kokoro-82m.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_kokoro-82m: 2 | 3 | ========== 4 | Kokoro-82M 5 | ========== 6 | 7 | - **Model Name:** Kokoro-82M 8 | - **Model Family:** Kokoro 9 | - **Abilities:** ['text2audio'] 10 | - **Multilingual:** True 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** hexgrad/Kokoro-82M 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name Kokoro-82M --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/megatts3.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_megatts3: 2 | 3 | ======== 4 | MegaTTS3 5 | ======== 6 | 7 | - **Model Name:** MegaTTS3 8 | - **Model Family:** MegaTTS 9 | - **Abilities:** ['text2audio'] 10 | - **Multilingual:** True 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** ByteDance/MegaTTS3 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name MegaTTS3 --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/melotts-chinese.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_melotts-chinese: 2 | 3 | =============== 4 | MeloTTS-Chinese 5 | =============== 6 | 7 | - **Model Name:** MeloTTS-Chinese 8 | - **Model Family:** MeloTTS 9 | - **Abilities:** ['text2audio'] 10 | - **Multilingual:** False 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** myshell-ai/MeloTTS-Chinese 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name MeloTTS-Chinese --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/melotts-english-v2.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_melotts-english-v2: 2 | 3 | ================== 4 | MeloTTS-English-v2 5 | ================== 6 | 7 | - **Model Name:** MeloTTS-English-v2 8 | - **Model Family:** MeloTTS 9 | - **Abilities:** ['text2audio'] 10 | - **Multilingual:** False 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** myshell-ai/MeloTTS-English-v2 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name MeloTTS-English-v2 --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/melotts-english-v3.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_melotts-english-v3: 2 | 3 | ================== 4 | MeloTTS-English-v3 5 | ================== 6 | 7 | - **Model Name:** MeloTTS-English-v3 8 | - **Model Family:** MeloTTS 9 | - **Abilities:** ['text2audio'] 10 | - **Multilingual:** False 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** myshell-ai/MeloTTS-English-v3 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name MeloTTS-English-v3 --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/melotts-english.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_melotts-english: 2 | 3 | =============== 4 | MeloTTS-English 5 | =============== 6 | 7 | - **Model Name:** MeloTTS-English 8 | - **Model Family:** MeloTTS 9 | - **Abilities:** ['text2audio'] 10 | - **Multilingual:** False 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** myshell-ai/MeloTTS-English 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name MeloTTS-English --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/melotts-french.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_melotts-french: 2 | 3 | ============== 4 | MeloTTS-French 5 | ============== 6 | 7 | - **Model Name:** MeloTTS-French 8 | - **Model Family:** MeloTTS 9 | - **Abilities:** ['text2audio'] 10 | - **Multilingual:** False 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** myshell-ai/MeloTTS-French 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name MeloTTS-French --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/melotts-japanese.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_melotts-japanese: 2 | 3 | ================ 4 | MeloTTS-Japanese 5 | ================ 6 | 7 | - **Model Name:** MeloTTS-Japanese 8 | - **Model Family:** MeloTTS 9 | - **Abilities:** ['text2audio'] 10 | - **Multilingual:** False 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** myshell-ai/MeloTTS-Japanese 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name MeloTTS-Japanese --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/melotts-korean.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_melotts-korean: 2 | 3 | ============== 4 | MeloTTS-Korean 5 | ============== 6 | 7 | - **Model Name:** MeloTTS-Korean 8 | - **Model Family:** MeloTTS 9 | - **Abilities:** ['text2audio'] 10 | - **Multilingual:** False 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** myshell-ai/MeloTTS-Korean 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name MeloTTS-Korean --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/melotts-spanish.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_melotts-spanish: 2 | 3 | =============== 4 | MeloTTS-Spanish 5 | =============== 6 | 7 | - **Model Name:** MeloTTS-Spanish 8 | - **Model Family:** MeloTTS 9 | - **Abilities:** ['text2audio'] 10 | - **Multilingual:** False 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** myshell-ai/MeloTTS-Spanish 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name MeloTTS-Spanish --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/paraformer-zh-hotword.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_paraformer-zh-hotword: 2 | 3 | ===================== 4 | paraformer-zh-hotword 5 | ===================== 6 | 7 | - **Model Name:** paraformer-zh-hotword 8 | - **Model Family:** funasr 9 | - **Abilities:** ['audio2text'] 10 | - **Multilingual:** False 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** JunHowie/speech_paraformer-large-contextual_asr_nat-zh-cn-16k-common-vocab8404 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name paraformer-zh-hotword --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/paraformer-zh-long.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_paraformer-zh-long: 2 | 3 | ================== 4 | paraformer-zh-long 5 | ================== 6 | 7 | - **Model Name:** paraformer-zh-long 8 | - **Model Family:** funasr 9 | - **Abilities:** ['audio2text'] 10 | - **Multilingual:** False 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** JunHowie/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name paraformer-zh-long --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/paraformer-zh-spk.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_paraformer-zh-spk: 2 | 3 | ================= 4 | paraformer-zh-spk 5 | ================= 6 | 7 | - **Model Name:** paraformer-zh-spk 8 | - **Model Family:** funasr 9 | - **Abilities:** ['audio2text'] 10 | - **Multilingual:** False 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** JunHowie/speech_paraformer-large-vad-punc-spk_asr_nat-zh-cn 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name paraformer-zh-spk --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/paraformer-zh.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_paraformer-zh: 2 | 3 | ============= 4 | paraformer-zh 5 | ============= 6 | 7 | - **Model Name:** paraformer-zh 8 | - **Model Family:** funasr 9 | - **Abilities:** ['audio2text'] 10 | - **Multilingual:** False 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** funasr/paraformer-zh 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name paraformer-zh --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/sensevoicesmall.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_sensevoicesmall: 2 | 3 | =============== 4 | SenseVoiceSmall 5 | =============== 6 | 7 | - **Model Name:** SenseVoiceSmall 8 | - **Model Family:** funasr 9 | - **Abilities:** ['audio2text'] 10 | - **Multilingual:** True 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** FunAudioLLM/SenseVoiceSmall 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name SenseVoiceSmall --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/whisper-base-mlx.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_whisper-base-mlx: 2 | 3 | ================ 4 | whisper-base-mlx 5 | ================ 6 | 7 | - **Model Name:** whisper-base-mlx 8 | - **Model Family:** whisper 9 | - **Abilities:** ['audio2text'] 10 | - **Multilingual:** True 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** mlx-community/whisper-base-mlx 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name whisper-base-mlx --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/whisper-base.en-mlx.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_whisper-base.en-mlx: 2 | 3 | =================== 4 | whisper-base.en-mlx 5 | =================== 6 | 7 | - **Model Name:** whisper-base.en-mlx 8 | - **Model Family:** whisper 9 | - **Abilities:** ['audio2text'] 10 | - **Multilingual:** False 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** mlx-community/whisper-base.en-mlx 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name whisper-base.en-mlx --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/whisper-base.en.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_whisper-base.en: 2 | 3 | =============== 4 | whisper-base.en 5 | =============== 6 | 7 | - **Model Name:** whisper-base.en 8 | - **Model Family:** whisper 9 | - **Abilities:** ['audio2text'] 10 | - **Multilingual:** False 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** openai/whisper-base.en 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name whisper-base.en --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/whisper-base.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_whisper-base: 2 | 3 | ============ 4 | whisper-base 5 | ============ 6 | 7 | - **Model Name:** whisper-base 8 | - **Model Family:** whisper 9 | - **Abilities:** ['audio2text'] 10 | - **Multilingual:** True 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** openai/whisper-base 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name whisper-base --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/whisper-large-v3-mlx.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_whisper-large-v3-mlx: 2 | 3 | ==================== 4 | whisper-large-v3-mlx 5 | ==================== 6 | 7 | - **Model Name:** whisper-large-v3-mlx 8 | - **Model Family:** whisper 9 | - **Abilities:** ['audio2text'] 10 | - **Multilingual:** True 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** mlx-community/whisper-large-v3-mlx 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name whisper-large-v3-mlx --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/whisper-large-v3-turbo-mlx.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_whisper-large-v3-turbo-mlx: 2 | 3 | ========================== 4 | whisper-large-v3-turbo-mlx 5 | ========================== 6 | 7 | - **Model Name:** whisper-large-v3-turbo-mlx 8 | - **Model Family:** whisper 9 | - **Abilities:** ['audio2text'] 10 | - **Multilingual:** True 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** mlx-community/whisper-large-v3-turbo 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name whisper-large-v3-turbo-mlx --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/whisper-large-v3-turbo.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_whisper-large-v3-turbo: 2 | 3 | ====================== 4 | whisper-large-v3-turbo 5 | ====================== 6 | 7 | - **Model Name:** whisper-large-v3-turbo 8 | - **Model Family:** whisper 9 | - **Abilities:** ['audio2text'] 10 | - **Multilingual:** True 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** openai/whisper-large-v3-turbo 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name whisper-large-v3-turbo --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/whisper-large-v3.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_whisper-large-v3: 2 | 3 | ================ 4 | whisper-large-v3 5 | ================ 6 | 7 | - **Model Name:** whisper-large-v3 8 | - **Model Family:** whisper 9 | - **Abilities:** ['audio2text'] 10 | - **Multilingual:** True 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** openai/whisper-large-v3 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name whisper-large-v3 --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/whisper-medium-mlx.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_whisper-medium-mlx: 2 | 3 | ================== 4 | whisper-medium-mlx 5 | ================== 6 | 7 | - **Model Name:** whisper-medium-mlx 8 | - **Model Family:** whisper 9 | - **Abilities:** ['audio2text'] 10 | - **Multilingual:** True 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** mlx-community/whisper-medium-mlx 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name whisper-medium-mlx --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/whisper-medium.en-mlx.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_whisper-medium.en-mlx: 2 | 3 | ===================== 4 | whisper-medium.en-mlx 5 | ===================== 6 | 7 | - **Model Name:** whisper-medium.en-mlx 8 | - **Model Family:** whisper 9 | - **Abilities:** ['audio2text'] 10 | - **Multilingual:** False 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** mlx-community/whisper-medium.en-mlx 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name whisper-medium.en-mlx --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/whisper-medium.en.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_whisper-medium.en: 2 | 3 | ================= 4 | whisper-medium.en 5 | ================= 6 | 7 | - **Model Name:** whisper-medium.en 8 | - **Model Family:** whisper 9 | - **Abilities:** ['audio2text'] 10 | - **Multilingual:** False 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** openai/whisper-medium.en 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name whisper-medium.en --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/whisper-medium.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_whisper-medium: 2 | 3 | ============== 4 | whisper-medium 5 | ============== 6 | 7 | - **Model Name:** whisper-medium 8 | - **Model Family:** whisper 9 | - **Abilities:** ['audio2text'] 10 | - **Multilingual:** True 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** openai/whisper-medium 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name whisper-medium --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/whisper-small-mlx.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_whisper-small-mlx: 2 | 3 | ================= 4 | whisper-small-mlx 5 | ================= 6 | 7 | - **Model Name:** whisper-small-mlx 8 | - **Model Family:** whisper 9 | - **Abilities:** ['audio2text'] 10 | - **Multilingual:** True 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** mlx-community/whisper-small-mlx 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name whisper-small-mlx --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/whisper-small.en-mlx.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_whisper-small.en-mlx: 2 | 3 | ==================== 4 | whisper-small.en-mlx 5 | ==================== 6 | 7 | - **Model Name:** whisper-small.en-mlx 8 | - **Model Family:** whisper 9 | - **Abilities:** ['audio2text'] 10 | - **Multilingual:** False 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** mlx-community/whisper-small.en-mlx 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name whisper-small.en-mlx --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/whisper-small.en.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_whisper-small.en: 2 | 3 | ================ 4 | whisper-small.en 5 | ================ 6 | 7 | - **Model Name:** whisper-small.en 8 | - **Model Family:** whisper 9 | - **Abilities:** ['audio2text'] 10 | - **Multilingual:** False 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** openai/whisper-small.en 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name whisper-small.en --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/whisper-small.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_whisper-small: 2 | 3 | ============= 4 | whisper-small 5 | ============= 6 | 7 | - **Model Name:** whisper-small 8 | - **Model Family:** whisper 9 | - **Abilities:** ['audio2text'] 10 | - **Multilingual:** True 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** openai/whisper-small 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name whisper-small --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/whisper-tiny-mlx.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_whisper-tiny-mlx: 2 | 3 | ================ 4 | whisper-tiny-mlx 5 | ================ 6 | 7 | - **Model Name:** whisper-tiny-mlx 8 | - **Model Family:** whisper 9 | - **Abilities:** ['audio2text'] 10 | - **Multilingual:** True 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** mlx-community/whisper-tiny 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name whisper-tiny-mlx --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/whisper-tiny.en-mlx.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_whisper-tiny.en-mlx: 2 | 3 | =================== 4 | whisper-tiny.en-mlx 5 | =================== 6 | 7 | - **Model Name:** whisper-tiny.en-mlx 8 | - **Model Family:** whisper 9 | - **Abilities:** ['audio2text'] 10 | - **Multilingual:** False 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** mlx-community/whisper-tiny.en-mlx 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name whisper-tiny.en-mlx --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/whisper-tiny.en.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_whisper-tiny.en: 2 | 3 | =============== 4 | whisper-tiny.en 5 | =============== 6 | 7 | - **Model Name:** whisper-tiny.en 8 | - **Model Family:** whisper 9 | - **Abilities:** ['audio2text'] 10 | - **Multilingual:** False 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** openai/whisper-tiny.en 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name whisper-tiny.en --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/audio/whisper-tiny.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_whisper-tiny: 2 | 3 | ============ 4 | whisper-tiny 5 | ============ 6 | 7 | - **Model Name:** whisper-tiny 8 | - **Model Family:** whisper 9 | - **Abilities:** ['audio2text'] 10 | - **Multilingual:** True 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** openai/whisper-tiny 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name whisper-tiny --model-type audio -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/bce-embedding-base_v1.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_bce-embedding-base_v1: 2 | 3 | ===================== 4 | bce-embedding-base_v1 5 | ===================== 6 | 7 | - **Model Name:** bce-embedding-base_v1 8 | - **Languages:** zh, en 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 768 15 | - **Max Tokens:** 512 16 | - **Model ID:** maidalun1020/bce-embedding-base_v1 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name bce-embedding-base_v1 --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/bge-base-en-v1.5.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_bge-base-en-v1.5: 2 | 3 | ================ 4 | bge-base-en-v1.5 5 | ================ 6 | 7 | - **Model Name:** bge-base-en-v1.5 8 | - **Languages:** en 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 768 15 | - **Max Tokens:** 512 16 | - **Model ID:** BAAI/bge-base-en-v1.5 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name bge-base-en-v1.5 --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/bge-base-en.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_bge-base-en: 2 | 3 | =========== 4 | bge-base-en 5 | =========== 6 | 7 | - **Model Name:** bge-base-en 8 | - **Languages:** en 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 768 15 | - **Max Tokens:** 512 16 | - **Model ID:** BAAI/bge-base-en 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name bge-base-en --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/bge-base-zh-v1.5.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_bge-base-zh-v1.5: 2 | 3 | ================ 4 | bge-base-zh-v1.5 5 | ================ 6 | 7 | - **Model Name:** bge-base-zh-v1.5 8 | - **Languages:** zh 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 768 15 | - **Max Tokens:** 512 16 | - **Model ID:** BAAI/bge-base-zh-v1.5 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name bge-base-zh-v1.5 --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/bge-base-zh.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_bge-base-zh: 2 | 3 | =========== 4 | bge-base-zh 5 | =========== 6 | 7 | - **Model Name:** bge-base-zh 8 | - **Languages:** zh 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 768 15 | - **Max Tokens:** 512 16 | - **Model ID:** BAAI/bge-base-zh 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name bge-base-zh --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/bge-large-en-v1.5.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_bge-large-en-v1.5: 2 | 3 | ================= 4 | bge-large-en-v1.5 5 | ================= 6 | 7 | - **Model Name:** bge-large-en-v1.5 8 | - **Languages:** en 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 1024 15 | - **Max Tokens:** 512 16 | - **Model ID:** BAAI/bge-large-en-v1.5 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name bge-large-en-v1.5 --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/bge-large-en.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_bge-large-en: 2 | 3 | ============ 4 | bge-large-en 5 | ============ 6 | 7 | - **Model Name:** bge-large-en 8 | - **Languages:** en 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 1024 15 | - **Max Tokens:** 512 16 | - **Model ID:** BAAI/bge-large-en 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name bge-large-en --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/bge-large-zh-noinstruct.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_bge-large-zh-noinstruct: 2 | 3 | ======================= 4 | bge-large-zh-noinstruct 5 | ======================= 6 | 7 | - **Model Name:** bge-large-zh-noinstruct 8 | - **Languages:** zh 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 1024 15 | - **Max Tokens:** 512 16 | - **Model ID:** BAAI/bge-large-zh-noinstruct 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name bge-large-zh-noinstruct --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/bge-large-zh-v1.5.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_bge-large-zh-v1.5: 2 | 3 | ================= 4 | bge-large-zh-v1.5 5 | ================= 6 | 7 | - **Model Name:** bge-large-zh-v1.5 8 | - **Languages:** zh 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 1024 15 | - **Max Tokens:** 512 16 | - **Model ID:** BAAI/bge-large-zh-v1.5 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name bge-large-zh-v1.5 --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/bge-large-zh.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_bge-large-zh: 2 | 3 | ============ 4 | bge-large-zh 5 | ============ 6 | 7 | - **Model Name:** bge-large-zh 8 | - **Languages:** zh 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 1024 15 | - **Max Tokens:** 512 16 | - **Model ID:** BAAI/bge-large-zh 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name bge-large-zh --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/bge-m3.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_bge-m3: 2 | 3 | ====== 4 | bge-m3 5 | ====== 6 | 7 | - **Model Name:** bge-m3 8 | - **Languages:** zh, en 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 1024 15 | - **Max Tokens:** 8192 16 | - **Model ID:** BAAI/bge-m3 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name bge-m3 --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/bge-small-en-v1.5.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_bge-small-en-v1.5: 2 | 3 | ================= 4 | bge-small-en-v1.5 5 | ================= 6 | 7 | - **Model Name:** bge-small-en-v1.5 8 | - **Languages:** en 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 384 15 | - **Max Tokens:** 512 16 | - **Model ID:** BAAI/bge-small-en-v1.5 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name bge-small-en-v1.5 --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/bge-small-zh-v1.5.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_bge-small-zh-v1.5: 2 | 3 | ================= 4 | bge-small-zh-v1.5 5 | ================= 6 | 7 | - **Model Name:** bge-small-zh-v1.5 8 | - **Languages:** zh 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 512 15 | - **Max Tokens:** 512 16 | - **Model ID:** BAAI/bge-small-zh-v1.5 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name bge-small-zh-v1.5 --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/bge-small-zh.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_bge-small-zh: 2 | 3 | ============ 4 | bge-small-zh 5 | ============ 6 | 7 | - **Model Name:** bge-small-zh 8 | - **Languages:** zh 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 512 15 | - **Max Tokens:** 512 16 | - **Model ID:** BAAI/bge-small-zh 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name bge-small-zh --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/e5-large-v2.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_e5-large-v2: 2 | 3 | =========== 4 | e5-large-v2 5 | =========== 6 | 7 | - **Model Name:** e5-large-v2 8 | - **Languages:** en 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 1024 15 | - **Max Tokens:** 512 16 | - **Model ID:** intfloat/e5-large-v2 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name e5-large-v2 --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/gte-base.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_gte-base: 2 | 3 | ======== 4 | gte-base 5 | ======== 6 | 7 | - **Model Name:** gte-base 8 | - **Languages:** en 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 768 15 | - **Max Tokens:** 512 16 | - **Model ID:** thenlper/gte-base 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name gte-base --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/gte-large.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_gte-large: 2 | 3 | ========= 4 | gte-large 5 | ========= 6 | 7 | - **Model Name:** gte-large 8 | - **Languages:** en 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 1024 15 | - **Max Tokens:** 512 16 | - **Model ID:** thenlper/gte-large 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name gte-large --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/gte-qwen2.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_gte-qwen2: 2 | 3 | ========= 4 | gte-Qwen2 5 | ========= 6 | 7 | - **Model Name:** gte-Qwen2 8 | - **Languages:** zh, en 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 3584 15 | - **Max Tokens:** 32000 16 | - **Model ID:** Alibaba-NLP/gte-Qwen2-7B-instruct 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name gte-Qwen2 --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/jina-clip-v2.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_jina-clip-v2: 2 | 3 | ============ 4 | jina-clip-v2 5 | ============ 6 | 7 | - **Model Name:** jina-clip-v2 8 | - **Languages:** 89 languages supported 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 1024 15 | - **Max Tokens:** 8192 16 | - **Model ID:** jinaai/jina-clip-v2 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name jina-clip-v2 --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/jina-embeddings-v2-base-en.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_jina-embeddings-v2-base-en: 2 | 3 | ========================== 4 | jina-embeddings-v2-base-en 5 | ========================== 6 | 7 | - **Model Name:** jina-embeddings-v2-base-en 8 | - **Languages:** en 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 512 15 | - **Max Tokens:** 8192 16 | - **Model ID:** jinaai/jina-embeddings-v2-base-en 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name jina-embeddings-v2-base-en --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/jina-embeddings-v2-base-zh.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_jina-embeddings-v2-base-zh: 2 | 3 | ========================== 4 | jina-embeddings-v2-base-zh 5 | ========================== 6 | 7 | - **Model Name:** jina-embeddings-v2-base-zh 8 | - **Languages:** zh, en 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 768 15 | - **Max Tokens:** 8192 16 | - **Model ID:** jinaai/jina-embeddings-v2-base-zh 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name jina-embeddings-v2-base-zh --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/jina-embeddings-v2-small-en.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_jina-embeddings-v2-small-en: 2 | 3 | =========================== 4 | jina-embeddings-v2-small-en 5 | =========================== 6 | 7 | - **Model Name:** jina-embeddings-v2-small-en 8 | - **Languages:** en 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 512 15 | - **Max Tokens:** 8192 16 | - **Model ID:** jinaai/jina-embeddings-v2-small-en 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name jina-embeddings-v2-small-en --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/jina-embeddings-v3.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_jina-embeddings-v3: 2 | 3 | ================== 4 | jina-embeddings-v3 5 | ================== 6 | 7 | - **Model Name:** jina-embeddings-v3 8 | - **Languages:** zh, en 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 1024 15 | - **Max Tokens:** 8192 16 | - **Model ID:** jinaai/jina-embeddings-v3 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name jina-embeddings-v3 --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/m3e-base.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_m3e-base: 2 | 3 | ======== 4 | m3e-base 5 | ======== 6 | 7 | - **Model Name:** m3e-base 8 | - **Languages:** zh, en 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 768 15 | - **Max Tokens:** 512 16 | - **Model ID:** moka-ai/m3e-base 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name m3e-base --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/m3e-large.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_m3e-large: 2 | 3 | ========= 4 | m3e-large 5 | ========= 6 | 7 | - **Model Name:** m3e-large 8 | - **Languages:** zh, en 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 1024 15 | - **Max Tokens:** 512 16 | - **Model ID:** moka-ai/m3e-large 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name m3e-large --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/m3e-small.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_m3e-small: 2 | 3 | ========= 4 | m3e-small 5 | ========= 6 | 7 | - **Model Name:** m3e-small 8 | - **Languages:** zh, en 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 512 15 | - **Max Tokens:** 512 16 | - **Model ID:** moka-ai/m3e-small 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name m3e-small --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/multilingual-e5-large.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_multilingual-e5-large: 2 | 3 | ===================== 4 | multilingual-e5-large 5 | ===================== 6 | 7 | - **Model Name:** multilingual-e5-large 8 | - **Languages:** zh 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 1024 15 | - **Max Tokens:** 514 16 | - **Model ID:** intfloat/multilingual-e5-large 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name multilingual-e5-large --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/text2vec-base-chinese-paraphrase.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_text2vec-base-chinese-paraphrase: 2 | 3 | ================================ 4 | text2vec-base-chinese-paraphrase 5 | ================================ 6 | 7 | - **Model Name:** text2vec-base-chinese-paraphrase 8 | - **Languages:** zh 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 768 15 | - **Max Tokens:** 256 16 | - **Model ID:** shibing624/text2vec-base-chinese-paraphrase 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name text2vec-base-chinese-paraphrase --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/text2vec-base-chinese-sentence.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_text2vec-base-chinese-sentence: 2 | 3 | ============================== 4 | text2vec-base-chinese-sentence 5 | ============================== 6 | 7 | - **Model Name:** text2vec-base-chinese-sentence 8 | - **Languages:** zh 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 768 15 | - **Max Tokens:** 256 16 | - **Model ID:** shibing624/text2vec-base-chinese-sentence 17 | - **Model Hubs**: `Hugging Face `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name text2vec-base-chinese-sentence --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/text2vec-base-chinese.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_text2vec-base-chinese: 2 | 3 | ===================== 4 | text2vec-base-chinese 5 | ===================== 6 | 7 | - **Model Name:** text2vec-base-chinese 8 | - **Languages:** zh 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 768 15 | - **Max Tokens:** 128 16 | - **Model ID:** shibing624/text2vec-base-chinese 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name text2vec-base-chinese --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/text2vec-base-multilingual.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_text2vec-base-multilingual: 2 | 3 | ========================== 4 | text2vec-base-multilingual 5 | ========================== 6 | 7 | - **Model Name:** text2vec-base-multilingual 8 | - **Languages:** zh 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 384 15 | - **Max Tokens:** 256 16 | - **Model ID:** shibing624/text2vec-base-multilingual 17 | - **Model Hubs**: `Hugging Face `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name text2vec-base-multilingual --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/embedding/text2vec-large-chinese.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_text2vec-large-chinese: 2 | 3 | ====================== 4 | text2vec-large-chinese 5 | ====================== 6 | 7 | - **Model Name:** text2vec-large-chinese 8 | - **Languages:** zh 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** 1024 15 | - **Max Tokens:** 256 16 | - **Model ID:** shibing624/text2vec-bge-large-chinese 17 | - **Model Hubs**: `Hugging Face `__, `ModelScope `__ 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name text2vec-large-chinese --model-type embedding -------------------------------------------------------------------------------- /doc/source/models/builtin/image/cogview4.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_cogview4: 2 | 3 | ======== 4 | cogview4 5 | ======== 6 | 7 | - **Model Name:** cogview4 8 | - **Model Family:** stable_diffusion 9 | - **Abilities:** text2image 10 | - **Available ControlNet:** None 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** THUDM/CogView4-6B 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name cogview4 --model-type image 20 | 21 | -------------------------------------------------------------------------------- /doc/source/models/builtin/image/flux.1-dev.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_flux.1-dev: 2 | 3 | ========== 4 | FLUX.1-dev 5 | ========== 6 | 7 | - **Model Name:** FLUX.1-dev 8 | - **Model Family:** stable_diffusion 9 | - **Abilities:** text2image, image2image, inpainting 10 | - **Available ControlNet:** None 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** black-forest-labs/FLUX.1-dev 16 | - **GGUF Model ID**: city96/FLUX.1-dev-gguf 17 | - **GGUF Quantizations**: F16, Q2_K, Q3_K_S, Q4_0, Q4_1, Q4_K_S, Q5_0, Q5_1, Q5_K_S, Q6_K, Q8_0 18 | 19 | 20 | Execute the following command to launch the model:: 21 | 22 | xinference launch --model-name FLUX.1-dev --model-type image 23 | 24 | 25 | For GGUF quantization, using below command: 26 | 27 | xinference launch --model-name FLUX.1-dev --model-type image --gguf_quantization ${gguf_quantization} --cpu_offload True 28 | -------------------------------------------------------------------------------- /doc/source/models/builtin/image/flux.1-schnell.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_flux.1-schnell: 2 | 3 | ============== 4 | FLUX.1-schnell 5 | ============== 6 | 7 | - **Model Name:** FLUX.1-schnell 8 | - **Model Family:** stable_diffusion 9 | - **Abilities:** text2image, image2image, inpainting 10 | - **Available ControlNet:** None 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** black-forest-labs/FLUX.1-schnell 16 | - **GGUF Model ID**: city96/FLUX.1-schnell-gguf 17 | - **GGUF Quantizations**: F16, Q2_K, Q3_K_S, Q4_0, Q4_1, Q4_K_S, Q5_0, Q5_1, Q5_K_S, Q6_K, Q8_0 18 | 19 | 20 | Execute the following command to launch the model:: 21 | 22 | xinference launch --model-name FLUX.1-schnell --model-type image 23 | 24 | 25 | For GGUF quantization, using below command: 26 | 27 | xinference launch --model-name FLUX.1-schnell --model-type image --gguf_quantization ${gguf_quantization} --cpu_offload True 28 | -------------------------------------------------------------------------------- /doc/source/models/builtin/image/got-ocr2_0.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_got-ocr2_0: 2 | 3 | ========== 4 | GOT-OCR2_0 5 | ========== 6 | 7 | - **Model Name:** GOT-OCR2_0 8 | - **Model Family:** ocr 9 | - **Abilities:** ocr 10 | - **Available ControlNet:** None 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** stepfun-ai/GOT-OCR2_0 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name GOT-OCR2_0 --model-type image 20 | 21 | -------------------------------------------------------------------------------- /doc/source/models/builtin/image/hunyuandit-v1.2-distilled.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_hunyuandit-v1.2-distilled: 2 | 3 | ========================= 4 | HunyuanDiT-v1.2-Distilled 5 | ========================= 6 | 7 | - **Model Name:** HunyuanDiT-v1.2-Distilled 8 | - **Model Family:** stable_diffusion 9 | - **Abilities:** text2image 10 | - **Available ControlNet:** None 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** Tencent-Hunyuan/HunyuanDiT-v1.2-Diffusers-Distilled 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name HunyuanDiT-v1.2-Distilled --model-type image 20 | 21 | -------------------------------------------------------------------------------- /doc/source/models/builtin/image/hunyuandit-v1.2.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_hunyuandit-v1.2: 2 | 3 | =============== 4 | HunyuanDiT-v1.2 5 | =============== 6 | 7 | - **Model Name:** HunyuanDiT-v1.2 8 | - **Model Family:** stable_diffusion 9 | - **Abilities:** text2image 10 | - **Available ControlNet:** None 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** Tencent-Hunyuan/HunyuanDiT-v1.2-Diffusers 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name HunyuanDiT-v1.2 --model-type image 20 | 21 | -------------------------------------------------------------------------------- /doc/source/models/builtin/image/index.rst: -------------------------------------------------------------------------------- 1 | .. _models_image_index: 2 | 3 | ================ 4 | Image Models 5 | ================ 6 | 7 | The following is a list of built-in image models in Xinference: 8 | 9 | 10 | .. toctree:: 11 | :maxdepth: 1 12 | 13 | 14 | cogview4 15 | 16 | flux.1-dev 17 | 18 | flux.1-schnell 19 | 20 | got-ocr2_0 21 | 22 | hunyuandit-v1.2 23 | 24 | hunyuandit-v1.2-distilled 25 | 26 | kolors 27 | 28 | sd-turbo 29 | 30 | sd3-medium 31 | 32 | sd3.5-large 33 | 34 | sd3.5-large-turbo 35 | 36 | sd3.5-medium 37 | 38 | sdxl-turbo 39 | 40 | stable-diffusion-2-inpainting 41 | 42 | stable-diffusion-inpainting 43 | 44 | stable-diffusion-v1.5 45 | 46 | stable-diffusion-xl-base-1.0 47 | 48 | stable-diffusion-xl-inpainting 49 | -------------------------------------------------------------------------------- /doc/source/models/builtin/image/kolors.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_kolors: 2 | 3 | ====== 4 | kolors 5 | ====== 6 | 7 | - **Model Name:** kolors 8 | - **Model Family:** stable_diffusion 9 | - **Abilities:** text2image, image2image 10 | - **Available ControlNet:** None 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** Kwai-Kolors/Kolors-diffusers 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name kolors --model-type image 20 | 21 | -------------------------------------------------------------------------------- /doc/source/models/builtin/image/sd-turbo.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_sd-turbo: 2 | 3 | ======== 4 | sd-turbo 5 | ======== 6 | 7 | - **Model Name:** sd-turbo 8 | - **Model Family:** stable_diffusion 9 | - **Abilities:** text2image 10 | - **Available ControlNet:** None 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** stabilityai/sd-turbo 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name sd-turbo --model-type image 20 | 21 | -------------------------------------------------------------------------------- /doc/source/models/builtin/image/sd3-medium.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_sd3-medium: 2 | 3 | ========== 4 | sd3-medium 5 | ========== 6 | 7 | - **Model Name:** sd3-medium 8 | - **Model Family:** stable_diffusion 9 | - **Abilities:** text2image, image2image, inpainting 10 | - **Available ControlNet:** None 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** stabilityai/stable-diffusion-3-medium-diffusers 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name sd3-medium --model-type image 20 | 21 | -------------------------------------------------------------------------------- /doc/source/models/builtin/image/sd3.5-large-turbo.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_sd3.5-large-turbo: 2 | 3 | ================= 4 | sd3.5-large-turbo 5 | ================= 6 | 7 | - **Model Name:** sd3.5-large-turbo 8 | - **Model Family:** stable_diffusion 9 | - **Abilities:** text2image, image2image, inpainting 10 | - **Available ControlNet:** None 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** stabilityai/stable-diffusion-3.5-large-turbo 16 | - **GGUF Model ID**: city96/stable-diffusion-3.5-large-turbo-gguf 17 | - **GGUF Quantizations**: F16, Q4_0, Q4_1, Q5_0, Q5_1, Q8_0 18 | 19 | 20 | Execute the following command to launch the model:: 21 | 22 | xinference launch --model-name sd3.5-large-turbo --model-type image 23 | 24 | 25 | For GGUF quantization, using below command: 26 | 27 | xinference launch --model-name sd3.5-large-turbo --model-type image --gguf_quantization ${gguf_quantization} --cpu_offload True 28 | -------------------------------------------------------------------------------- /doc/source/models/builtin/image/sd3.5-large.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_sd3.5-large: 2 | 3 | =========== 4 | sd3.5-large 5 | =========== 6 | 7 | - **Model Name:** sd3.5-large 8 | - **Model Family:** stable_diffusion 9 | - **Abilities:** text2image, image2image, inpainting 10 | - **Available ControlNet:** None 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** stabilityai/stable-diffusion-3.5-large 16 | - **GGUF Model ID**: city96/stable-diffusion-3.5-large-gguf 17 | - **GGUF Quantizations**: F16, Q4_0, Q4_1, Q5_0, Q5_1, Q8_0 18 | 19 | 20 | Execute the following command to launch the model:: 21 | 22 | xinference launch --model-name sd3.5-large --model-type image 23 | 24 | 25 | For GGUF quantization, using below command: 26 | 27 | xinference launch --model-name sd3.5-large --model-type image --gguf_quantization ${gguf_quantization} --cpu_offload True 28 | -------------------------------------------------------------------------------- /doc/source/models/builtin/image/sd3.5-medium.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_sd3.5-medium: 2 | 3 | ============ 4 | sd3.5-medium 5 | ============ 6 | 7 | - **Model Name:** sd3.5-medium 8 | - **Model Family:** stable_diffusion 9 | - **Abilities:** text2image, image2image, inpainting 10 | - **Available ControlNet:** None 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** stabilityai/stable-diffusion-3.5-medium 16 | - **GGUF Model ID**: city96/stable-diffusion-3.5-medium-gguf 17 | - **GGUF Quantizations**: F16, Q3_K_M, Q3_K_S, Q4_0, Q4_1, Q4_K_M, Q4_K_S, Q5_0, Q5_1, Q5_K_M, Q5_K_S, Q6_K, Q8_0 18 | 19 | 20 | Execute the following command to launch the model:: 21 | 22 | xinference launch --model-name sd3.5-medium --model-type image 23 | 24 | 25 | For GGUF quantization, using below command: 26 | 27 | xinference launch --model-name sd3.5-medium --model-type image --gguf_quantization ${gguf_quantization} --cpu_offload True 28 | -------------------------------------------------------------------------------- /doc/source/models/builtin/image/sdxl-turbo.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_sdxl-turbo: 2 | 3 | ========== 4 | sdxl-turbo 5 | ========== 6 | 7 | - **Model Name:** sdxl-turbo 8 | - **Model Family:** stable_diffusion 9 | - **Abilities:** text2image 10 | - **Available ControlNet:** None 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** stabilityai/sdxl-turbo 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name sdxl-turbo --model-type image 20 | 21 | -------------------------------------------------------------------------------- /doc/source/models/builtin/image/stable-diffusion-2-inpainting.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_stable-diffusion-2-inpainting: 2 | 3 | ============================= 4 | stable-diffusion-2-inpainting 5 | ============================= 6 | 7 | - **Model Name:** stable-diffusion-2-inpainting 8 | - **Model Family:** stable_diffusion 9 | - **Abilities:** inpainting 10 | - **Available ControlNet:** None 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** stabilityai/stable-diffusion-2-inpainting 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name stable-diffusion-2-inpainting --model-type image 20 | 21 | -------------------------------------------------------------------------------- /doc/source/models/builtin/image/stable-diffusion-inpainting.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_stable-diffusion-inpainting: 2 | 3 | =========================== 4 | stable-diffusion-inpainting 5 | =========================== 6 | 7 | - **Model Name:** stable-diffusion-inpainting 8 | - **Model Family:** stable_diffusion 9 | - **Abilities:** inpainting 10 | - **Available ControlNet:** None 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** runwayml/stable-diffusion-inpainting 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name stable-diffusion-inpainting --model-type image 20 | 21 | -------------------------------------------------------------------------------- /doc/source/models/builtin/image/stable-diffusion-v1.5.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_stable-diffusion-v1.5: 2 | 3 | ===================== 4 | stable-diffusion-v1.5 5 | ===================== 6 | 7 | - **Model Name:** stable-diffusion-v1.5 8 | - **Model Family:** stable_diffusion 9 | - **Abilities:** text2image, image2image 10 | - **Available ControlNet:** ['canny', 'mlsd', 'hed', 'scribble', 'openpose', 'normal', 'seg'] 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** runwayml/stable-diffusion-v1-5 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name stable-diffusion-v1.5 --model-type image 20 | 21 | -------------------------------------------------------------------------------- /doc/source/models/builtin/image/stable-diffusion-xl-base-1.0.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_stable-diffusion-xl-base-1.0: 2 | 3 | ============================ 4 | stable-diffusion-xl-base-1.0 5 | ============================ 6 | 7 | - **Model Name:** stable-diffusion-xl-base-1.0 8 | - **Model Family:** stable_diffusion 9 | - **Abilities:** text2image, image2image 10 | - **Available ControlNet:** ['canny', 'depth', 'zoe-depth'] 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** stabilityai/stable-diffusion-xl-base-1.0 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name stable-diffusion-xl-base-1.0 --model-type image 20 | 21 | -------------------------------------------------------------------------------- /doc/source/models/builtin/image/stable-diffusion-xl-inpainting.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_stable-diffusion-xl-inpainting: 2 | 3 | ============================== 4 | stable-diffusion-xl-inpainting 5 | ============================== 6 | 7 | - **Model Name:** stable-diffusion-xl-inpainting 8 | - **Model Family:** stable_diffusion 9 | - **Abilities:** inpainting 10 | - **Available ControlNet:** None 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** diffusers/stable-diffusion-xl-1.0-inpainting-0.1 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name stable-diffusion-xl-inpainting --model-type image 20 | 21 | -------------------------------------------------------------------------------- /doc/source/models/builtin/index.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_index: 2 | 3 | ============== 4 | Builtin Models 5 | ============== 6 | 7 | .. toctree:: 8 | :maxdepth: 1 9 | 10 | llm/index 11 | embedding/index 12 | image/index 13 | audio/index 14 | rerank/index 15 | video/index 16 | -------------------------------------------------------------------------------- /doc/source/models/builtin/llm/gpt-2.rst: -------------------------------------------------------------------------------- 1 | .. _models_llm_gpt-2: 2 | 3 | ======================================== 4 | gpt-2 5 | ======================================== 6 | 7 | - **Context Length:** 1024 8 | - **Model Name:** gpt-2 9 | - **Languages:** en 10 | - **Abilities:** generate 11 | - **Description:** GPT-2 is a Transformer-based LLM that is trained on WebTest, a 40 GB dataset of Reddit posts with 3+ upvotes. 12 | 13 | Specifications 14 | ^^^^^^^^^^^^^^ 15 | 16 | 17 | Model Spec 1 (pytorch, 1_5 Billion) 18 | ++++++++++++++++++++++++++++++++++++++++ 19 | 20 | - **Model Format:** pytorch 21 | - **Model Size (in billions):** 1_5 22 | - **Quantizations:** none 23 | - **Engines**: Transformers 24 | - **Model ID:** openai-community/gpt2 25 | - **Model Hubs**: `Hugging Face `__ 26 | 27 | Execute the following command to launch the model, remember to replace ``${quantization}`` with your 28 | chosen quantization method from the options listed above:: 29 | 30 | xinference launch --model-engine ${engine} --model-name gpt-2 --size-in-billions 1_5 --model-format pytorch --quantization ${quantization} 31 | 32 | -------------------------------------------------------------------------------- /doc/source/models/builtin/llm/opt.rst: -------------------------------------------------------------------------------- 1 | .. _models_llm_opt: 2 | 3 | ======================================== 4 | opt 5 | ======================================== 6 | 7 | - **Context Length:** 2048 8 | - **Model Name:** opt 9 | - **Languages:** en 10 | - **Abilities:** generate 11 | - **Description:** Opt is an open-source, decoder-only, Transformer based LLM that was designed to replicate GPT-3. 12 | 13 | Specifications 14 | ^^^^^^^^^^^^^^ 15 | 16 | 17 | Model Spec 1 (pytorch, 1 Billion) 18 | ++++++++++++++++++++++++++++++++++++++++ 19 | 20 | - **Model Format:** pytorch 21 | - **Model Size (in billions):** 1 22 | - **Quantizations:** none 23 | - **Engines**: Transformers 24 | - **Model ID:** facebook/opt-125m 25 | - **Model Hubs**: `Hugging Face `__ 26 | 27 | Execute the following command to launch the model, remember to replace ``${quantization}`` with your 28 | chosen quantization method from the options listed above:: 29 | 30 | xinference launch --model-engine ${engine} --model-name opt --size-in-billions 1 --model-format pytorch --quantization ${quantization} 31 | 32 | -------------------------------------------------------------------------------- /doc/source/models/builtin/rerank/bce-reranker-base_v1.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_bce-reranker-base_v1: 2 | 3 | ==================== 4 | bce-reranker-base_v1 5 | ==================== 6 | 7 | - **Model Name:** bce-reranker-base_v1 8 | - **Languages:** en, zh 9 | - **Abilities:** rerank 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Model ID:** maidalun1020/bce-reranker-base_v1 15 | 16 | Execute the following command to launch the model:: 17 | 18 | xinference launch --model-name bce-reranker-base_v1 --model-type rerank -------------------------------------------------------------------------------- /doc/source/models/builtin/rerank/bge-reranker-base.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_bge-reranker-base: 2 | 3 | ================= 4 | bge-reranker-base 5 | ================= 6 | 7 | - **Model Name:** bge-reranker-base 8 | - **Languages:** en, zh 9 | - **Abilities:** rerank 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Model ID:** BAAI/bge-reranker-base 15 | 16 | Execute the following command to launch the model:: 17 | 18 | xinference launch --model-name bge-reranker-base --model-type rerank -------------------------------------------------------------------------------- /doc/source/models/builtin/rerank/bge-reranker-large.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_bge-reranker-large: 2 | 3 | ================== 4 | bge-reranker-large 5 | ================== 6 | 7 | - **Model Name:** bge-reranker-large 8 | - **Languages:** en, zh 9 | - **Abilities:** rerank 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Model ID:** BAAI/bge-reranker-large 15 | 16 | Execute the following command to launch the model:: 17 | 18 | xinference launch --model-name bge-reranker-large --model-type rerank -------------------------------------------------------------------------------- /doc/source/models/builtin/rerank/bge-reranker-v2-gemma.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_bge-reranker-v2-gemma: 2 | 3 | ===================== 4 | bge-reranker-v2-gemma 5 | ===================== 6 | 7 | - **Model Name:** bge-reranker-v2-gemma 8 | - **Languages:** en, zh, multilingual 9 | - **Abilities:** rerank 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Model ID:** BAAI/bge-reranker-v2-gemma 15 | 16 | Execute the following command to launch the model:: 17 | 18 | xinference launch --model-name bge-reranker-v2-gemma --model-type rerank -------------------------------------------------------------------------------- /doc/source/models/builtin/rerank/bge-reranker-v2-m3.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_bge-reranker-v2-m3: 2 | 3 | ================== 4 | bge-reranker-v2-m3 5 | ================== 6 | 7 | - **Model Name:** bge-reranker-v2-m3 8 | - **Languages:** en, zh, multilingual 9 | - **Abilities:** rerank 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Model ID:** BAAI/bge-reranker-v2-m3 15 | 16 | Execute the following command to launch the model:: 17 | 18 | xinference launch --model-name bge-reranker-v2-m3 --model-type rerank -------------------------------------------------------------------------------- /doc/source/models/builtin/rerank/bge-reranker-v2-minicpm-layerwise.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_bge-reranker-v2-minicpm-layerwise: 2 | 3 | ================================= 4 | bge-reranker-v2-minicpm-layerwise 5 | ================================= 6 | 7 | - **Model Name:** bge-reranker-v2-minicpm-layerwise 8 | - **Languages:** en, zh, multilingual 9 | - **Abilities:** rerank 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Model ID:** BAAI/bge-reranker-v2-minicpm-layerwise 15 | 16 | Execute the following command to launch the model:: 17 | 18 | xinference launch --model-name bge-reranker-v2-minicpm-layerwise --model-type rerank -------------------------------------------------------------------------------- /doc/source/models/builtin/rerank/index.rst: -------------------------------------------------------------------------------- 1 | .. _models_rerank_index: 2 | 3 | ================ 4 | Rerank Models 5 | ================ 6 | 7 | The following is a list of built-in rerank models in Xinference: 8 | 9 | 10 | .. toctree:: 11 | :maxdepth: 1 12 | 13 | 14 | bce-reranker-base_v1 15 | 16 | bge-reranker-base 17 | 18 | bge-reranker-large 19 | 20 | bge-reranker-v2-gemma 21 | 22 | bge-reranker-v2-m3 23 | 24 | bge-reranker-v2-minicpm-layerwise 25 | 26 | jina-reranker-v2 27 | 28 | minicpm-reranker 29 | -------------------------------------------------------------------------------- /doc/source/models/builtin/rerank/jina-reranker-v2.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_jina-reranker-v2: 2 | 3 | ================ 4 | jina-reranker-v2 5 | ================ 6 | 7 | - **Model Name:** jina-reranker-v2 8 | - **Languages:** en, zh, multilingual 9 | - **Abilities:** rerank 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Model ID:** jinaai/jina-reranker-v2-base-multilingual 15 | 16 | Execute the following command to launch the model:: 17 | 18 | xinference launch --model-name jina-reranker-v2 --model-type rerank -------------------------------------------------------------------------------- /doc/source/models/builtin/rerank/minicpm-reranker.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_minicpm-reranker: 2 | 3 | ================ 4 | minicpm-reranker 5 | ================ 6 | 7 | - **Model Name:** minicpm-reranker 8 | - **Languages:** en, zh 9 | - **Abilities:** rerank 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Model ID:** openbmb/MiniCPM-Reranker 15 | 16 | Execute the following command to launch the model:: 17 | 18 | xinference launch --model-name minicpm-reranker --model-type rerank -------------------------------------------------------------------------------- /doc/source/models/builtin/video/cogvideox-2b.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_cogvideox-2b: 2 | 3 | ============ 4 | CogVideoX-2b 5 | ============ 6 | 7 | - **Model Name:** CogVideoX-2b 8 | - **Model Family:** CogVideoX 9 | - **Abilities:** text2video 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Model ID:** THUDM/CogVideoX-2b 15 | 16 | Execute the following command to launch the model:: 17 | 18 | xinference launch --model-name CogVideoX-2b --model-type video -------------------------------------------------------------------------------- /doc/source/models/builtin/video/cogvideox-5b.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_cogvideox-5b: 2 | 3 | ============ 4 | CogVideoX-5b 5 | ============ 6 | 7 | - **Model Name:** CogVideoX-5b 8 | - **Model Family:** CogVideoX 9 | - **Abilities:** text2video 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Model ID:** THUDM/CogVideoX-5b 15 | 16 | Execute the following command to launch the model:: 17 | 18 | xinference launch --model-name CogVideoX-5b --model-type video -------------------------------------------------------------------------------- /doc/source/models/builtin/video/hunyuanvideo.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_hunyuanvideo: 2 | 3 | ============ 4 | HunyuanVideo 5 | ============ 6 | 7 | - **Model Name:** HunyuanVideo 8 | - **Model Family:** HunyuanVideo 9 | - **Abilities:** text2video 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Model ID:** hunyuanvideo-community/HunyuanVideo 15 | 16 | Execute the following command to launch the model:: 17 | 18 | xinference launch --model-name HunyuanVideo --model-type video -------------------------------------------------------------------------------- /doc/source/models/builtin/video/index.rst: -------------------------------------------------------------------------------- 1 | .. _models_video_index: 2 | 3 | ================ 4 | Video Models 5 | ================ 6 | 7 | The following is a list of built-in video models in Xinference: 8 | 9 | 10 | .. toctree:: 11 | :maxdepth: 1 12 | 13 | 14 | cogvideox-2b 15 | 16 | cogvideox-5b 17 | 18 | hunyuanvideo 19 | 20 | wan2.1-1.3b 21 | 22 | wan2.1-14b 23 | 24 | wan2.1-flf2v-14b-720p 25 | 26 | wan2.1-i2v-14b-480p 27 | 28 | wan2.1-i2v-14b-720p 29 | -------------------------------------------------------------------------------- /doc/source/models/builtin/video/wan2.1-1.3b.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_wan2.1-1.3b: 2 | 3 | =========== 4 | Wan2.1-1.3B 5 | =========== 6 | 7 | - **Model Name:** Wan2.1-1.3B 8 | - **Model Family:** Wan 9 | - **Abilities:** text2video 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Model ID:** Wan-AI/Wan2.1-T2V-1.3B-Diffusers 15 | 16 | Execute the following command to launch the model:: 17 | 18 | xinference launch --model-name Wan2.1-1.3B --model-type video -------------------------------------------------------------------------------- /doc/source/models/builtin/video/wan2.1-14b.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_wan2.1-14b: 2 | 3 | ========== 4 | Wan2.1-14B 5 | ========== 6 | 7 | - **Model Name:** Wan2.1-14B 8 | - **Model Family:** Wan 9 | - **Abilities:** text2video 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Model ID:** Wan-AI/Wan2.1-T2V-14B-Diffusers 15 | 16 | Execute the following command to launch the model:: 17 | 18 | xinference launch --model-name Wan2.1-14B --model-type video -------------------------------------------------------------------------------- /doc/source/models/builtin/video/wan2.1-flf2v-14b-720p.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_wan2.1-flf2v-14b-720p: 2 | 3 | ===================== 4 | Wan2.1-flf2v-14B-720p 5 | ===================== 6 | 7 | - **Model Name:** Wan2.1-flf2v-14B-720p 8 | - **Model Family:** Wan 9 | - **Abilities:** firstlastframe2video 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Model ID:** Wan-AI/Wan2.1-FLF2V-14B-720P-diffusers 15 | 16 | Execute the following command to launch the model:: 17 | 18 | xinference launch --model-name Wan2.1-flf2v-14B-720p --model-type video -------------------------------------------------------------------------------- /doc/source/models/builtin/video/wan2.1-i2v-14b-480p.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_wan2.1-i2v-14b-480p: 2 | 3 | =================== 4 | Wan2.1-i2v-14B-480p 5 | =================== 6 | 7 | - **Model Name:** Wan2.1-i2v-14B-480p 8 | - **Model Family:** Wan 9 | - **Abilities:** image2video 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Model ID:** Wan-AI/Wan2.1-I2V-14B-480P-Diffusers 15 | 16 | Execute the following command to launch the model:: 17 | 18 | xinference launch --model-name Wan2.1-i2v-14B-480p --model-type video -------------------------------------------------------------------------------- /doc/source/models/builtin/video/wan2.1-i2v-14b-720p.rst: -------------------------------------------------------------------------------- 1 | .. _models_builtin_wan2.1-i2v-14b-720p: 2 | 3 | =================== 4 | Wan2.1-i2v-14B-720p 5 | =================== 6 | 7 | - **Model Name:** Wan2.1-i2v-14B-720p 8 | - **Model Family:** Wan 9 | - **Abilities:** image2video 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Model ID:** Wan-AI/Wan2.1-I2V-14B-720P-Diffusers 15 | 16 | Execute the following command to launch the model:: 17 | 18 | xinference launch --model-name Wan2.1-i2v-14B-720p --model-type video -------------------------------------------------------------------------------- /doc/source/models/model_abilities/index.rst: -------------------------------------------------------------------------------- 1 | .. _abilities_index: 2 | 3 | =============== 4 | Model Abilities 5 | =============== 6 | 7 | .. toctree:: 8 | :maxdepth: 2 9 | 10 | chat 11 | tools 12 | multimodal 13 | embed 14 | rerank 15 | image 16 | audio 17 | video 18 | -------------------------------------------------------------------------------- /doc/source/user_guide/index.rst: -------------------------------------------------------------------------------- 1 | .. _user_guide_index: 2 | 3 | ========== 4 | User Guide 5 | ========== 6 | 7 | .. toctree:: 8 | :maxdepth: 2 9 | 10 | backends 11 | client_api 12 | auth_system 13 | metrics 14 | distributed_inference 15 | continuous_batching 16 | vllm_enhancement 17 | -------------------------------------------------------------------------------- /doc/source/user_guide/vllm_enhancement.rst: -------------------------------------------------------------------------------- 1 | .. _user_guide_vllm_enhancement: 2 | 3 | ############################################ 4 | Xavier: Share KV Cache between vllm replicas 5 | ############################################ 6 | For scenarios such as long document queries and multi-round conversations, 7 | the computation during the inference prefill phase can be particularly heavy, 8 | which affects overall throughput and the latency of individual inferences. 9 | Xinference enhances the vllm engine by introducing the ``Xavier`` framework, 10 | enabling KV cache sharing across multiple vllm instances. 11 | This allows KV cache computed by other replicas to be directly reused, avoiding redundant computations. 12 | 13 | ***** 14 | Usage 15 | ***** 16 | Simply add the parameter ``enable_xavier=True`` when starting the vllm model. 17 | 18 | *********** 19 | Limitations 20 | *********** 21 | * Xavier requires vllm version >= ``0.7.0``. 22 | * Due to the underlying communication not recognizing ``0.0.0.0``, the actual IP address needs to be passed when starting Xinference, for example: ``xinference-local -H 192.168.xx.xx``. 23 | -------------------------------------------------------------------------------- /doc/templates/audio.rst.jinja: -------------------------------------------------------------------------------- 1 | .. _models_builtin_{{ model_name|lower }}: 2 | 3 | {{ "=" * model_name|length }} 4 | {{ model_name }} 5 | {{ "=" * model_name|length }} 6 | 7 | - **Model Name:** {{ model_name }} 8 | - **Model Family:** {{ model_family }} 9 | - **Abilities:** {{ model_ability }} 10 | - **Multilingual:** {{ multilingual }} 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** {{ model_id }} 16 | 17 | Execute the following command to launch the model:: 18 | 19 | xinference launch --model-name {{ model_name }} --model-type audio -------------------------------------------------------------------------------- /doc/templates/audio_index.rst.jinja: -------------------------------------------------------------------------------- 1 | .. _models_audio_index: 2 | 3 | ================ 4 | Audio Models 5 | ================ 6 | 7 | The following is a list of built-in audio models in Xinference: 8 | 9 | 10 | .. toctree:: 11 | :maxdepth: 1 12 | 13 | {% for model in models %} 14 | {{ model.model_name|lower }} 15 | {% endfor %} -------------------------------------------------------------------------------- /doc/templates/embedding.rst.jinja: -------------------------------------------------------------------------------- 1 | .. _models_builtin_{{ model_name|lower }}: 2 | 3 | {{ "=" * model_name|length }} 4 | {{ model_name }} 5 | {{ "=" * model_name|length }} 6 | 7 | - **Model Name:** {{ model_name }} 8 | - **Languages:** {{ ', '.join(language) }} 9 | - **Abilities:** embed 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Dimensions:** {{ dimensions }} 15 | - **Max Tokens:** {{ max_tokens }} 16 | - **Model ID:** {{ model_id }} 17 | - **Model Hubs**: {% for hub in model_hubs -%}`{{ hub.name }} <{{ hub.url }}>`__{% if not loop.last %}, {% endif %} {%- endfor %} 18 | 19 | Execute the following command to launch the model:: 20 | 21 | xinference launch --model-name {{ model_name }} --model-type embedding -------------------------------------------------------------------------------- /doc/templates/embedding_index.rst.jinja: -------------------------------------------------------------------------------- 1 | .. _models_embedding_index: 2 | 3 | ================ 4 | Embedding Models 5 | ================ 6 | 7 | The following is a list of built-in embedding models in Xinference: 8 | 9 | 10 | .. toctree:: 11 | :maxdepth: 1 12 | 13 | {% for model in models %} 14 | {{ model.model_name|lower }} 15 | {% endfor %} -------------------------------------------------------------------------------- /doc/templates/image.rst.jinja: -------------------------------------------------------------------------------- 1 | .. _models_builtin_{{ model_name|lower }}: 2 | 3 | {{ "=" * model_name|length }} 4 | {{ model_name }} 5 | {{ "=" * model_name|length }} 6 | 7 | - **Model Name:** {{ model_name }} 8 | - **Model Family:** {{ model_family }} 9 | - **Abilities:** {{ model_ability }} 10 | - **Available ControlNet:** {{ available_controlnet }} 11 | 12 | Specifications 13 | ^^^^^^^^^^^^^^ 14 | 15 | - **Model ID:** {{ model_id }} 16 | {%- if gguf_quantizations %} 17 | - **GGUF Model ID**: {{ gguf_model_id }} 18 | - **GGUF Quantizations**: {{ gguf_quantizations }} 19 | {% endif %} 20 | 21 | Execute the following command to launch the model:: 22 | 23 | xinference launch --model-name {{ model_name }} --model-type image 24 | 25 | {% if gguf_quantizations %} 26 | For GGUF quantization, using below command: 27 | 28 | xinference launch --model-name {{ model_name }} --model-type image --gguf_quantization ${{ '{' }}gguf_quantization{{ '}' }} --cpu_offload True 29 | {% endif %} -------------------------------------------------------------------------------- /doc/templates/image_index.rst.jinja: -------------------------------------------------------------------------------- 1 | .. _models_image_index: 2 | 3 | ================ 4 | Image Models 5 | ================ 6 | 7 | The following is a list of built-in image models in Xinference: 8 | 9 | 10 | .. toctree:: 11 | :maxdepth: 1 12 | 13 | {% for model in models %} 14 | {{ model.model_name|lower }} 15 | {% endfor %} -------------------------------------------------------------------------------- /doc/templates/llm_index.rst.jinja: -------------------------------------------------------------------------------- 1 | .. _models_llm_index: 2 | 3 | ===================== 4 | Large language Models 5 | ===================== 6 | 7 | The following is a list of built-in LLM in Xinference: 8 | 9 | .. list-table:: 10 | :widths: 25 25 25 50 11 | :header-rows: 1 12 | 13 | * - MODEL NAME 14 | - ABILITIES 15 | - COTNEXT_LENGTH 16 | - DESCRIPTION 17 | 18 | {% for model in models %} 19 | * - :ref:`{{ model.model_name|lower }} ` 20 | - {{ model.model_ability | join(', ') }} 21 | - {{ model.context_length }} 22 | - {{ model.model_description }} 23 | {% endfor %} 24 | 25 | .. toctree:: 26 | :maxdepth: 3 27 | 28 | {% for model in models %} 29 | {{ model.model_name|lower }} 30 | {% endfor %} 31 | 32 | 33 | -------------------------------------------------------------------------------- /doc/templates/metrics.jinja: -------------------------------------------------------------------------------- 1 | .. _metrics: 2 | 3 | ================== 4 | Metrics 5 | ================== 6 | 7 | There are two types of metrics exporters in an Xinference cluster: 8 | 9 | - Supervisor metrics exporter at `/metrics`, e.g. `http://127.0.0.1:9997/metrics`. 10 | - Worker metrics exporter at each worker node, the exporter host and port can be set by `--metrics-exporter-host` and `--metrics-exporter-port` options in `xinference-local` or `xinference-worker` command. 11 | 12 | Supervisor Metrics 13 | ^^^^^^^^^^^^^^^^^^ 14 | 15 | 16 | {% for m in supervisor_metrics %} 17 | - **{{ m.name }}** ({{ m.type }}): {{ m.help }} 18 | {% endfor %} 19 | 20 | 21 | Worker Metrics 22 | ^^^^^^^^^^^^^^ 23 | 24 | 25 | {% for m in worker_metrics %} 26 | - **{{ m.name }}** ({{ m.type }}): {{ m.help }} 27 | {% endfor %} 28 | -------------------------------------------------------------------------------- /doc/templates/rerank.rst.jinja: -------------------------------------------------------------------------------- 1 | .. _models_builtin_{{ model_name|lower }}: 2 | 3 | {{ "=" * model_name|length }} 4 | {{ model_name }} 5 | {{ "=" * model_name|length }} 6 | 7 | - **Model Name:** {{ model_name }} 8 | - **Languages:** {{ ', '.join(language) }} 9 | - **Abilities:** rerank 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Model ID:** {{ model_id }} 15 | 16 | Execute the following command to launch the model:: 17 | 18 | xinference launch --model-name {{ model_name }} --model-type rerank -------------------------------------------------------------------------------- /doc/templates/rerank_index.rst.jinja: -------------------------------------------------------------------------------- 1 | .. _models_rerank_index: 2 | 3 | ================ 4 | Rerank Models 5 | ================ 6 | 7 | The following is a list of built-in rerank models in Xinference: 8 | 9 | 10 | .. toctree:: 11 | :maxdepth: 1 12 | 13 | {% for model in models %} 14 | {{ model.model_name|lower }} 15 | {% endfor %} -------------------------------------------------------------------------------- /doc/templates/video.rst.jinja: -------------------------------------------------------------------------------- 1 | .. _models_builtin_{{ model_name|lower }}: 2 | 3 | {{ "=" * model_name|length }} 4 | {{ model_name }} 5 | {{ "=" * model_name|length }} 6 | 7 | - **Model Name:** {{ model_name }} 8 | - **Model Family:** {{ model_family }} 9 | - **Abilities:** {{ model_ability }} 10 | 11 | Specifications 12 | ^^^^^^^^^^^^^^ 13 | 14 | - **Model ID:** {{ model_id }} 15 | 16 | Execute the following command to launch the model:: 17 | 18 | xinference launch --model-name {{ model_name }} --model-type video -------------------------------------------------------------------------------- /doc/templates/video_index.rst.jinja: -------------------------------------------------------------------------------- 1 | .. _models_video_index: 2 | 3 | ================ 4 | Video Models 5 | ================ 6 | 7 | The following is a list of built-in video models in Xinference: 8 | 9 | 10 | .. toctree:: 11 | :maxdepth: 1 12 | 13 | {% for model in models %} 14 | {{ model.model_name|lower }} 15 | {% endfor %} -------------------------------------------------------------------------------- /examples/draft.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/examples/draft.png -------------------------------------------------------------------------------- /pyproject.toml: -------------------------------------------------------------------------------- 1 | [build-system] 2 | requires = [ 3 | "setuptools>=64; python_version<'3.12'", 4 | "setuptools>=75; python_version>='3.12'" 5 | ] 6 | build-backend = "setuptools.build_meta" 7 | 8 | [tool.black] 9 | include = '\.pyi?$' 10 | extend-exclude = ''' 11 | /( 12 | | ^/xinference/(_version.py) 13 | | thirdparty 14 | )/ 15 | ''' 16 | 17 | [tool.pytest.ini_options] 18 | asyncio_mode = "auto" 19 | 20 | [tool.cibuildwheel] 21 | build = ["cp39-*", "cp310-*", "cp311-*", "cp312-*"] 22 | skip = "pp* *musllinux* *i686 cp36* cp39-win32 cp310-win32 cp311-win32 cp312-win32" 23 | manylinux-x86_64-image = "manylinux2014" 24 | -------------------------------------------------------------------------------- /xinference/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2023 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | 15 | from . import _version 16 | 17 | __version__ = _version.get_versions()["version"] 18 | 19 | 20 | try: 21 | import intel_extension_for_pytorch # noqa: F401 22 | except: 23 | pass 24 | 25 | 26 | def _install(): 27 | from xoscar.backends.router import Router 28 | 29 | default_router = Router.get_instance_or_empty() 30 | Router.set_instance(default_router) 31 | 32 | 33 | _install() 34 | del _install 35 | -------------------------------------------------------------------------------- /xinference/api/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2023 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/api/oauth2/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2023 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/client/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2023 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | 15 | from .restful.restful_client import Client 16 | 17 | # For compatibility 18 | RESTfulClient = Client 19 | -------------------------------------------------------------------------------- /xinference/client/handlers.py: -------------------------------------------------------------------------------- 1 | from .restful.restful_client import ( # noqa: F401 2 | RESTfulAudioModelHandle as AudioModelHandle, 3 | ) 4 | from .restful.restful_client import ( # noqa: F401 5 | RESTfulChatModelHandle as ChatModelHandle, 6 | ) 7 | from .restful.restful_client import ( # noqa: F401 8 | RESTfulEmbeddingModelHandle as EmbeddingModelHandle, 9 | ) 10 | from .restful.restful_client import ( # noqa: F401 11 | RESTfulGenerateModelHandle as GenerateModelHandle, 12 | ) 13 | from .restful.restful_client import ( # noqa: F401 14 | RESTfulImageModelHandle as ImageModelHandle, 15 | ) 16 | from .restful.restful_client import ( # noqa: F401 17 | RESTfulVideoModelHandle as VideoModelHandle, 18 | ) 19 | -------------------------------------------------------------------------------- /xinference/client/restful/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2023 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/client/tests/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2023 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/core/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2023 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/core/tests/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2023 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/deploy/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2023 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/deploy/docker/requirements-base.txt: -------------------------------------------------------------------------------- 1 | xoscar>=0.7.2 2 | gradio==5.22.0 3 | pillow 4 | click 5 | tqdm>=4.27 6 | tabulate 7 | requests 8 | pydantic>2 9 | fastapi>=0.110.3 10 | uvicorn 11 | huggingface-hub>=0.19.4 12 | typing_extensions 13 | modelscope>=1.10.0 14 | sse_starlette>=1.6.5 # ensure_bytes API break change: https://github.com/sysid/sse-starlette/issues/65 15 | openai>=1.40.0 # For typing 16 | python-jose[cryptography] 17 | passlib[bcrypt] 18 | aioprometheus[starlette]>=23.12.0 19 | nvidia-ml-py 20 | pynvml>=12 21 | async-timeout 22 | peft 23 | opencv-contrib-python-headless 24 | setproctitle 25 | gguf 26 | -------------------------------------------------------------------------------- /xinference/deploy/docker/requirements_cpu-base.txt: -------------------------------------------------------------------------------- 1 | xoscar>=0.7.2 2 | gradio==5.22.0 3 | pillow 4 | click 5 | tqdm>=4.27 6 | tabulate 7 | requests 8 | pydantic>2 9 | fastapi>=0.110.3 10 | uvicorn 11 | huggingface-hub>=0.19.4 12 | typing_extensions 13 | boto3>=1.28.55,<1.28.65 14 | sse_starlette>=1.6.5 15 | openai>1 16 | python-jose[cryptography] 17 | passlib[bcrypt] 18 | aioprometheus[starlette]>=23.12.0 19 | nvidia-ml-py 20 | async-timeout 21 | orjson 22 | protobuf 23 | setproctitle 24 | uv 25 | loguru # For Fish Speech 26 | natsort # For Fish Speech 27 | ormsgpack # For Fish Speech 28 | cachetools # For Fish Speech 29 | imageio-ffmpeg # For video 30 | opencv-contrib-python-headless 31 | gguf 32 | -------------------------------------------------------------------------------- /xinference/deploy/docker/requirements_cpu-ml.txt: -------------------------------------------------------------------------------- 1 | torch>=2.0.0 2 | torchaudio 3 | sentencepiece 4 | sentence-transformers>=3.1.0 5 | transformers>=4.46.0 6 | transformers_stream_generator 7 | accelerate>=0.28.0 8 | bitsandbytes 9 | tiktoken>=0.6.0 10 | gguf 11 | gptqmodel 12 | autoawq!=0.2.6 13 | optimum 14 | peft 15 | timm 16 | tensorizer~=2.9.0 17 | modelscope>=1.19.0 18 | FlagEmbedding 19 | controlnet_aux 20 | einops 21 | -------------------------------------------------------------------------------- /xinference/deploy/test/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2023 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/model/audio/tests/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2023 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/model/audio/tests/basic_ref_en.wav: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/model/audio/tests/basic_ref_en.wav -------------------------------------------------------------------------------- /xinference/model/audio/tests/bbc_news.npy: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/model/audio/tests/bbc_news.npy -------------------------------------------------------------------------------- /xinference/model/audio/tests/bbc_news.wav: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/model/audio/tests/bbc_news.wav -------------------------------------------------------------------------------- /xinference/model/audio/tests/common_voice_zh-CN_38026095.mp3: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/model/audio/tests/common_voice_zh-CN_38026095.mp3 -------------------------------------------------------------------------------- /xinference/model/audio/tests/cross_lingual_prompt.wav: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/model/audio/tests/cross_lingual_prompt.wav -------------------------------------------------------------------------------- /xinference/model/audio/tests/jfk.flac: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/model/audio/tests/jfk.flac -------------------------------------------------------------------------------- /xinference/model/audio/tests/zero_shot_prompt.wav: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/model/audio/tests/zero_shot_prompt.wav -------------------------------------------------------------------------------- /xinference/model/embedding/tests/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2023 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/model/embedding/utils.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2024 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | from .core import EmbeddingModelSpec 15 | 16 | 17 | def get_model_version(embedding_model: EmbeddingModelSpec) -> str: 18 | return f"{embedding_model.model_name}--{embedding_model.max_tokens}--{embedding_model.dimensions}" 19 | -------------------------------------------------------------------------------- /xinference/model/flexible/launchers/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2024 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | 15 | from .image_process_launcher import launcher as image_process 16 | from .transformers_launcher import launcher as transformers 17 | -------------------------------------------------------------------------------- /xinference/model/flexible/tests/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2024 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/model/image/ocr/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2023 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/model/image/scheduler/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2024 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/model/image/stable_diffusion/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2023 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/model/image/tests/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2023 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/model/llm/llama_cpp/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2023 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/model/llm/llama_cpp/tests/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2023 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/model/llm/llama_cpp/tests/dummy.gguf: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/model/llm/llama_cpp/tests/dummy.gguf -------------------------------------------------------------------------------- /xinference/model/llm/lmdeploy/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/model/llm/lmdeploy/__init__.py -------------------------------------------------------------------------------- /xinference/model/llm/lmdeploy/tests/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2023 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/model/llm/mlx/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2023 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/model/llm/mlx/tests/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2023 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/model/llm/mlx/tests/fish.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/model/llm/mlx/tests/fish.png -------------------------------------------------------------------------------- /xinference/model/llm/sglang/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2024 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/model/llm/tests/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2023 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/model/llm/transformers/multimodal/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2025 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/model/llm/transformers/tests/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2023 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/model/llm/vllm/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2023 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/model/llm/vllm/tests/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/model/llm/vllm/tests/__init__.py -------------------------------------------------------------------------------- /xinference/model/llm/vllm/xavier/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2025 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/model/llm/vllm/xavier/test/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2025 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/model/rerank/tests/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2023 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/model/tests/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2025 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/model/video/tests/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2023 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/thirdparty/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/cosyvoice/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/cosyvoice/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/cosyvoice/bin/spk2info.pt: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/cosyvoice/bin/spk2info.pt -------------------------------------------------------------------------------- /xinference/thirdparty/cosyvoice/cli/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/cosyvoice/cli/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/cosyvoice/dataset/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/cosyvoice/dataset/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/cosyvoice/transformer/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/cosyvoice/transformer/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/cosyvoice/utils/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/cosyvoice/utils/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/cosyvoice/utils/losses.py: -------------------------------------------------------------------------------- 1 | import torch 2 | import torch.nn.functional as F 3 | 4 | 5 | def tpr_loss(disc_real_outputs, disc_generated_outputs, tau): 6 | loss = 0 7 | for dr, dg in zip(disc_real_outputs, disc_generated_outputs): 8 | m_DG = torch.median((dr - dg)) 9 | L_rel = torch.mean((((dr - dg) - m_DG) ** 2)[dr < dg + m_DG]) 10 | loss += tau - F.relu(tau - L_rel) 11 | return loss 12 | 13 | 14 | def mel_loss(real_speech, generated_speech, mel_transforms): 15 | loss = 0 16 | for transform in mel_transforms: 17 | mel_r = transform(real_speech) 18 | mel_g = transform(generated_speech) 19 | loss += F.l1_loss(mel_g, mel_r) 20 | return loss 21 | -------------------------------------------------------------------------------- /xinference/thirdparty/deepseek_vl/serve/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2023 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/thirdparty/deepseek_vl/serve/app_modules/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2023 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/thirdparty/deepseek_vl/serve/assets/avatar.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/deepseek_vl/serve/assets/avatar.png -------------------------------------------------------------------------------- /xinference/thirdparty/deepseek_vl/serve/assets/favicon.ico: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/deepseek_vl/serve/assets/favicon.ico -------------------------------------------------------------------------------- /xinference/thirdparty/deepseek_vl/serve/examples/app.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/deepseek_vl/serve/examples/app.png -------------------------------------------------------------------------------- /xinference/thirdparty/deepseek_vl/serve/examples/chart.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/deepseek_vl/serve/examples/chart.png -------------------------------------------------------------------------------- /xinference/thirdparty/deepseek_vl/serve/examples/mirror.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/deepseek_vl/serve/examples/mirror.png -------------------------------------------------------------------------------- /xinference/thirdparty/deepseek_vl/serve/examples/pipeline.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/deepseek_vl/serve/examples/pipeline.png -------------------------------------------------------------------------------- /xinference/thirdparty/deepseek_vl/serve/examples/puzzle.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/deepseek_vl/serve/examples/puzzle.png -------------------------------------------------------------------------------- /xinference/thirdparty/deepseek_vl/serve/examples/rap.jpeg: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/deepseek_vl/serve/examples/rap.jpeg -------------------------------------------------------------------------------- /xinference/thirdparty/deepseek_vl2/serve/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/deepseek_vl2/serve/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/deepseek_vl2/serve/app_modules/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/deepseek_vl2/serve/app_modules/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/deepseek_vl2/serve/assets/avatar.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/deepseek_vl2/serve/assets/avatar.png -------------------------------------------------------------------------------- /xinference/thirdparty/deepseek_vl2/serve/assets/favicon.ico: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/deepseek_vl2/serve/assets/favicon.ico -------------------------------------------------------------------------------- /xinference/thirdparty/deepseek_vl2/serve/assets/simsun.ttc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/deepseek_vl2/serve/assets/simsun.ttc -------------------------------------------------------------------------------- /xinference/thirdparty/f5_tts/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/f5_tts/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/f5_tts/eval/eval_infer_batch.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | 3 | # e.g. F5-TTS, 16 NFE 4 | accelerate launch src/f5_tts/eval/eval_infer_batch.py -s 0 -n "F5TTS_Base" -t "seedtts_test_zh" -nfe 16 5 | accelerate launch src/f5_tts/eval/eval_infer_batch.py -s 0 -n "F5TTS_Base" -t "seedtts_test_en" -nfe 16 6 | accelerate launch src/f5_tts/eval/eval_infer_batch.py -s 0 -n "F5TTS_Base" -t "ls_pc_test_clean" -nfe 16 7 | 8 | # e.g. Vanilla E2 TTS, 32 NFE 9 | accelerate launch src/f5_tts/eval/eval_infer_batch.py -s 0 -n "E2TTS_Base" -t "seedtts_test_zh" -o "midpoint" -ss 0 10 | accelerate launch src/f5_tts/eval/eval_infer_batch.py -s 0 -n "E2TTS_Base" -t "seedtts_test_en" -o "midpoint" -ss 0 11 | accelerate launch src/f5_tts/eval/eval_infer_batch.py -s 0 -n "E2TTS_Base" -t "ls_pc_test_clean" -o "midpoint" -ss 0 12 | 13 | # etc. 14 | -------------------------------------------------------------------------------- /xinference/thirdparty/f5_tts/infer/examples/basic/basic.toml: -------------------------------------------------------------------------------- 1 | # F5-TTS | E2-TTS 2 | model = "F5-TTS" 3 | ref_audio = "infer/examples/basic/basic_ref_en.wav" 4 | # If an empty "", transcribes the reference audio automatically. 5 | ref_text = "Some call me nature, others call me mother nature." 6 | gen_text = "I don't really care what you call me. I've been a silent spectator, watching species evolve, empires rise and fall. But always remember, I am mighty and enduring." 7 | # File with text to generate. Ignores the text above. 8 | gen_file = "" 9 | remove_silence = false 10 | output_dir = "tests" 11 | output_file = "infer_cli_out.wav" 12 | -------------------------------------------------------------------------------- /xinference/thirdparty/f5_tts/infer/examples/basic/basic_ref_en.wav: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/f5_tts/infer/examples/basic/basic_ref_en.wav -------------------------------------------------------------------------------- /xinference/thirdparty/f5_tts/infer/examples/basic/basic_ref_zh.wav: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/f5_tts/infer/examples/basic/basic_ref_zh.wav -------------------------------------------------------------------------------- /xinference/thirdparty/f5_tts/infer/examples/multi/country.flac: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/f5_tts/infer/examples/multi/country.flac -------------------------------------------------------------------------------- /xinference/thirdparty/f5_tts/infer/examples/multi/main.flac: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/f5_tts/infer/examples/multi/main.flac -------------------------------------------------------------------------------- /xinference/thirdparty/f5_tts/infer/examples/multi/story.toml: -------------------------------------------------------------------------------- 1 | # F5-TTS | E2-TTS 2 | model = "F5-TTS" 3 | ref_audio = "infer/examples/multi/main.flac" 4 | # If an empty "", transcribes the reference audio automatically. 5 | ref_text = "" 6 | gen_text = "" 7 | # File with text to generate. Ignores the text above. 8 | gen_file = "infer/examples/multi/story.txt" 9 | remove_silence = true 10 | output_dir = "tests" 11 | 12 | [voices.town] 13 | ref_audio = "infer/examples/multi/town.flac" 14 | ref_text = "" 15 | 16 | [voices.country] 17 | ref_audio = "infer/examples/multi/country.flac" 18 | ref_text = "" 19 | 20 | -------------------------------------------------------------------------------- /xinference/thirdparty/f5_tts/infer/examples/multi/town.flac: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/f5_tts/infer/examples/multi/town.flac -------------------------------------------------------------------------------- /xinference/thirdparty/f5_tts/model/__init__.py: -------------------------------------------------------------------------------- 1 | from f5_tts.model.cfm import CFM 2 | 3 | from f5_tts.model.backbones.unett import UNetT 4 | from f5_tts.model.backbones.dit import DiT 5 | from f5_tts.model.backbones.mmdit import MMDiT 6 | 7 | # from f5_tts.model.trainer import Trainer 8 | 9 | 10 | __all__ = ["CFM", "UNetT", "DiT", "MMDiT"] # , "Trainer"] 11 | -------------------------------------------------------------------------------- /xinference/thirdparty/f5_tts/model/backbones/README.md: -------------------------------------------------------------------------------- 1 | ## Backbones quick introduction 2 | 3 | 4 | ### unett.py 5 | - flat unet transformer 6 | - structure same as in e2-tts & voicebox paper except using rotary pos emb 7 | - update: allow possible abs pos emb & convnextv2 blocks for embedded text before concat 8 | 9 | ### dit.py 10 | - adaln-zero dit 11 | - embedded timestep as condition 12 | - concatted noised_input + masked_cond + embedded_text, linear proj in 13 | - possible abs pos emb & convnextv2 blocks for embedded text before concat 14 | - possible long skip connection (first layer to last layer) 15 | 16 | ### mmdit.py 17 | - sd3 structure 18 | - timestep as condition 19 | - left stream: text embedded and applied a abs pos emb 20 | - right stream: masked_cond & noised_input concatted and with same conv pos emb as unett 21 | -------------------------------------------------------------------------------- /xinference/thirdparty/fish_speech/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/fish_speech/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/fish_speech/fish_speech/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/fish_speech/fish_speech/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/fish_speech/fish_speech/callbacks/__init__.py: -------------------------------------------------------------------------------- 1 | from .grad_norm import GradNormMonitor 2 | 3 | __all__ = ["GradNormMonitor"] 4 | -------------------------------------------------------------------------------- /xinference/thirdparty/fish_speech/fish_speech/configs/lora/r_8_alpha_16.yaml: -------------------------------------------------------------------------------- 1 | _target_: fish_speech.models.text2semantic.lora.LoraConfig 2 | r: 8 3 | lora_alpha: 16 4 | lora_dropout: 0.01 5 | -------------------------------------------------------------------------------- /xinference/thirdparty/fish_speech/fish_speech/datasets/protos/text-data.proto: -------------------------------------------------------------------------------- 1 | syntax = "proto3"; 2 | 3 | package text_data; 4 | 5 | message Semantics { 6 | repeated uint32 values = 1; 7 | } 8 | 9 | message Sentence { 10 | repeated string texts = 1; 11 | repeated Semantics semantics = 3; 12 | } 13 | 14 | message TextData { 15 | string source = 1; 16 | string name = 2; 17 | repeated Sentence sentences = 4; 18 | } 19 | 20 | message SampledData { 21 | string source = 1; 22 | string name = 2; 23 | repeated Sentence samples = 3; 24 | } 25 | -------------------------------------------------------------------------------- /xinference/thirdparty/fish_speech/fish_speech/datasets/protos/text_data_stream.py: -------------------------------------------------------------------------------- 1 | import struct 2 | 3 | from .text_data_pb2 import TextData 4 | 5 | 6 | def read_pb_stream(f): 7 | while True: 8 | buf = f.read(4) 9 | if len(buf) == 0: 10 | break 11 | size = struct.unpack("I", buf)[0] 12 | buf = f.read(size) 13 | text_data = TextData() 14 | text_data.ParseFromString(buf) 15 | yield text_data 16 | 17 | 18 | def write_pb_stream(f, text_data): 19 | buf = text_data.SerializeToString() 20 | f.write(struct.pack("I", len(buf))) 21 | f.write(buf) 22 | 23 | 24 | def pack_pb_stream(text_data): 25 | buf = text_data.SerializeToString() 26 | return struct.pack("I", len(buf)) + buf 27 | 28 | 29 | def split_pb_stream(f): 30 | while True: 31 | head = f.read(4) 32 | if len(head) == 0: 33 | break 34 | size = struct.unpack("I", head)[0] 35 | buf = f.read(size) 36 | yield head + buf 37 | -------------------------------------------------------------------------------- /xinference/thirdparty/fish_speech/fish_speech/i18n/__init__.py: -------------------------------------------------------------------------------- 1 | from .core import i18n 2 | 3 | __all__ = ["i18n"] 4 | -------------------------------------------------------------------------------- /xinference/thirdparty/fish_speech/fish_speech/models/text2semantic/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/fish_speech/fish_speech/models/text2semantic/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/fish_speech/fish_speech/models/vqgan/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/fish_speech/fish_speech/models/vqgan/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/fish_speech/fish_speech/text/__init__.py: -------------------------------------------------------------------------------- 1 | from .clean import clean_text 2 | from .spliter import split_text 3 | 4 | __all__ = ["clean_text", "split_text"] 5 | -------------------------------------------------------------------------------- /xinference/thirdparty/fish_speech/fish_speech/text/chn_text_norm/README.md: -------------------------------------------------------------------------------- 1 | # This account is no longer in use, see [Atomicoo](https://github.com/atomicoo) for my latest works. 2 | 3 | # Chn Text Norm 4 | 5 | this is a repository for chinese text normalization (no longer maintained). 6 | 7 | ## Quick Start ## 8 | 9 | ### Git Clone Repo ### 10 | 11 | git clone this repo to the root directory of your project which need to use it. 12 | 13 | cd /path/to/proj 14 | git clone https://github.com/Joee1995/chn-text-norm.git 15 | 16 | after that, your doc tree should be: 17 | ``` 18 | proj # root of your project 19 | |--- chn_text_norm # this chn-text-norm tool 20 | |--- text.py 21 | |--- ... 22 | |--- text_normalize.py # your text normalization code 23 | |--- ... 24 | ``` 25 | 26 | ### How to Use ? ### 27 | 28 | # text_normalize.py 29 | from chn_text_norm.text import * 30 | 31 | raw_text = 'your raw text' 32 | text = Text(raw_text=raw_text).normalize() 33 | 34 | ### How to add quantums ### 35 | 36 | 打开test.py,然后你就知道怎么做了。 37 | -------------------------------------------------------------------------------- /xinference/thirdparty/fish_speech/fish_speech/text/chn_text_norm/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/fish_speech/fish_speech/text/chn_text_norm/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/fish_speech/fish_speech/text/chn_text_norm/basic_constant.py: -------------------------------------------------------------------------------- 1 | # -*- coding: utf-8 -*- 2 | """基本常量 3 | 中文数字/数位/符号字符常量 4 | """ 5 | 6 | __author__ = "Zhiyang Zhou " 7 | __data__ = "2019-05-02" 8 | 9 | CHINESE_DIGIS = "零一二三四五六七八九" 10 | BIG_CHINESE_DIGIS_SIMPLIFIED = "零壹贰叁肆伍陆柒捌玖" 11 | BIG_CHINESE_DIGIS_TRADITIONAL = "零壹貳參肆伍陸柒捌玖" 12 | SMALLER_BIG_CHINESE_UNITS_SIMPLIFIED = "十百千万" 13 | SMALLER_BIG_CHINESE_UNITS_TRADITIONAL = "拾佰仟萬" 14 | LARGER_CHINESE_NUMERING_UNITS_SIMPLIFIED = "亿兆京垓秭穰沟涧正载" 15 | LARGER_CHINESE_NUMERING_UNITS_TRADITIONAL = "億兆京垓秭穰溝澗正載" 16 | SMALLER_CHINESE_NUMERING_UNITS_SIMPLIFIED = "十百千万" 17 | SMALLER_CHINESE_NUMERING_UNITS_TRADITIONAL = "拾佰仟萬" 18 | 19 | ZERO_ALT = "〇" 20 | ONE_ALT = "幺" 21 | TWO_ALTS = ["两", "兩"] 22 | 23 | POSITIVE = ["正", "正"] 24 | NEGATIVE = ["负", "負"] 25 | POINT = ["点", "點"] 26 | # PLUS = [u'加', u'加'] 27 | # SIL = [u'杠', u'槓'] 28 | 29 | # 中文数字系统类型 30 | NUMBERING_TYPES = ["low", "mid", "high"] 31 | -------------------------------------------------------------------------------- /xinference/thirdparty/fish_speech/fish_speech/text/chn_text_norm/cardinal.py: -------------------------------------------------------------------------------- 1 | # -*- coding: utf-8 -*- 2 | """CARDINAL类 (包含小数DECIMAL类) 3 | 纯数 <=> 中文字符串 方法 4 | 中文字符串 <=> 纯数 方法 5 | """ 6 | 7 | __author__ = "Zhiyang Zhou " 8 | __data__ = "2019-05-03" 9 | 10 | from fish_speech.text.chn_text_norm.basic_util import * 11 | 12 | 13 | class Cardinal: 14 | """ 15 | CARDINAL类 16 | """ 17 | 18 | def __init__(self, cardinal=None, chntext=None): 19 | self.cardinal = cardinal 20 | self.chntext = chntext 21 | 22 | def chntext2cardinal(self): 23 | return chn2num(self.chntext) 24 | 25 | def cardinal2chntext(self): 26 | return num2chn(self.cardinal) 27 | 28 | 29 | if __name__ == "__main__": 30 | 31 | # 测试程序 32 | print(Cardinal(cardinal="21357.230").cardinal2chntext()) 33 | -------------------------------------------------------------------------------- /xinference/thirdparty/fish_speech/fish_speech/text/chn_text_norm/digit.py: -------------------------------------------------------------------------------- 1 | # -*- coding: utf-8 -*- 2 | """DIGIT类 3 | 数字串 <=> 中文字符串 方法 4 | 中文字符串 <=> 数字串 方法 5 | """ 6 | 7 | __author__ = "Zhiyang Zhou " 8 | __data__ = "2019-05-03" 9 | 10 | from fish_speech.text.chn_text_norm.basic_util import * 11 | 12 | 13 | class Digit: 14 | """ 15 | DIGIT类 16 | """ 17 | 18 | def __init__(self, digit=None, chntext=None): 19 | self.digit = digit 20 | self.chntext = chntext 21 | 22 | # def chntext2digit(self): 23 | # return chn2num(self.chntext) 24 | 25 | def digit2chntext(self): 26 | return num2chn(self.digit, alt_two=False, use_units=False) 27 | 28 | 29 | if __name__ == "__main__": 30 | 31 | # 测试程序 32 | print(Digit(digit="2016").digit2chntext()) 33 | -------------------------------------------------------------------------------- /xinference/thirdparty/fish_speech/fish_speech/text/chn_text_norm/fraction.py: -------------------------------------------------------------------------------- 1 | # -*- coding: utf-8 -*- 2 | """FRACTION类 3 | 分数 <=> 中文字符串 方法 4 | 中文字符串 <=> 分数 方法 5 | """ 6 | 7 | __author__ = "Zhiyang Zhou " 8 | __data__ = "2019-05-03" 9 | 10 | from fish_speech.text.chn_text_norm.basic_util import * 11 | 12 | 13 | class Fraction: 14 | """ 15 | FRACTION类 16 | """ 17 | 18 | def __init__(self, fraction=None, chntext=None): 19 | self.fraction = fraction 20 | self.chntext = chntext 21 | 22 | def chntext2fraction(self): 23 | denominator, numerator = self.chntext.split("分之") 24 | return chn2num(numerator) + "/" + chn2num(denominator) 25 | 26 | def fraction2chntext(self): 27 | numerator, denominator = self.fraction.split("/") 28 | return num2chn(denominator) + "分之" + num2chn(numerator) 29 | 30 | 31 | if __name__ == "__main__": 32 | 33 | # 测试程序 34 | print(Fraction(fraction="2135/7230").fraction2chntext()) 35 | print(Fraction(chntext="五百八十一分之三百六十九").chntext2fraction()) 36 | -------------------------------------------------------------------------------- /xinference/thirdparty/fish_speech/fish_speech/text/chn_text_norm/percentage.py: -------------------------------------------------------------------------------- 1 | # -*- coding: utf-8 -*- 2 | """PERCENTAGE类 3 | 百分数 <=> 中文字符串 方法 4 | 中文字符串 <=> 百分数 方法 5 | """ 6 | 7 | __author__ = "Zhiyang Zhou " 8 | __data__ = "2019-05-06" 9 | 10 | from fish_speech.text.chn_text_norm.basic_util import * 11 | 12 | 13 | class Percentage: 14 | """ 15 | PERCENTAGE类 16 | """ 17 | 18 | def __init__(self, percentage=None, chntext=None): 19 | self.percentage = percentage 20 | self.chntext = chntext 21 | 22 | def chntext2percentage(self): 23 | return chn2num(self.chntext.strip().strip("百分之")) + "%" 24 | 25 | def percentage2chntext(self): 26 | return "百分之" + num2chn(self.percentage.strip().strip("%")) 27 | 28 | 29 | if __name__ == "__main__": 30 | 31 | # 测试程序 32 | print(Percentage(chntext="百分之五十六点零三").chntext2percentage()) 33 | print(Percentage(percentage="65.3%").percentage2chntext()) 34 | -------------------------------------------------------------------------------- /xinference/thirdparty/fish_speech/fish_speech/text/clean.py: -------------------------------------------------------------------------------- 1 | import re 2 | 3 | SYMBOLS_MAPPING = { 4 | "‘": "'", 5 | "’": "'", 6 | } 7 | 8 | REPLACE_SYMBOL_REGEX = re.compile( 9 | "|".join(re.escape(p) for p in SYMBOLS_MAPPING.keys()) 10 | ) 11 | 12 | 13 | EMOJI_REGEX = re.compile( 14 | "[" 15 | "\U0001F600-\U0001F64F" # emoticons 16 | "\U0001F300-\U0001F5FF" # symbols & pictographs 17 | "\U0001F680-\U0001F6FF" # transport & map symbols 18 | "\U0001F1E0-\U0001F1FF" # flags (iOS) 19 | "]+", 20 | flags=re.UNICODE, 21 | ) 22 | 23 | 24 | def clean_text(text): 25 | # Clean the text 26 | text = text.strip() 27 | 28 | # Replace all chinese symbols with their english counterparts 29 | text = REPLACE_SYMBOL_REGEX.sub(lambda x: SYMBOLS_MAPPING[x.group()], text) 30 | 31 | # Remove emojis 32 | text = EMOJI_REGEX.sub(r"", text) 33 | 34 | # Remove continuous periods (...) and commas (,,,) 35 | text = re.sub(r"[,]{2,}", lambda m: m.group()[0], text) 36 | 37 | return text 38 | -------------------------------------------------------------------------------- /xinference/thirdparty/fish_speech/fish_speech/utils/__init__.py: -------------------------------------------------------------------------------- 1 | from .braceexpand import braceexpand 2 | from .context import autocast_exclude_mps 3 | from .file import get_latest_checkpoint 4 | from .instantiators import instantiate_callbacks, instantiate_loggers 5 | from .logger import RankedLogger 6 | from .logging_utils import log_hyperparameters 7 | from .rich_utils import enforce_tags, print_config_tree 8 | from .utils import extras, get_metric_value, set_seed, task_wrapper 9 | 10 | __all__ = [ 11 | "enforce_tags", 12 | "extras", 13 | "get_metric_value", 14 | "RankedLogger", 15 | "instantiate_callbacks", 16 | "instantiate_loggers", 17 | "log_hyperparameters", 18 | "print_config_tree", 19 | "task_wrapper", 20 | "braceexpand", 21 | "get_latest_checkpoint", 22 | "autocast_exclude_mps", 23 | "set_seed", 24 | ] 25 | -------------------------------------------------------------------------------- /xinference/thirdparty/fish_speech/fish_speech/utils/context.py: -------------------------------------------------------------------------------- 1 | from contextlib import nullcontext 2 | 3 | import torch 4 | 5 | 6 | def autocast_exclude_mps( 7 | device_type: str, dtype: torch.dtype 8 | ) -> nullcontext | torch.autocast: 9 | return ( 10 | nullcontext() 11 | if torch.backends.mps.is_available() 12 | else torch.autocast(device_type, dtype) 13 | ) 14 | -------------------------------------------------------------------------------- /xinference/thirdparty/fish_speech/fish_speech/utils/file.py: -------------------------------------------------------------------------------- 1 | import os 2 | from pathlib import Path 3 | 4 | 5 | def get_latest_checkpoint(path: Path | str) -> Path | None: 6 | # Find the latest checkpoint 7 | ckpt_dir = Path(path) 8 | 9 | if ckpt_dir.exists() is False: 10 | return None 11 | 12 | ckpts = sorted(ckpt_dir.glob("*.ckpt"), key=os.path.getmtime) 13 | if len(ckpts) == 0: 14 | return None 15 | 16 | return ckpts[-1] 17 | -------------------------------------------------------------------------------- /xinference/thirdparty/fish_speech/fish_speech/webui/html/footer.html: -------------------------------------------------------------------------------- 1 |
2 | API 3 |  •  4 | Github 5 |  •  6 | Gradio 7 |
8 |
9 |
10 | {versions} 11 |
12 | -------------------------------------------------------------------------------- /xinference/thirdparty/fish_speech/tools/extract_model.py: -------------------------------------------------------------------------------- 1 | import click 2 | import torch 3 | from loguru import logger 4 | 5 | 6 | @click.command() 7 | @click.argument("model_path") 8 | @click.argument("output_path") 9 | def main(model_path, output_path): 10 | if model_path == output_path: 11 | logger.error("Model path and output path are the same") 12 | return 13 | 14 | logger.info(f"Loading model from {model_path}") 15 | state_dict = torch.load(model_path, map_location="cpu")["state_dict"] 16 | torch.save(state_dict, output_path) 17 | logger.info(f"Model saved to {output_path}") 18 | 19 | 20 | if __name__ == "__main__": 21 | main() 22 | -------------------------------------------------------------------------------- /xinference/thirdparty/fish_speech/tools/sensevoice/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/fish_speech/tools/sensevoice/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/fish_speech/tools/server/exception_handler.py: -------------------------------------------------------------------------------- 1 | import traceback 2 | from http import HTTPStatus 3 | 4 | from kui.asgi import HTTPException, JSONResponse 5 | 6 | 7 | class ExceptionHandler: 8 | 9 | async def http_exception_handler(self, exc: HTTPException): 10 | return JSONResponse( 11 | dict( 12 | statusCode=exc.status_code, 13 | message=exc.content, 14 | error=HTTPStatus(exc.status_code).phrase, 15 | ), 16 | exc.status_code, 17 | exc.headers, 18 | ) 19 | 20 | async def other_exception_handler(self, exc: Exception): 21 | traceback.print_exc() 22 | 23 | status = HTTPStatus.INTERNAL_SERVER_ERROR 24 | return JSONResponse( 25 | dict(statusCode=status, message=str(exc), error=status.phrase), 26 | status, 27 | ) 28 | -------------------------------------------------------------------------------- /xinference/thirdparty/fish_speech/tools/webui/variables.py: -------------------------------------------------------------------------------- 1 | from fish_speech.i18n import i18n 2 | 3 | HEADER_MD = f"""# Fish Speech 4 | 5 | {i18n("A text-to-speech model based on VQ-GAN and Llama developed by [Fish Audio](https://fish.audio).")} 6 | 7 | {i18n("You can find the source code [here](https://github.com/fishaudio/fish-speech) and models [here](https://huggingface.co/fishaudio/fish-speech-1.5).")} 8 | 9 | {i18n("Related code and weights are released under CC BY-NC-SA 4.0 License.")} 10 | 11 | {i18n("We are not responsible for any misuse of the model, please consider your local laws and regulations before using it.")} 12 | """ 13 | 14 | TEXTBOX_PLACEHOLDER = i18n("Put your text here.") 15 | -------------------------------------------------------------------------------- /xinference/thirdparty/internvl/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/internvl/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/llava/__init__.py: -------------------------------------------------------------------------------- 1 | from .model import LlavaLlamaForCausalLM 2 | -------------------------------------------------------------------------------- /xinference/thirdparty/llava/model/__init__.py: -------------------------------------------------------------------------------- 1 | from .llava_llama import LlavaConfig, LlavaLlamaForCausalLM 2 | -------------------------------------------------------------------------------- /xinference/thirdparty/llava/model/clip_encoder/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/llava/model/clip_encoder/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/llava/model/clip_encoder/builder.py: -------------------------------------------------------------------------------- 1 | from .clip_encoder import CLIPVisionTower 2 | 3 | 4 | def build_vision_tower(vision_tower_cfg, **kwargs): 5 | vision_tower = getattr( 6 | vision_tower_cfg, 7 | "mm_vision_tower", 8 | getattr(vision_tower_cfg, "vision_tower", None), 9 | ) 10 | 11 | return CLIPVisionTower(vision_tower, args=vision_tower_cfg, **kwargs) 12 | -------------------------------------------------------------------------------- /xinference/thirdparty/llava/model/constants.py: -------------------------------------------------------------------------------- 1 | # Model Constants 2 | IGNORE_INDEX = -100 3 | IMAGE_TOKEN_INDEX = -200 4 | DEFAULT_IMAGE_TOKEN = "" 5 | 6 | key_info = {"model_path": None} 7 | -------------------------------------------------------------------------------- /xinference/thirdparty/llava/model/multimodal_projector/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/llava/model/multimodal_projector/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/matcha/VERSION: -------------------------------------------------------------------------------- 1 | 0.0.7.0 2 | -------------------------------------------------------------------------------- /xinference/thirdparty/matcha/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/matcha/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/matcha/data/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/matcha/data/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/matcha/data/components/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/matcha/data/components/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/matcha/hifigan/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/matcha/hifigan/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/matcha/hifigan/config.py: -------------------------------------------------------------------------------- 1 | v1 = { 2 | "resblock": "1", 3 | "num_gpus": 0, 4 | "batch_size": 16, 5 | "learning_rate": 0.0004, 6 | "adam_b1": 0.8, 7 | "adam_b2": 0.99, 8 | "lr_decay": 0.999, 9 | "seed": 1234, 10 | "upsample_rates": [8, 8, 2, 2], 11 | "upsample_kernel_sizes": [16, 16, 4, 4], 12 | "upsample_initial_channel": 512, 13 | "resblock_kernel_sizes": [3, 7, 11], 14 | "resblock_dilation_sizes": [[1, 3, 5], [1, 3, 5], [1, 3, 5]], 15 | "resblock_initial_channel": 256, 16 | "segment_size": 8192, 17 | "num_mels": 80, 18 | "num_freq": 1025, 19 | "n_fft": 1024, 20 | "hop_size": 256, 21 | "win_size": 1024, 22 | "sampling_rate": 22050, 23 | "fmin": 0, 24 | "fmax": 8000, 25 | "fmax_loss": None, 26 | "num_workers": 4, 27 | "dist_config": {"dist_backend": "nccl", "dist_url": "tcp://localhost:54321", "world_size": 1}, 28 | } 29 | -------------------------------------------------------------------------------- /xinference/thirdparty/matcha/hifigan/env.py: -------------------------------------------------------------------------------- 1 | """ from https://github.com/jik876/hifi-gan """ 2 | 3 | import os 4 | import shutil 5 | 6 | 7 | class AttrDict(dict): 8 | def __init__(self, *args, **kwargs): 9 | super().__init__(*args, **kwargs) 10 | self.__dict__ = self 11 | 12 | 13 | def build_env(config, config_name, path): 14 | t_path = os.path.join(path, config_name) 15 | if config != t_path: 16 | os.makedirs(path, exist_ok=True) 17 | shutil.copyfile(config, os.path.join(path, config_name)) 18 | -------------------------------------------------------------------------------- /xinference/thirdparty/matcha/models/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/matcha/models/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/matcha/models/components/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/matcha/models/components/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/matcha/onnx/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/matcha/onnx/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/matcha/text/symbols.py: -------------------------------------------------------------------------------- 1 | """ from https://github.com/keithito/tacotron 2 | 3 | Defines the set of symbols used in text input to the model. 4 | """ 5 | _pad = "_" 6 | _punctuation = ';:,.!?¡¿—…"«»“” ' 7 | _letters = "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz" 8 | _letters_ipa = ( 9 | "ɑɐɒæɓʙβɔɕçɗɖðʤəɘɚɛɜɝɞɟʄɡɠɢʛɦɧħɥʜɨɪʝɭɬɫɮʟɱɯɰŋɳɲɴøɵɸθœɶʘɹɺɾɻʀʁɽʂʃʈʧʉʊʋⱱʌɣɤʍχʎʏʑʐʒʔʡʕʢǀǁǂǃˈˌːˑʼʴʰʱʲʷˠˤ˞↓↑→↗↘'̩'ᵻ" 10 | ) 11 | 12 | 13 | # Export all symbols: 14 | symbols = [_pad] + list(_punctuation) + list(_letters) + list(_letters_ipa) 15 | 16 | # Special symbol ids 17 | SPACE_ID = symbols.index(" ") 18 | -------------------------------------------------------------------------------- /xinference/thirdparty/matcha/utils/__init__.py: -------------------------------------------------------------------------------- 1 | from matcha.utils.instantiators import instantiate_callbacks, instantiate_loggers 2 | from matcha.utils.logging_utils import log_hyperparameters 3 | from matcha.utils.pylogger import get_pylogger 4 | from matcha.utils.rich_utils import enforce_tags, print_config_tree 5 | from matcha.utils.utils import extras, get_metric_value, task_wrapper 6 | -------------------------------------------------------------------------------- /xinference/thirdparty/matcha/utils/monotonic_align/__init__.py: -------------------------------------------------------------------------------- 1 | import numpy as np 2 | import torch 3 | 4 | from matcha.utils.monotonic_align.core import maximum_path_c 5 | 6 | 7 | def maximum_path(value, mask): 8 | """Cython optimised version. 9 | value: [b, t_x, t_y] 10 | mask: [b, t_x, t_y] 11 | """ 12 | value = value * mask 13 | device = value.device 14 | dtype = value.dtype 15 | value = value.data.cpu().numpy().astype(np.float32) 16 | path = np.zeros_like(value).astype(np.int32) 17 | mask = mask.data.cpu().numpy() 18 | 19 | t_x_max = mask.sum(1)[:, 0].astype(np.int32) 20 | t_y_max = mask.sum(2)[:, 0].astype(np.int32) 21 | maximum_path_c(path, value, t_x_max, t_y_max) 22 | return torch.from_numpy(path).to(device=device, dtype=dtype) 23 | -------------------------------------------------------------------------------- /xinference/thirdparty/matcha/utils/monotonic_align/setup.py: -------------------------------------------------------------------------------- 1 | # from distutils.core import setup 2 | # from Cython.Build import cythonize 3 | # import numpy 4 | 5 | # setup(name='monotonic_align', 6 | # ext_modules=cythonize("core.pyx"), 7 | # include_dirs=[numpy.get_include()]) 8 | -------------------------------------------------------------------------------- /xinference/thirdparty/matcha/utils/pylogger.py: -------------------------------------------------------------------------------- 1 | import logging 2 | 3 | from lightning.pytorch.utilities import rank_zero_only 4 | 5 | 6 | def get_pylogger(name: str = __name__) -> logging.Logger: 7 | """Initializes a multi-GPU-friendly python command line logger. 8 | 9 | :param name: The name of the logger, defaults to ``__name__``. 10 | 11 | :return: A logger object. 12 | """ 13 | logger = logging.getLogger(name) 14 | 15 | # this ensures all logging levels get marked with the rank zero decorator 16 | # otherwise logs would get multiplied for each GPU process in multi-GPU setup 17 | logging_levels = ("debug", "info", "warning", "error", "exception", "fatal", "critical") 18 | for level in logging_levels: 19 | setattr(logger, level, rank_zero_only(getattr(logger, level))) 20 | 21 | return logger 22 | -------------------------------------------------------------------------------- /xinference/thirdparty/megatts3/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/megatts3/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/melo/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/melo/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/melo/infer.py: -------------------------------------------------------------------------------- 1 | import os 2 | import click 3 | from melo.api import TTS 4 | 5 | 6 | 7 | @click.command() 8 | @click.option('--ckpt_path', '-m', type=str, default=None, help="Path to the checkpoint file") 9 | @click.option('--text', '-t', type=str, default=None, help="Text to speak") 10 | @click.option('--language', '-l', type=str, default="EN", help="Language of the model") 11 | @click.option('--output_dir', '-o', type=str, default="outputs", help="Path to the output") 12 | def main(ckpt_path, text, language, output_dir): 13 | if ckpt_path is None: 14 | raise ValueError("The model_path must be specified") 15 | 16 | config_path = os.path.join(os.path.dirname(ckpt_path), 'config.json') 17 | model = TTS(language=language, config_path=config_path, ckpt_path=ckpt_path) 18 | 19 | for spk_name, spk_id in model.hps.data.spk2id.items(): 20 | save_path = f'{output_dir}/{spk_name}/output.wav' 21 | os.makedirs(os.path.dirname(save_path), exist_ok=True) 22 | model.tts_to_file(text, spk_id, save_path) 23 | 24 | if __name__ == "__main__": 25 | main() 26 | -------------------------------------------------------------------------------- /xinference/thirdparty/melo/init_downloads.py: -------------------------------------------------------------------------------- 1 | 2 | 3 | if __name__ == '__main__': 4 | 5 | from melo.api import TTS 6 | device = 'auto' 7 | models = { 8 | 'EN': TTS(language='EN', device=device), 9 | 'ES': TTS(language='ES', device=device), 10 | 'FR': TTS(language='FR', device=device), 11 | 'ZH': TTS(language='ZH', device=device), 12 | 'JP': TTS(language='JP', device=device), 13 | 'KR': TTS(language='KR', device=device), 14 | } -------------------------------------------------------------------------------- /xinference/thirdparty/melo/monotonic_align/__init__.py: -------------------------------------------------------------------------------- 1 | from numpy import zeros, int32, float32 2 | from torch import from_numpy 3 | 4 | from .core import maximum_path_jit 5 | 6 | 7 | def maximum_path(neg_cent, mask): 8 | device = neg_cent.device 9 | dtype = neg_cent.dtype 10 | neg_cent = neg_cent.data.cpu().numpy().astype(float32) 11 | path = zeros(neg_cent.shape, dtype=int32) 12 | 13 | t_t_max = mask.sum(1)[:, 0].data.cpu().numpy().astype(int32) 14 | t_s_max = mask.sum(2)[:, 0].data.cpu().numpy().astype(int32) 15 | maximum_path_jit(path, neg_cent, t_t_max, t_s_max) 16 | return from_numpy(path).to(device=device, dtype=dtype) 17 | -------------------------------------------------------------------------------- /xinference/thirdparty/melo/text/cmudict_cache.pickle: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/melo/text/cmudict_cache.pickle -------------------------------------------------------------------------------- /xinference/thirdparty/melo/text/english_utils/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/melo/text/english_utils/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/melo/text/english_utils/abbreviations.py: -------------------------------------------------------------------------------- 1 | import re 2 | 3 | # List of (regular expression, replacement) pairs for abbreviations in english: 4 | abbreviations_en = [ 5 | (re.compile("\\b%s\\." % x[0], re.IGNORECASE), x[1]) 6 | for x in [ 7 | ("mrs", "misess"), 8 | ("mr", "mister"), 9 | ("dr", "doctor"), 10 | ("st", "saint"), 11 | ("co", "company"), 12 | ("jr", "junior"), 13 | ("maj", "major"), 14 | ("gen", "general"), 15 | ("drs", "doctors"), 16 | ("rev", "reverend"), 17 | ("lt", "lieutenant"), 18 | ("hon", "honorable"), 19 | ("sgt", "sergeant"), 20 | ("capt", "captain"), 21 | ("esq", "esquire"), 22 | ("ltd", "limited"), 23 | ("col", "colonel"), 24 | ("ft", "fort"), 25 | ] 26 | ] 27 | 28 | def expand_abbreviations(text, lang="en"): 29 | if lang == "en": 30 | _abbreviations = abbreviations_en 31 | else: 32 | raise NotImplementedError() 33 | for regex, replacement in _abbreviations: 34 | text = re.sub(regex, replacement, text) 35 | return text -------------------------------------------------------------------------------- /xinference/thirdparty/melo/text/es_phonemizer/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/melo/text/es_phonemizer/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/melo/text/es_phonemizer/es_symbols.txt: -------------------------------------------------------------------------------- 1 | _,.!?-~…NQabdefghijklmnopstuvwxyzɑæʃʑçɯɪɔɛɹðəɫɥɸʊɾʒθβŋɦ⁼ʰ`^#*=ˈˌ→↓↑ ɡrɲʝɣʎː—¿¡ -------------------------------------------------------------------------------- /xinference/thirdparty/melo/text/es_phonemizer/es_to_ipa.py: -------------------------------------------------------------------------------- 1 | from .cleaner import spanish_cleaners 2 | from .gruut_wrapper import Gruut 3 | 4 | def es2ipa(text): 5 | e = Gruut(language="es-es", keep_puncs=True, keep_stress=True, use_espeak_phonemes=True) 6 | # text = spanish_cleaners(text) 7 | phonemes = e.phonemize(text, separator="") 8 | return phonemes 9 | 10 | 11 | if __name__ == '__main__': 12 | print(es2ipa('¿Y a quién echaría de menos, en el mundo si no fuese a vos?')) -------------------------------------------------------------------------------- /xinference/thirdparty/melo/text/es_phonemizer/spanish_symbols.txt: -------------------------------------------------------------------------------- 1 | dˌaβˈiðkopeɾfjl unθsbmtʃwɛxɪŋʊɣɡrɲʝʎː -------------------------------------------------------------------------------- /xinference/thirdparty/melo/text/fr_phonemizer/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/melo/text/fr_phonemizer/__init__.py -------------------------------------------------------------------------------- /xinference/thirdparty/melo/text/fr_phonemizer/fr_to_ipa.py: -------------------------------------------------------------------------------- 1 | from .cleaner import french_cleaners 2 | from .gruut_wrapper import Gruut 3 | 4 | 5 | def remove_consecutive_t(input_str): 6 | result = [] 7 | count = 0 8 | 9 | for char in input_str: 10 | if char == 't': 11 | count += 1 12 | else: 13 | if count < 3: 14 | result.extend(['t'] * count) 15 | count = 0 16 | result.append(char) 17 | 18 | if count < 3: 19 | result.extend(['t'] * count) 20 | 21 | return ''.join(result) 22 | 23 | def fr2ipa(text): 24 | e = Gruut(language="fr-fr", keep_puncs=True, keep_stress=True, use_espeak_phonemes=True) 25 | # text = french_cleaners(text) 26 | phonemes = e.phonemize(text, separator="") 27 | # print(phonemes) 28 | phonemes = remove_consecutive_t(phonemes) 29 | # print(phonemes) 30 | return phonemes -------------------------------------------------------------------------------- /xinference/thirdparty/melo/text/fr_phonemizer/french_symbols.txt: -------------------------------------------------------------------------------- 1 | _,.!?-~…NQabdefghijklmnopstuvwxyzɑæʃʑçɯɪɔɛɹðəɫɥɸʊɾʒθβŋɦ⁼ʰ`^#*=ˈˌ→↓↑ ɣɡrɲʝʎː̃œøʁɒʌ—ɜɐ -------------------------------------------------------------------------------- /xinference/thirdparty/melo/text/ko_dictionary.py: -------------------------------------------------------------------------------- 1 | # coding: utf-8 2 | # Add the word you want to the dictionary. 3 | etc_dictionary = {"1+1": "원플러스원", "2+1": "투플러스원"} 4 | 5 | 6 | english_dictionary = { 7 | "KOREA": "코리아", 8 | "IDOL": "아이돌", 9 | "IT": "아이티", 10 | "IQ": "아이큐", 11 | "UP": "업", 12 | "DOWN": "다운", 13 | "PC": "피씨", 14 | "CCTV": "씨씨티비", 15 | "SNS": "에스엔에스", 16 | "AI": "에이아이", 17 | "CEO": "씨이오", 18 | "A": "에이", 19 | "B": "비", 20 | "C": "씨", 21 | "D": "디", 22 | "E": "이", 23 | "F": "에프", 24 | "G": "지", 25 | "H": "에이치", 26 | "I": "아이", 27 | "J": "제이", 28 | "K": "케이", 29 | "L": "엘", 30 | "M": "엠", 31 | "N": "엔", 32 | "O": "오", 33 | "P": "피", 34 | "Q": "큐", 35 | "R": "알", 36 | "S": "에스", 37 | "T": "티", 38 | "U": "유", 39 | "V": "브이", 40 | "W": "더블유", 41 | "X": "엑스", 42 | "Y": "와이", 43 | "Z": "제트", 44 | } 45 | -------------------------------------------------------------------------------- /xinference/thirdparty/melo/train.sh: -------------------------------------------------------------------------------- 1 | CONFIG=$1 2 | GPUS=$2 3 | MODEL_NAME=$(basename "$(dirname $CONFIG)") 4 | 5 | PORT=10902 6 | 7 | while : # auto-resume: the code sometimes crash due to bug of gloo on some gpus 8 | do 9 | torchrun --nproc_per_node=$GPUS \ 10 | --master_port=$PORT \ 11 | train.py --c $CONFIG --model $MODEL_NAME 12 | 13 | for PID in $(ps -aux | grep $CONFIG | grep python | awk '{print $2}') 14 | do 15 | echo $PID 16 | kill -9 $PID 17 | done 18 | sleep 30 19 | done -------------------------------------------------------------------------------- /xinference/thirdparty/mlx/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright 2022-2023 XProbe Inc. 2 | # 3 | # Licensed under the Apache License, Version 2.0 (the "License"); 4 | # you may not use this file except in compliance with the License. 5 | # You may obtain a copy of the License at 6 | # 7 | # http://www.apache.org/licenses/LICENSE-2.0 8 | # 9 | # Unless required by applicable law or agreed to in writing, software 10 | # distributed under the License is distributed on an "AS IS" BASIS, 11 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 | # See the License for the specific language governing permissions and 13 | # limitations under the License. 14 | -------------------------------------------------------------------------------- /xinference/thirdparty/mlx/flux/__init__.py: -------------------------------------------------------------------------------- 1 | # Copyright © 2024 Apple Inc. 2 | 3 | from .datasets import Dataset, load_dataset 4 | from .flux import FluxPipeline 5 | from .lora import LoRALinear 6 | from .sampler import FluxSampler 7 | from .trainer import Trainer 8 | from .utils import ( 9 | load_ae, 10 | load_clip, 11 | load_clip_tokenizer, 12 | load_flow_model, 13 | load_t5, 14 | load_t5_tokenizer, 15 | ) 16 | -------------------------------------------------------------------------------- /xinference/thirdparty/whisper/__main__.py: -------------------------------------------------------------------------------- 1 | from .transcribe import cli 2 | 3 | cli() 4 | -------------------------------------------------------------------------------- /xinference/thirdparty/whisper/assets/mel_filters.npz: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/thirdparty/whisper/assets/mel_filters.npz -------------------------------------------------------------------------------- /xinference/thirdparty/whisper/normalizers/__init__.py: -------------------------------------------------------------------------------- 1 | from .basic import BasicTextNormalizer as BasicTextNormalizer 2 | from .english import EnglishTextNormalizer as EnglishTextNormalizer 3 | -------------------------------------------------------------------------------- /xinference/thirdparty/whisper/version.py: -------------------------------------------------------------------------------- 1 | __version__ = "20231117" 2 | -------------------------------------------------------------------------------- /xinference/web/ui/.eslintignore: -------------------------------------------------------------------------------- 1 | .idea 2 | .github 3 | node_modules 4 | build 5 | public 6 | -------------------------------------------------------------------------------- /xinference/web/ui/.eslintrc.yml: -------------------------------------------------------------------------------- 1 | env: 2 | browser: true 3 | es2021: true 4 | node: true 5 | extends: 6 | - 'eslint:recommended' 7 | - 'plugin:react/recommended' 8 | - 'prettier' 9 | parserOptions: 10 | parser: '@babel/eslint-parser' 11 | requireConfigFile: false 12 | ecmaFeatures: 13 | jsx: true 14 | ecmaVersion: 12 15 | sourceType: module 16 | plugins: 17 | - react 18 | - simple-import-sort 19 | rules: 20 | new-cap: 'error' 21 | no-var: 'error' 22 | simple-import-sort/imports: 'error' 23 | simple-import-sort/exports: 'error' 24 | quote-props: ['error', 'consistent'] 25 | 'react/react-in-jsx-scope': 'off' 26 | 'react/prop-types': 'off' 27 | 'react/jsx-key': 'off' 28 | settings: 29 | react: 30 | version: 'detect' 31 | -------------------------------------------------------------------------------- /xinference/web/ui/.gitignore: -------------------------------------------------------------------------------- 1 | # See https://help.github.com/articles/ignoring-files/ for more about ignoring files. 2 | 3 | # dependencies 4 | /node_modules 5 | /.pnp 6 | .pnp.js 7 | 8 | # testing 9 | /coverage 10 | 11 | # production 12 | /build 13 | 14 | # misc 15 | .DS_Store 16 | .env.local 17 | .env.development.local 18 | .env.test.local 19 | .env.production.local 20 | 21 | npm-debug.log* 22 | yarn-debug.log* 23 | yarn-error.log* 24 | -------------------------------------------------------------------------------- /xinference/web/ui/.prettierignore: -------------------------------------------------------------------------------- 1 | .idea 2 | .github 3 | node_modules 4 | build 5 | public 6 | -------------------------------------------------------------------------------- /xinference/web/ui/.prettierrc.yml: -------------------------------------------------------------------------------- 1 | trailingComma: 'es5' 2 | tabWidth: 2 3 | semi: false 4 | singleQuote: true 5 | printWidth: 80 6 | bracketSpacing: true 7 | bracketSameLine: false 8 | arrowParens: 'always' 9 | quoteProps: 'consistent' 10 | -------------------------------------------------------------------------------- /xinference/web/ui/src/components/alertComponent.js: -------------------------------------------------------------------------------- 1 | import MuiAlert from '@mui/material/Alert' 2 | import React from 'react' 3 | 4 | const Alert = React.forwardRef(function Alert(props, ref) { 5 | return 6 | }) 7 | 8 | export { Alert } 9 | -------------------------------------------------------------------------------- /xinference/web/ui/src/components/apiContext.js: -------------------------------------------------------------------------------- 1 | import React, { createContext, useState } from 'react' 2 | 3 | import { getEndpoint } from './utils' 4 | 5 | export const ApiContext = createContext() 6 | 7 | export const ApiContextProvider = ({ children }) => { 8 | const [isCallingApi, setIsCallingApi] = useState(false) 9 | const [isUpdatingModel, setIsUpdatingModel] = useState(false) 10 | const [errorMsg, setErrorMsg] = useState('') 11 | const [successMsg, setSuccessMsg] = useState('') 12 | const endPoint = getEndpoint() 13 | 14 | return ( 15 | 28 | {children} 29 | 30 | ) 31 | } 32 | -------------------------------------------------------------------------------- /xinference/web/ui/src/components/copyComponent/style.css: -------------------------------------------------------------------------------- 1 | .copyText { 2 | font-size: 14px !important; 3 | color: #666; 4 | cursor: pointer; 5 | } 6 | 7 | .copyText:hover { 8 | color: #1976d2; 9 | } 10 | -------------------------------------------------------------------------------- /xinference/web/ui/src/components/errorMessageSnackBar.js: -------------------------------------------------------------------------------- 1 | import Snackbar from '@mui/material/Snackbar' 2 | import React, { useContext } from 'react' 3 | 4 | import { Alert } from './alertComponent' 5 | import { ApiContext } from './apiContext' 6 | 7 | const ErrorMessageSnackBar = () => { 8 | const { errorMsg, setErrorMsg } = useContext(ApiContext) 9 | 10 | const handleClose = (event, reason) => { 11 | if (reason === 'clickaway') { 12 | return 13 | } 14 | setErrorMsg('') 15 | } 16 | 17 | return ( 18 | 24 | 25 | {errorMsg} 26 | 27 | 28 | ) 29 | } 30 | 31 | export default ErrorMessageSnackBar 32 | -------------------------------------------------------------------------------- /xinference/web/ui/src/components/successMessageSnackBar.js: -------------------------------------------------------------------------------- 1 | import Snackbar from '@mui/material/Snackbar' 2 | import React, { useContext } from 'react' 3 | 4 | import { Alert } from './alertComponent' 5 | import { ApiContext } from './apiContext' 6 | 7 | const SuccessMessageSnackBar = () => { 8 | const { successMsg, setSuccessMsg } = useContext(ApiContext) 9 | 10 | const handleClose = (event, reason) => { 11 | if (reason === 'clickaway') { 12 | return 13 | } 14 | setSuccessMsg('') 15 | } 16 | 17 | return ( 18 | 24 | 25 | {successMsg} 26 | 27 | 28 | ) 29 | } 30 | 31 | export default SuccessMessageSnackBar 32 | -------------------------------------------------------------------------------- /xinference/web/ui/src/components/tableTitle.js: -------------------------------------------------------------------------------- 1 | import Typography from '@mui/material/Typography' 2 | import PropTypes from 'prop-types' 3 | import React from 'react' 4 | 5 | export default function TableTitle(props) { 6 | return ( 7 | 13 | {props.children} 14 | 15 | ) 16 | } 17 | 18 | TableTitle.propTypes = { 19 | component: PropTypes.elementType, 20 | children: PropTypes.node, 21 | } 22 | -------------------------------------------------------------------------------- /xinference/web/ui/src/components/themeButton.js: -------------------------------------------------------------------------------- 1 | import DarkModeIcon from '@mui/icons-material/DarkMode' 2 | import LightModeIcon from '@mui/icons-material/LightMode' 3 | import { Box, IconButton } from '@mui/material' 4 | import React from 'react' 5 | 6 | import { useThemeContext } from './themeContext' 7 | 8 | const ThemeButton = ({ sx }) => { 9 | const { themeMode, toggleTheme } = useThemeContext() 10 | 11 | return ( 12 | 13 | 14 | {themeMode === 'light' ? : } 15 | 16 | 17 | ) 18 | } 19 | 20 | export default ThemeButton 21 | -------------------------------------------------------------------------------- /xinference/web/ui/src/components/titleTypography.js: -------------------------------------------------------------------------------- 1 | import Typography from '@mui/material/Typography' 2 | import React from 'react' 3 | 4 | const h2Style = { 5 | margin: '10px 10px', 6 | fontSize: '20px', 7 | fontWeight: 'bold', 8 | } 9 | 10 | export default function TitleTypography({ value }) { 11 | return ( 12 | 19 | {value} 20 | 21 | ) 22 | } 23 | -------------------------------------------------------------------------------- /xinference/web/ui/src/components/utils.js: -------------------------------------------------------------------------------- 1 | const getEndpoint = () => { 2 | let endPoint = '' 3 | if (!process.env.NODE_ENV || process.env.NODE_ENV === 'development') { 4 | endPoint = 'http://127.0.0.1:9997' 5 | } else { 6 | const fullUrl = window.location.href 7 | endPoint = fullUrl.split('/ui')[0] 8 | } 9 | return endPoint 10 | } 11 | 12 | const isValidBearerToken = (token) => { 13 | return ( 14 | token !== '' && token !== undefined && token !== null && token.length > 10 15 | ) 16 | } 17 | 18 | const toReadableSize = (size) => { 19 | const res_size = size / 1024.0 ** 2 20 | return res_size.toFixed(2) + 'MiB' 21 | } 22 | 23 | export { getEndpoint, isValidBearerToken, toReadableSize } 24 | -------------------------------------------------------------------------------- /xinference/web/ui/src/components/versionLabel.js: -------------------------------------------------------------------------------- 1 | import { Box, Typography } from '@mui/material' 2 | import React, { useEffect, useState } from 'react' 3 | 4 | import fetchWrapper from '../components/fetchWrapper' 5 | 6 | const VersionLabel = ({ sx }) => { 7 | const [version, setVersion] = useState('') 8 | 9 | useEffect(() => { 10 | fetchWrapper 11 | .get('/v1/cluster/version') 12 | .then((data) => { 13 | setVersion('v' + data['version']) 14 | }) 15 | .catch((error) => { 16 | console.error('Error:', error) 17 | }) 18 | }, []) 19 | 20 | return ( 21 | 22 | {version} 23 | 24 | ) 25 | } 26 | 27 | export default VersionLabel 28 | -------------------------------------------------------------------------------- /xinference/web/ui/src/i18n.js: -------------------------------------------------------------------------------- 1 | import i18n from 'i18next' 2 | import { initReactI18next } from 'react-i18next' 3 | 4 | import en from './locales/en.json' 5 | import ja from './locales/ja.json' 6 | import ko from './locales/ko.json' 7 | import zh from './locales/zh.json' 8 | 9 | i18n.use(initReactI18next).init({ 10 | fallbackLng: 'en', 11 | lng: localStorage.getItem('language') || 'en', 12 | debug: true, 13 | interpolation: { 14 | escapeValue: false, 15 | }, 16 | resources: { 17 | en: { translation: en }, 18 | zh: { translation: zh }, 19 | ja: { translation: ja }, 20 | ko: { translation: ko }, 21 | }, 22 | }) 23 | 24 | i18n.on('languageChanged', (lng) => { 25 | localStorage.setItem('language', lng) 26 | }) 27 | 28 | export default i18n 29 | -------------------------------------------------------------------------------- /xinference/web/ui/src/index.css: -------------------------------------------------------------------------------- 1 | @import url('https://fonts.googleapis.com/css2?family=Inter:wght@400;600;700&display=swap'); 2 | 3 | html, 4 | body, 5 | #root, 6 | .app { 7 | height: 100%; 8 | width: 100%; 9 | font-family: 'Inter', sans-serif; 10 | } 11 | 12 | ::-webkit-scrollbar { 13 | width: 8px; 14 | } 15 | 16 | /* Track */ 17 | ::-webkit-scrollbar-track { 18 | background: #00000000; 19 | } 20 | 21 | /* Handle */ 22 | ::-webkit-scrollbar-thumb { 23 | background: #7c767640; 24 | border-radius: 120px; 25 | } 26 | 27 | /* Handle on Hover */ 28 | ::-webkit-scrollbar-track:hover { 29 | background: #b3b4ba10; 30 | } 31 | -------------------------------------------------------------------------------- /xinference/web/ui/src/index.js: -------------------------------------------------------------------------------- 1 | import './i18n' 2 | 3 | import React from 'react' 4 | import { CookiesProvider } from 'react-cookie' 5 | import ReactDOM from 'react-dom/client' 6 | 7 | import App from './App' 8 | 9 | const root = ReactDOM.createRoot(document.getElementById('root')) 10 | root.render( 11 | 12 | 13 | 14 | 15 | 16 | ) 17 | -------------------------------------------------------------------------------- /xinference/web/ui/src/media/icon.webp: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/xorbitsai/inference/1d2b6948947fb6870f0a61d82d5f4d9e312e3f3b/xinference/web/ui/src/media/icon.webp -------------------------------------------------------------------------------- /xinference/web/ui/src/scenes/_layout/index.js: -------------------------------------------------------------------------------- 1 | import { Box } from '@mui/material' 2 | import React from 'react' 3 | import { Outlet } from 'react-router-dom' 4 | 5 | import MenuSide from '../../components/MenuSide' 6 | 7 | const Layout = () => { 8 | return ( 9 | 10 | 11 | 12 | 13 | 14 | 15 | ) 16 | } 17 | 18 | export default Layout 19 | -------------------------------------------------------------------------------- /xinference/web/ui/src/scenes/cluster_info/style.js: -------------------------------------------------------------------------------- 1 | import { createTheme, tableCellClasses } from '@mui/material' 2 | import Paper from '@mui/material/Paper' 3 | import TableCell from '@mui/material/TableCell' 4 | import TableRow from '@mui/material/TableRow' 5 | import { styled } from '@mui/system' 6 | 7 | export const theme = createTheme() 8 | 9 | export const StyledTableCell = styled(TableCell)(() => ({ 10 | [`&.${tableCellClasses.body}`]: { 11 | fontSize: 14, 12 | }, 13 | })) 14 | 15 | export const StyledTableRow = styled(TableRow)(({ theme }) => ({ 16 | '&:nth-of-type(odd)': { 17 | backgroundColor: theme.palette.action.hover, 18 | }, 19 | // hide last border 20 | '&:last-child td, &:last-child th': { 21 | border: 0, 22 | }, 23 | })) 24 | 25 | export const StyledPaper = styled(Paper)({ 26 | padding: theme.spacing(2), 27 | display: 'flex', 28 | overflow: 'auto', 29 | flexDirection: 'column', 30 | }) 31 | -------------------------------------------------------------------------------- /xinference/web/ui/src/scenes/launch_model/components/drawer.js: -------------------------------------------------------------------------------- 1 | import { useTheme } from '@mui/material' 2 | import React, { useEffect } from 'react' 3 | 4 | const Drawer = ({ isOpen, onClose, children }) => { 5 | const theme = useTheme() 6 | 7 | useEffect(() => { 8 | if (isOpen) { 9 | document.body.style.overflow = 'hidden' 10 | } else { 11 | document.body.style.overflow = '' 12 | } 13 | 14 | return () => { 15 | document.body.style.overflow = '' 16 | } 17 | }, [isOpen]) 18 | 19 | return ( 20 |
21 |
22 |
28 | {isOpen && children} 29 |
30 |
31 | ) 32 | } 33 | 34 | export default Drawer 35 | -------------------------------------------------------------------------------- /xinference/web/ui/src/scenes/launch_model/components/progress.js: -------------------------------------------------------------------------------- 1 | import LinearProgress from '@mui/material/LinearProgress' 2 | import React from 'react' 3 | 4 | const Progress = ({ progress }) => { 5 | return ( 6 |
7 | 8 |
9 | ) 10 | } 11 | 12 | export default Progress 13 | -------------------------------------------------------------------------------- /xinference/web/ui/src/scenes/login/header.js: -------------------------------------------------------------------------------- 1 | import { AppBar, Box, Toolbar } from '@mui/material' 2 | import Typography from '@mui/material/Typography' 3 | import * as React from 'react' 4 | 5 | import icon from '../../media/icon.webp' 6 | 7 | export default function Header() { 8 | return ( 9 | theme.zIndex.drawer + 1, 17 | }} 18 | > 19 | 20 | 29 | 30 | 31 | {'Xinference'} 32 | 33 | 34 | 35 | 36 | ) 37 | } 38 | --------------------------------------------------------------------------------