# ModelScope — Open-Source Model Hub and ML Framework > ModelScope provides a unified Python interface for downloading, running, and fine-tuning thousands of pre-trained models across NLP, computer vision, audio, and multimodal tasks. ## Install Save in your project root: # ModelScope — Open-Source Model Hub and ML Framework ## Quick Use ```bash pip install modelscope ``` ```python from modelscope.pipelines import pipeline p = pipeline("text-generation", model="qwen/Qwen2-7B-Instruct") result = p("Explain what ModelScope does in one sentence.") print(result) ``` ## Introduction ModelScope is an open-source platform and Python library for accessing pre-trained models. It hosts thousands of models for NLP, vision, audio, and multimodal tasks, and provides a consistent API for inference, training, and evaluation regardless of the underlying framework. ## What ModelScope Does - Hosts thousands of pre-trained models with one-line download and inference - Provides a unified pipeline API that works across PyTorch, TensorFlow, and ONNX - Supports fine-tuning with built-in trainers and dataset loaders - Includes model evaluation tools with standard benchmarks - Offers a model card system with documentation, metrics, and licensing info ## Architecture Overview ModelScope wraps models from various frameworks behind a pipeline abstraction. When you call pipeline() with a task name and model ID, the library resolves the model checkpoint from the hub, loads the appropriate preprocessor and postprocessor, and returns a callable object. The hub layer handles model versioning, caching, and access control. Trainers extend PyTorch training loops with dataset integration and metric computation. ## Self-Hosting & Configuration - Install from PyPI; models are cached locally after first download - Set MODELSCOPE_CACHE to control where model files are stored - Configure hub access tokens for gated or private models - Use the modelscope server command to host a local model registry - Fine-tune with the built-in Trainer class or export models for external training ## Key Features - Unified pipeline API makes switching between models a one-line change - Thousands of models spanning text, image, audio, video, and multimodal tasks - Built-in dataset library with preprocessing pipelines for common benchmarks - Framework-agnostic design supports PyTorch, TensorFlow, and ONNX models - Active community with regular model updates and new architecture support ## Comparison with Similar Tools - **Hugging Face** — larger global community and model count; ModelScope has stronger coverage of Chinese-language and multimodal models - **TorchHub** — PyTorch-only; no built-in training or evaluation pipeline - **ONNX Model Zoo** — inference-focused; no fine-tuning or dataset integration - **Replicate** — cloud API for running models; not self-hostable as a library ## FAQ **Q: How does ModelScope compare to Hugging Face?** A: Both host pre-trained models with a Python API. ModelScope has particularly strong coverage of Chinese-language models and multimodal architectures. **Q: Can I use models offline?** A: Yes. Once downloaded, models are cached locally and work without internet access. **Q: Does it support GPU acceleration?** A: Yes. Models automatically use CUDA when available. You can specify device placement in the pipeline constructor. **Q: Can I upload my own models?** A: Yes. The hub supports model uploads with documentation, licensing, and versioning through the web interface or CLI. ## Sources - https://github.com/modelscope/modelscope - https://modelscope.cn/ --- Source: https://tokrepo.com/en/workflows/2f0e6af0-3fdb-11f1-9bc6-00163e2b0d79 Author: AI Open Source