{"version":"1.0","workflow_uuid":"20408a97-5017-11f1-9bc6-00163e2b0d79","workflow_title":"mistral.rs — Blazingly Fast LLM Inference in Rust","install_contract":{"version":"1.0","installReady":false,"title":"mistral.rs — Blazingly Fast LLM Inference in Rust","summary":"mistral.rs is a cross-platform LLM inference engine written in Rust that supports 40+ model families including text, vision, and speech. It provides OpenAI-compatible APIs, quantization, PagedAttention, and both Rust and Python SDKs.","assetType":"Configs","pageUrl":"https://tokrepo.com/en/workflows/asset-20408a97","sourceUrl":"http://localhost:1234/v1/chat/completions","intendedFor":[],"firstActions":[],"agentFirstSteps":[],"targetPaths":[],"verification":[],"startingPoints":[],"example":"","successOutcome":"","boundaries":[],"askUserIf":["the current workspace stack cannot be matched to a safe upstream template","the target path is not the project root, or an existing file should be merged instead of overwritten"]}}