{"version":"1.0","workflow_uuid":"617d3446-4cd0-11f1-9bc6-00163e2b0d79","workflow_title":"ONNX Runtime — Cross-Platform ML Inference Accelerator","install_contract":{"version":"1.0","installReady":false,"title":"ONNX Runtime — Cross-Platform ML Inference Accelerator","summary":"A high-performance inference engine for ONNX models that runs on CPU, GPU, and specialized hardware across cloud, edge, and mobile.","assetType":"Configs","pageUrl":"https://tokrepo.com/en/workflows/asset-617d3446","sourceUrl":"https://github.com/microsoft/onnxruntime","intendedFor":[],"firstActions":[],"agentFirstSteps":[],"targetPaths":[],"verification":[],"startingPoints":[],"example":"","successOutcome":"","boundaries":[],"askUserIf":["the current workspace stack cannot be matched to a safe upstream template","the target path is not the project root, or an existing file should be merged instead of overwritten"]}}