{"version":"1.0","workflow_uuid":"0e90de1c-3d9d-11f1-9bc6-00163e2b0d79","workflow_title":"ONNX Runtime — Cross-Platform ML Model Inference Engine","install_contract":{"version":"1.0","installReady":false,"title":"ONNX Runtime — Cross-Platform ML Model Inference Engine","summary":"ONNX Runtime is a high-performance inference engine for machine learning models in the ONNX format. Developed by Microsoft, it accelerates model serving across CPU, GPU, and specialized hardware with a unified API for Python, C++, C#, Java, and JavaScript.","assetType":"Scripts","pageUrl":"https://tokrepo.com/en/workflows/0e90de1c-3d9d-11f1-9bc6-00163e2b0d79","sourceUrl":"https://github.com/microsoft/onnxruntime","intendedFor":[],"firstActions":[],"agentFirstSteps":[],"targetPaths":[],"verification":[],"startingPoints":[],"example":"","successOutcome":"","boundaries":[],"askUserIf":["the current workspace stack cannot be matched to a safe upstream template","the target path is not the project root, or an existing file should be merged instead of overwritten"]}}