{"version":"1.0","workflow_uuid":"8a35faad-0abb-421b-ae18-e869579db4b4","workflow_title":"Helicone — LLM Observability and Prompt Management","install_contract":{"version":"1.0","installReady":false,"title":"Helicone — LLM Observability and Prompt Management","summary":"Open-source LLM observability platform. One-line proxy integration for request logging, cost tracking, caching, rate limiting, and prompt versioning across all providers.","assetType":"Configs","pageUrl":"https://tokrepo.com/en/workflows/8a35faad-0abb-421b-ae18-e869579db4b4","sourceUrl":"https://oai.helicone.ai/v1","intendedFor":[],"firstActions":[],"agentFirstSteps":[],"targetPaths":[],"verification":[],"startingPoints":[],"example":"","successOutcome":"","boundaries":[],"askUserIf":["the current workspace stack cannot be matched to a safe upstream template","the target path is not the project root, or an existing file should be merged instead of overwritten"]}}