{"version":"1.0","workflow_uuid":"2db0b23f-39ec-11f1-9bc6-00163e2b0d79","workflow_title":"Apache Hudi — Incremental Data Processing for Data Lakehouses","install_contract":{"version":"1.0","installReady":false,"title":"Apache Hudi — Incremental Data Processing for Data Lakehouses","summary":"Apache Hudi (Hadoop Upserts Deletes and Incrementals) is an open-source data lakehouse platform that provides record-level insert, update, and delete capabilities on data lakes. It powers incremental pipelines, CDC ingestion, and near-real-time analytics on S3, GCS, and HDFS.","assetType":"Configs","pageUrl":"https://tokrepo.com/en/workflows/2db0b23f-39ec-11f1-9bc6-00163e2b0d79","sourceUrl":"https://github.com/apache/hudi","intendedFor":[],"firstActions":[],"agentFirstSteps":[],"targetPaths":[],"verification":[],"startingPoints":[],"example":"","successOutcome":"","boundaries":[],"askUserIf":["the current workspace stack cannot be matched to a safe upstream template","the target path is not the project root, or an existing file should be merged instead of overwritten"]}}