[< BACK]
// POSTED: May 5, 2026

Technical Program Manager, Data Lake

APPLY NOW
This a Full Remote job, the offer is available from: Missouri (USA) Role Summary Ansira is consolidating and scaling a unified, enterprise Data Lake to integrate product, media, and business data, standardize reporting, and accelerate decision-making across the organization. We are seeking a Technical Program Manager (TPM) to lead this cross-functional program end-to-end — aligning product and engineering roadmaps, driving ingestion and migration from legacy systems, maturing data governance and quality, and ensuring business adoption of standardized, self-serve analytics. This leader will orchestrate work across Ansira’s Product solutions, Data Engineering, Data Science/BI, Media, and Client Partnership teams, with a clear mandate: deliver consistent, governed, and performant data to downstream products and reporting while deprecating redundant systems and minimizing operational cost. What You’ll Do Program Leadership & Delivery • Own the multi-quarter program plan for the unified Data Lake: scope, roadmap, milestones, budgets (OPEX/CAPEX), risks, and dependencies. • Stand up and run the operating model: weekly workstream standups, cross-functional syncs, monthly steering committee, and a transparent executive status rhythm. • Build and maintain a single-source-of-truth for delivery: program charter, RACI, RAID log, decision log, intake/triage process, and dashboards for progress/risks. • Drive the migration plan from legacy pipelines and tools (e.g., Alteryx, Insighter) to the target stack (e.g., Snowflake, Power BI embedded via platform connectors). • Coordinate parallel workstreams (ingestion, modeling, governance, reporting cutover) to hit time-bound deliverables with predictable quality. Product Management & Roadmap • Define and maintain the Data Lake program backlog, translating business use cases into technical epics, data contracts, and acceptance criteria. • Partner with Product and Data Science teams to standardize media and product reporting packages and ensure they’re backed by governed, contract-driven data. • Prioritize sources and domains for ingestion based on business value, client impact, and technical feasibility; establish clear go/no-go gates. • Align with platform architecture to ensure scalable patterns for batch/stream ELT/CDC, cost control, observability, and reusability across domains. Data Governance, Quality, and Security • Establish practical data contracts with upstream product and business owners; define schema, SLAs, lineage, and DQ checks at ingestion. • Stand up governance ceremonies and roles (data owners, stewards) and implement data catalog/lineage practices to improve discoverability and trust. • Define and monitor quality KPIs (completeness, timeliness, accuracy) and drive remediation plans with accountable teams. • Ensure data privacy, compliance, and security best practices (e.g., PII handling, role-based access, data masking) across environments. Stakeholder Management & Change Adoption • Serve as the connective tissue across Product, Engineering, Data Science, Media, Finance, and Client Partnership — communicating decisions, trade-offs, and timelines. • Lead change management for reporting standardization (e.g., Media (AdTech/LBN)-based standard reports), business onboarding to the lake, and client-facing cutovers. • Create enablement assets (runbooks, playbooks, onboarding guides) and training plans to accelerate adoption and reduce support burden. Technical Fluency • Partner effectively with architects and data engineers on Snowflake/BigQuery/Databricks, Azure/AWS/GCP services, orchestration (ADF/Airflow), and transformation (dbt). • Understand ELT/CDC patterns, API/file ingestion, schema design for analytics, and BI tooling (Power BI, Looker). Write and review basic SQL for validation. • Apply FinOps and performance/cost optimization practices (storage tiers, compute sizing, job scheduling, caching strategies). Minimum Qualifications • 8+ years in Program/Project/Product Management, with 5+ years leading complex data platform initiatives in a cloud environment. • Proven delivery of cross-functional data programs involving multiple product lines and business stakeholders; strong executive communication. • Hands-on experience with modern data stacks: one or more of Snowflake/BigQuery/Databricks; Azure Data Factory/Airflow; dbt; Kafka/Kinesis; Git/Terraform; REST/SFTP integrations. • Strong grounding in data governance and quality practices, data contracts, catalog/lineage, and secure data access. • Demonstrated expertise in Agile at scale (Scrum/Kanban), Jira/Confluence, dependency/risk management, and budget tracking (including CAPEX/OPEX). • Competent SQL skills for validation/triage; fluency in reading pipeline/log artifacts and interpreting BI/semantic model requirements. Preferred Qualifications • Background in marketing/media data and standardized performance reporting (e.g., Media (AdTech/LBN), campaign hierarchies, Power BI embedded). • Prior experience migrating from legacy ETL/BI ecosystems (e.g., Alteryx/Insighter/Tableau) to a lakehouse with standardized semantic layers. • Experience establishing data domains and productizing data (SLAs, contracts, versioning, lifecycle) to accelerate downstream analytics. • Familiarity with privacy, security, and compliance standards (e.g., RBAC/ABAC, PII governance) and enterprise SSO/permissions models for embedded analytics. • FinOps mindset: cost observability, unit economics, and right-sizing compute/storage. Success Metrics (What Great Looks Like) • Sources Onboarded: number of prioritized sources/domains ingested to the lake with production-grade contracts and SLAs. • Time-to-Data: cycle time from intake approval to governed data available for downstream consumption. • Data Quality & Reliability: sustained improvement in DQ scorecards; incident MTTR reduction; SLA adherence for freshness and availability. • Migration Progress: percent of targeted legacy pipelines and reports decommissioned; client reporting cutovers delivered on schedule. • Adoption & Reuse: growth in standardized reporting/package usage; reduction in ad hoc one-off pipelines. • Cost & Performance: measurable storage/compute cost per use case; query performance improvements aligned to agreed SLOs. 30/60/90-Day Plan • 30 Days: Confirm program charter, governance model, and delivery rhythm. Baseline current-state architecture, sources, and reporting dependencies. Publish the first integrated roadmap/milestone plan and RAID. • 60 Days: Land 1–2 ingestion patterns as reusable templates (contracts, lineage, observability). Execute the first cutover to standardized reporting for a targeted use case. Stand up DQ scorecards and weekly KPI review. • 90 Days: Complete a tranche of source onboardings and at least one significant reporting migration. Retire targeted legacy jobs. Publish quarterly executive readout with outcomes, cost/performance improvements, and next-wave priorities. Tools You’ll Use • Jira, Confluence, Power BI; Snowflake/BigQuery/Databricks; Azure/AWS/GCP services (e.g., ADF, IAM); Airflow; dbt; Git; catalog/lineage tools. Why This Role This is a high-visibility opportunity to unify Ansira’s data foundation and materially improve the speed, quality, and consistency of insights across products and services. You will own the operating model, drive cross-functional alignment, and deliver tangible business outcomes through a modern, governed data platform. This offer from "Ansira" has been enriched by Jobgether.com and got a 75% flex score. Apply tot his job
Interested in this role?Apply on iHire