SJ
SponsoredJobs
.online
← Back to All Jobs

EDL Tech lead

Sign in to view the employer name
🇺🇸 Dallas, Texas, United States
Posted 27 Apr 2026

Location

Job description

Enterprise Data Lake Technical Lead Any Visa works Relocation works C2C works H1b works Hybrid 3 days onsite in Dallas Texas Direct Client Role 12+ Months Contract The Enterprise Data Lake (EDL) Technical Lead is responsible for owning the design, implementation, and engineering leadership of the enterprise data lake platform. This role ensures the platform is scalable, reliable, secure, and consumable across analytics, reporting, and operational workloads. The Tech Lead drives technical decisions and collaborates closely with Architecture, Product Management, Data Engineering, Security, and downstream consumers to deliver robust data platform capabilities. Enterprise Data Lake Technical Lead Primary Skills : Apache Kafka Databricks platform DataHub or Collibra/Alation/Apache Atlas RDBMS and Modern Data lake Kubernetes - Functional knowledge Secondary : Cloud Platform (Azure and/ or AWS) C# and or Go or java Terraform ELT kind of servicesPosition Overview The Enterprise Data Lake (EDL) Technical Lead is responsible for owning the design, implementation, and engineering leadership of the enterprise data lake platform. This role ensures the platform is scalable, reliable, secure, and consumable across analytics, reporting, and operational workloads. The Tech Lead drives technical decisions and collaborates closely with Architecture, Product Management, Data Engineering, Security, and downstream consumers to deliver robust data platform capabilities. Key Responsibilities Technical Leadership
  • Own the design and implementation of enterprise-scale data lake solutions on Azure/AWS Cloud
  • Define technical standards and best practices for data platform components
  • Drive technical decision-making across data ingestion, storage, processing, and governance layers
  • Provide technical mentorship and guidance to data engineering and platform teams
  • Collaborate with Architecture team on overall architectural alignment and strategy
  • Partner with Product Management, Data Engineering, Security, SRE, and downstream consumers

Data Governance & Catalog Management (Critical)
  • Own the design and implementation of enterprise data catalog and metadata management layer
  • Build and operate data catalog using DataHub or equivalent metadata platform
  • Design and implement automated metadata ingestion, lineage tracking, and data discovery capabilities
  • Design and implement data governance policies, data quality rules, and compliance frameworks
  • Enable self-service data discovery and access management for downstream consumers
  • Establish data ownership, stewardship models, and metadata standards across the enterprise
  • Build APIs and services for metadata-driven data operations and workflows

Platform Engineering & Development
  • Design and implement real-time and batch data ingestion pipelines using Apache Kafka
  • Optimize data processing workflows on Databricks platform (Delta Lake, Spark optimization, Unity Catalog)
  • Design integration patterns between RDBMS sources and data lake (CDC, batch ETL, replication)
  • Optimize database performance, query tuning, and indexing strategies across relational and distributed systems
  • Implement infrastructure as code using Terraform for automated provisioning and management
  • Design and deploy containerized data services on Kubernetes clusters
  • Develop data platform services and tooling using C# and Go
  • Build database migration tools, data validation frameworks, and connectivity services

Operations & Reliability
  • Ensure platform scalability, reliability, and security across all data lake components
  • Implement monitoring, logging, and observability solutions for data infrastructure
  • Optimize Linux-based systems for data processing workloads
  • Establish CI/CD pipelines for data platform deployments
  • Ensure platform security, compliance, and data governance standards (GDPR, SOC2, etc.)
  • Drive cost optimization and performance tuning initiatives
  • Implement database backup/recovery strategies and disaster recovery planning

Required Qualifications Technical Expertise Data Streaming & Processing
  • 5+ years with Apache Kafka (streaming architecture, Kafka Connect, Schema Registry, stream processing)
  • 3+ years with Databricks (Delta Lake, Apache Spark optimization, Unity Catalog, cluster management)
  • Deep understanding of data lake architecture patterns (Bronze/Silver/Gold, medallion architecture)

Data Governance & Cataloging
  • 3+ years with DataHub or similar metadata management platforms (Alation, Collibra, Apache Atlas)
  • Deep experience building and operating enterprise data catalog systems
  • Expertise in automated metadata extraction, lineage tracking, and impact analysis
  • Experience with data quality frameworks and metadata-driven data operations
  • Knowledge of data governance policies, data classification, and compliance automation
  • Understanding of data discovery, access control, and self-service analytics enablement

Relational Database Systems
  • 5+ years with enterprise RDBMS platforms including:
  • SQL Server (T-SQL, SSIS, SSRS, replication, Always On Availability Groups)
  • PostgreSQL (advanced query optimization, partitioning, extensions, streaming replication)
  • MySQL (replication, clustering, performance tuning)
  • Strong SQL skills (complex queries, stored procedures, window functions, query optimization)
  • Database design principles (normalization, indexing strategies, schema design, partitioning)
  • Change Data Capture (CDC) patterns and implementation (Debezium, Azure Data Factory, AWS DMS, custom solutions)
  • Database migration experience (schema migration, data migration, zero-downtime migrations)

Cloud & Infrastructure
  • 5+ years with Azure or AWS Cloud including:
  • Azure: Data Lake Storage (Gen2), Event Hubs, AKS, Azure SQL, Key Vault, Azure AD, Monitor
  • AWS: S3, MSK/Kinesis, EKS, RDS/Aurora, Secrets Manager, IAM, CloudWatch
  • Cloud-native data services, networking, security, and IAM
  • 3+ years with Kubernetes (deployment strategies, scaling, monitoring, service mesh, Helm)
  • 3+ years with Terraform (modules, state management, multi-environment deployments, multi-cloud)

Programming & Development
  • Strong proficiency in any programming languages like C#, Go
  • Expert-level SQL across multiple database platforms
  • Experience with Python for data engineering tasks (preferred)
  • Familiarity with Shell scripting (Bash)

Leadership & Experience
  • 7+ years in data engineering, platform engineering, or database engineering roles
  • 3+ years in technical leadership capacity (Tech Lead, Principal Engineer)
  • Proven track record of delivering large-scale data infrastructure projects
  • Experience leading teams of 5-10+ engineers
  • Strong architectural design and system thinking capabilities
  • Experience migrating legacy RDBMS workloads to modern data lake architectures
  • Demonstrated ability to balance technical excellence with business needs

Preferred Qualifications Additional Technologies
  • Experience with cloud ETL services (Azure Data Factory, AWS Glue, Azure Stream Analytics, AWS Lambda)
  • Knowledge of managed database features (Azure SQL elastic pools/hyperscale, AWS RDS/Aurora serverless)
  • Knowledge of additional streaming technologies (Apache Flink)
  • Experience with database sharding and horizontal partitioning strategies at scale
  • Familiarity with NoSQL databases (Cosmos DB, DynamoDB, MongoDB, Cassandra, Redis)
  • Experience with Apache Iceberg
  • Knowledge of data observability tools
  • Education: Any Bachelors or Similar

Job details

EmployerSign in to view the employer name
LocationDallas, Texas, United States
Posted27 Apr 2026
SalaryNot specified
SponsorshipVisa Sponsored ✓
Categories
Information Technology

Prep this job with AI