Lead Cloud Data Engineer - Remote Opportunity

Remote Full-time
Job DescriptionWe are building a next-generation Cloud Data Platform to unify data from Policy, Claims, Billing, and Administration systems into a single source of truth. We are seeking a Lead Cloud Data Engineer who will be 75% hand-on and play a critical role in designing, building, and optimizing our modern data ecosystem leveraging Medallion architecture, bolthires Lake, and modern data warehouse technologies such as Snowflake, Synapse, or Redshift. As a technical leader, the Lead Data Engineer will define and execute the end-to-end data engineering strategy from data ingestion and modeling to governance and performance optimization enabling scalable, high-quality, and analytics-ready data assets.This role requires hands on deep technical expertise in cloud-native data engineering, automation, and architecture design, coupled with strong leadership to mentor teams and align solutions with business goals. ResponsibilitiesData Platform Design & Architecture• Define the strategic roadmap for the enterprise data platform, ensuring scalability, performance, and interoperability across business domains. • Architect and implement cloud-native, Medallion-based data architectures (Bronze–Silver–Gold layers) for unified and governed data delivery.• Drive standardization of data models, pipelines, and quality frameworks across Policy, Claims, Billing, and Administrative data assets. • Evaluate and implement emerging data technologies to strengthen the platform’s performance, bolthires efficiency, and resilienceData Integration & Ingestion• Design, build, and optimize high-performance ingestion pipelines, using AWS Glue, Databricks, or custom Spark applications. • Automate ingestion of structured, semi-structured, and unstructured data from APIs databases, and external data feeds.• Tune and monitor ingestion pipelines for throughput, bolthires control, and reliability across dev/test/prod environments. Data Transformation & Modeling• Hands on Development of ETL/ELT pipelines using Databricks or similar frameworks to transform raw data into curated and consumption-ready datasets. • Design and develop relational, Vault, and dimensional data models to support analytics, BI, and AI/ML workloads. • Define and enforce data quality standards, validation frameworks, and enrichment rules to ensure trusted business data.• Apply data quality, cleansing, and enrichment logic to ensure accuracy and completeness of business-critical data. Cloud Infrastructure, Automation and Performance• Collaborate with DevOps and Cloud Engineering teams to design automated, infrastructure-as-code environments using Terraform, CloudFormation, or equivalent tools. • Implement bolthires/CD pipelines for data pipeline deployment, versioning, and testing. • Lead performance tuning and scalability optimization to ensure highly available, bolthires-efficient data platform.Governance, Security & Compliance• Implement and enforce data governance, cataloging, and lineage practices using tools such as Purview, Alation, or Collibra. • Partner with InfoSec to implement data privacy, access control, and compliance frameworks aligned with regulatory standards. • Drive consistency and accountability in data stewardship across business and IT teams. Leadership, Collaboration & Mentorship• Lead a team of data engineers, providing technical guidance, coaching, and performance mentorship.• Collaborate with Data Architects, Analysts, and Business Leaders to align data solutions with enterprise strategy. • Promote a culture of engineering excellence, reusability, and knowledge sharing across data organization. • Influence enterprise-wide standards for data engineering, automation, and governance. Skills/RequirementsQualifications:• Bachelor’s or master’s degree in computer science, Data Engineering, or a related field. • 12+ years of experience in data engineering with at 3+ years in a lead or architect-level role and least 8+ years on cloud platforms (AWS, Azure, or GCP).• Deep hands-on experience with Python, SQL, and data modeling (relational, and Dimensional), Databricks, Spark, AWS Glue, bolthires Lake, Snowflake, Synapse, or Redshift• Proven experience with Medallion architecture, modern data warehousing principles., data governance, lineage, and bolthires/CD for data pipelines• Excellent leadership, communication, and cross-functional collaboration skills. • Experience in the Property & Casualty (P&C) Insurance domain such as Policy, Claims, or Billing data preferred.• Familiarity with event-driven architectures (Kafka, Kinesis) and real-time data streaming. • Knowledge of machine learning pipeline integration and feature engineering. • Proven ability to lead large-scale data modernization or cloud migration initiatives. Compensation• $140,000 - $165,000 commensurate with experience, plus bonus eligibilityBenefitsWe are proud to offer a robust benefits suite that includes:• Competitive base salary plus incentive plans for eligible team members• 401(K) retirement plan that includes a company match of up to 6% of your eligible salary• Free basic life and AD&D, long-term disability and short-term disability insurance• Medical, dental and vision plans to meet your unique healthcare needs• Wellness incentives• Generous time off program that includes personal, holiday and volunteer paid time off• Flexible work schedules and hybrid/remote options for eligible positions• Educational assistance#TMG Apply tot his job
Apply Now →
← Back to Home