Data Engineer (Remote, working in the US time zone)
This a Full Remote job, the offer is available from: Indiana (USA) This is a remote position. We are looking for an experienced Senior Data Engineer to join a global pharmaceutical client on a long-term project. The role focuses on designing, building, and optimizing scalable, reliable data pipelines and data platforms that support advanced analytics, experimentation, and data-driven decision-making across manufacturing and business domains. This is a senior-level position requiring deep technical expertise, strong problem-solving skills, and the ability to collaborate effectively with cross-functional teams in a regulated environment. Key Responsibilities • Design, build, and maintain scalable and reliable data pipelines using modern data engineering best practices • Develop and optimize ETL/ELT pipelines for structured and semi-structured data • Work extensively with SQL and Python or Scala to process and transform large datasets • Build and manage data solutions on AWS cloud infrastructure • Leverage big data technologies such as Spark, Kafka, and Hadoop • Design and maintain data models and data warehousing solutions • Orchestrate data workflows using tools such as Apache Airflow • Ensure data integrity, quality, and accuracy through validation, cleaning, and preprocessing techniques • Optimize data collection, processing, and storage for performance, scalability, and cost efficiency • Design and conduct experiments to test hypotheses and validate data-driven solutions • Collaborate with analytics, data science, engineering, and business stakeholders to understand data needs and align solutions with organizational goals • Create templates, dashboards, and visualizations to communicate insights clearly to stakeholders • Apply and support data governance, security, and compliance standards, especially in a regulated pharmaceutical environment Requirements • 10+ years of experience in Data Engineering or similar roles • Strong expertise in SQL and Python or Scala • Hands-on experience with AWS (profiles with strong AWS exposure will be prioritized) • Solid experience with Spark, Kafka, and Hadoop • Proven experience designing and implementing ETL/ELT pipelines • Strong understanding of data modeling, data warehousing concepts, and analytics use cases • Experience with workflow orchestration tools such as Apache Airflow • Strong analytical thinking and problem-solving skills • Excellent communication and collaboration skills • Ability to work independently in a remote, distributed team Nice to Have • Experience in the pharmaceutical or life sciences industry (top priority) • Knowledge of manufacturing processes and manufacturing data • Experience working in regulated environments (GxP, compliance-driven projects) Benefits • Project Duration: 10 months+ • Work Model: Remote (US time zone collaboration required) • Location: Open to candidates based in US or Europe • Contract Type: B2B / Contract This offer from "futureproof consulting" has been enriched by Jobgether.com and got a 74% flex score. Apply tot his job