Big Data Engineer (Back fill) @ Remote, Richmond VA (C2C conract)
REQUIREMENTData Engineer (Back fill)Remote / Richmond VALong term contractNeed Ex-Capital One consultants is MUSTFast Path -Data Engineer(Backfill Role)Main Skillset: Big Data engineering, Spark, Hadoop, Hive, HDFS, Kafka, AWS EMR and S3, ETL pipelines, distributed systems, NoSQL databases, and strong Java/Scala programming. What Manager likes a true Big Data engineer who understands the internals of Spark and Hadoop, can optimize large pipelines, and has strong AWS and Java/Scala skills to deliver production-ready systems at scale.Location, Richmond, orRemote is fine. Job Description:We are seeking a highly skilled Big Data Engineer with deep expertise in building, optimizing, and scaling data pipelines across distributed systems. The ideal candidate has hands-on experience with Spark, Hadoop, Hive, Kafka, AWS EMR, and S3, along with strong programming skills in Java or Scala. This role requires someone who not only builds data workflows but also understands the internals of big data frameworks to deliver high-performance, production-ready solutions.Key Responsibilities:• Design, develop, and optimize large-scale ETL pipelines using Spark, Hive, and Hadoop. • Implement and manage distributed data processing systems on AWS EMR, S3, and related cloud data services. • Work with Kafka and streaming data frameworks to handle real-time data ingestion and processing. • Optimize performance of Spark jobs and Hadoop clusters to improve reliability and throughput. • Collaborate with data scientists, analysts, and product teams to deliver scalable and maintainable data solutions.• Develop data models and integrate structured and unstructured data using NoSQL databases. • Ensure best practices for data quality, governance, and security in the data engineering lifecycle. Required Skills &Experience:• Strong programming skills in Java and/or Scala. • Deep understanding of Apache Spark internals, Hadoop architecture, and distributed computing concepts. • Hands-on experience with AWS services EMR, S3, Lambda, and related tools. • Proficiency with Hive, HDFS, Kafka, and ETL frameworks.• Experience working with NoSQL databases (e.g., HBase, Cassandra, DynamoDB). • Proven ability to design, deploy, and optimize large-scale production data systems. PreferredQualifications:• Experience with Airflow or similar workflow orchestration tools. • Familiarity with CI/CD pipelines and data infrastructure automation. • Strong problem-solving skills and ability to tune large distributed systems for performance. Thanks & Regards,John Stanley Sr. BDM / Delivery ManagerMaintec Technologies Inc8801 Fast Park Drive, Ste.301, Raleigh, NC 27617Mobile: +1 (919) 267-1887 / +91- 98411-45549Email: [email protected]; www.maintec.in | www.maintec.comLinkedIn : www.linkedin.com/in/johnstanley1/Bangalore | Chennai | Hyderabad | Pune | Noida | USA Apply tot his job