We are seeking a highly skilled and motivated Data Engineer with over 5 years of experience in big data technologies, specializing in Spark, Python, and Hadoop. The ideal candidate will have a strong track record of designing, building, and maintaining scalable data pipelines and solutions. In this role, you will develop and optimize robust data workflows using Apache Spark and Python, ensuring data transformation, cleaning, and enrichment processes meet the needs of analytics and machine learning initiatives. You will also design and implement scalable solutions leveraging the Hadoop ecosystem while collaborating with cross-functional teams to gather requirements and ensure seamless data integration.
Key responsibilities include monitoring and troubleshooting production pipelines to ensure reliability and performance, implementing best practices for data governance, security, and compliance, and optimizing workflows for maximum efficiency. Candidates should have a bachelor’s degree in Computer Science, Engineering, or a related field, with strong expertise in distributed computing principles and big data architecture. Experience with ETL processes, data warehousing, and cloud platforms (AWS, Azure, or GCP) is a plus. If you are passionate about big data technologies and enjoy problem-solving in a collaborative environment, we invite you to apply and contribute to our dynamic and innovative team.