Job Summary
We are seeking a skilled Data Engineer to design, develop, and maintain scalable data pipelines and infrastructure. The ideal candidate will have hands-on experience with Big Data technologies, strong programming skills, and the ability to work effectively in cloud environments such as AWS and GCP. This role requires collaboration with data scientists, analysts, and engineers to deliver high-quality data solutions that support business needs.
Key Responsibilities
Develop, test, and deploy robust data pipelines using technologies including Hadoop, HDFS, Spark, Scala, and Python. You will be responsible for writing efficient, clean, and maintainable code primarily in Java, Scala, and Python to ensure optimal performance and scalability.
Design and implement data models and SQL queries that enable efficient data access and analysis. Your work will involve building data solutions within cloud platforms like AWS and GCP, leveraging their services to enhance data processing and storage capabilities.
Contribute actively to the development and ongoing maintenance of data infrastructure, ensuring its reliability and scalability. Participate in code reviews to uphold coding standards, improve code quality, and promote best practices across the team.
Collaborate closely with data scientists, analysts, and fellow engineers to understand data requirements and translate them into technical solutions. Troubleshoot and resolve issues related to data pipelines and applications promptly to minimize downtime.
Document data pipelines, workflows, and system architecture clearly to facilitate knowledge sharing and future development. Stay informed about emerging Big Data technologies and industry trends to continuously improve our data engineering practices.
Required Qualifications
Strong proficiency in Java programming is essential for this role, alongside solid experience with Big Data technologies such as Hadoop and HDFS. You should have hands-on expertise in Spark and Scala, enabling you to build efficient data processing workflows.
Proficiency in Python programming is required to support various data engineering tasks and automation. A solid understanding of SQL and relational database concepts is necessary for designing effective data models and queries.
Experience working within cloud environments, specifically AWS or GCP, is critical for building and deploying scalable data solutions. Familiarity with Unix/Linux operating systems is also important for managing and troubleshooting data infrastructure.
Excellent problem-solving and analytical skills are needed to address complex data challenges. Strong communication and collaboration abilities will enable you to work effectively both independently and as part of a multidisciplinary team.
Preferred Qualifications and Benefits
While not explicitly listed, candidates with additional certifications in cloud platforms or Big Data technologies will have an advantage. The role offers opportunities to work with cutting-edge technologies and contribute to impactful data projects in a dynamic environment.
We provide a collaborative workplace that encourages continuous learning and professional growth, supporting your development as a data engineering expert.
---
This position is ideal for a motivated data engineer eager to leverage their technical skills and cloud experience to build scalable, high-performance data solutions. If you are passionate about Big Data and enjoy working in a fast-paced, team-oriented setting, we encourage you to apply.