Contract
On-site

We are looking for a highly motivated and skilled Data Engineer

Key Responsibilities

  • Build and maintain robust, scalable ETL pipelines across batch and real-time data sources.
  • Design and implement data transformations using Spark (PySpark/Scala/Java) on Hadoop/Hive.
  • Stream data from Kafka topics into data lakes or analytics layers using Spark Streaming.
  • Collaborate with cross-functional teams on data modeling, ingestion strategies, and performance optimization.
  • Implement and support CI/CD pipelines using Git, Jenkins, and container technologies like Docker/Kubernetes.
  • Work within cloud and on-prem hybrid data platforms, contributing to automation, deployment, and monitoring of data workflows.

Skills

  • Strong programming skills in Python, Scala, or Java.
  • Hands-on experience with Apache Spark, Hadoop, Hive, Kafka, HBase, or related tools.
  • Sound understanding of data warehousing, dimensional modeling, and SQL.
  • Familiarity with Airflow, Git, Jenkins, and containerization tools (Docker/Kubernetes).
  • Exposure to cloud platforms such as AWS or GCP is a plus.
  • Experience with Agile delivery models and collaborative tools like Jira and Confluence.

Nice to Have

  • Experience with streaming data pipelines, machine learning workflows, or feature engineering.
  • Familiarity with Terraform, Ansible, or other infrastructure-as-code tools.
  • Exposure to Snowflake, Databricks, or modern data lakehouse architecture is a bonus.