logo

Leading Data Engineer

Apply

Responsibilities

  • Design, build, and operate Spark Structured Streaming pipelines (event-driven, CDC, API-based)
  • Implement and maintain Slowly Changing Dimensions (Type 1/2) at scale
  • Own data modeling and table design using Apache Iceberg (partitioning, compaction, schema evolution)
  • Ensure pipeline reliability, SLAs, and data correctness across batch and streaming workloads
  • Develop and operate Airflow-orchestrated data workflows
  • Collaborate closely with Senior Data Engineers and DataOps on platform stability and performance
  • Enforce data access controls, governance, and secure data sharing

Requirements

·       5+ years experience in data engineering

·       2+ years experience in production Spark Streaming

·       Deep understanding of Iceberg internals, lakehouse patterns, and large-scale analytics

·       Proven experience implementing SCDs in distributed systems

·       Strong Python & SQL, comfortable reading JVM-level Spark behavior

·       Experience running Spark on Kubernetes

·       Operational mindset: monitoring, alerting, incident response

·       Zero tolerance for fragile pipelines or undocumented logic

Nice to have

·    CDC systems (Debezium, Kafka-based ingestion)

·    Performance tuning under cost and latency constraints

·    On-prem or hybrid data platforms

This role is not for dashboard builders or pure analysts.
You will own core data flows in production.

Conditions

  • Work from one of the best business centers in Baku
  • Competitive market-level salary
  • Private medical insurance, paid vacation, sick leave, corporate events and bonuses
  • Opportunity to influence the product, open-minded management, and a strong, energetic team
  • Be part of first ever ecosystem of Azerbaijan
  • Be part of “Dəyərlisən”
  • Shuttle service
Share this job opening

Application:

I agree to the processing of my personal data in accordance with the PashaPay Privacy Policy