Top
Middle Data Engineer
NEWRemote, hubs in Dubai, Yerevan, ...Full-timeGlobal
š Midš Remote
RemoteRemote work position availableActivePosted within the last 30 days
Job Description
[AI-summarized by JobStash]
You will build and maintain data pipelines and data-related services that power analytics across products. You will extend SQL-based pipelines, add new data sources to ETL processes, refactor Python scripts into modular production-quality code, configure CI for linting and tests, and update pipeline documentation. You will work closely with senior engineers and analysts, ask clarifying questions, and gradually take ownership of data platform components while learning modern data workflows.
Requirements
- āConfident communication and proactive clarification seeking
- āResponsibility, ownership, and proactive communication of challenges
- āComfortable with IDEs and version control systems like Git
- āBasic understanding of clean code principles and software delivery workflows
- āEssential Python skills including language fundamentals and data structures
- āConfident with SQL basics
- āRegular and thoughtful use of AI tools
- āMotivation to learn and grow in data engineering
- āKnowledge of data engineering fundamentals including ETL, data modeling, data quality, and storage systems
- āExperience using Apache Airflow
- āFamiliarity with containerization tools such as Docker
- āExposure to cloud platforms such as GCP
- āBasic experience with cloud platforms (GCP, AWS, or Azure) is a plus
- āUnderstanding of orchestration tools such as Airflow is a plus
- āBasic Docker usage (build, run, logs) is a plus
- āExperience with BI tools (Superset, Metabase, Power BI) is a plus
- āPersonal data projects (ETL scripts, dashboards, analytics) are a plus
Responsibilities
- āBuild and maintain data pipelines and data-related services
- āContribute to shared tools and libraries
- āUpgrade data platform components and services
- āCommunicate with analysts to understand data needs
- āExtend SQL-based pipelines with new transformations
- āAdd new data sources to ETL processes
- āRefactor Python scripts into modular code and add logging
- āConfigure CI for linting and tests for data repositories
- āUpdate pipeline documentation after logic or schema changes
Benefits & Perks
- āRemote work setup with access to hubs in Dubai, Yerevan, London and Belgrade
- āCompensation for medical expenses
- āProvision of necessary equipment
- ā20 working days of paid vacation annually
- ā11 days off per year
- ā14 days of paid sick leave
- āAccess to internal conferences
- āAccess to English courses
- āAccess to corporate events
- āRegular performance reviews
Tech Stack
data pipelineloggingPythonSQLETLCIdata modelingtestingGCPGitproject:The Open Platform