Junior Data Engineer - Mobile Apps

Leadtech·Remote(Spain)
Software Development

WFA Digital Insight

As the demand for skilled data engineers continues to rise, with a 25% growth in remote data engineering roles over the past year, professionals with expertise in cloud-native data solutions are becoming increasingly sought after. Leadtech, a company at the forefront of digital innovation, is now hiring a Junior Data Engineer to join their team in Spain. This role stands out for its focus on building robust and cost-effective data pipelines, making it an exciting opportunity for those looking to grow their skills in data architecture and pipeline development. With the right combination of technical expertise and passion for data engineering, candidates can excel in this position and contribute to the company's continued success.

Job Description

About the Role

The Junior Data Engineer role at Leadtech is a unique opportunity to work on designing, developing, and optimizing the company's data infrastructure on Databricks. As a key member of the data engineering team, you will be responsible for architecting pipelines using BigQuery, Google Cloud Storage, Apache Airflow, dbt, Dataflow, and Pub/Sub, ensuring high availability and performance across ETL/ELT processes. This role requires a strong understanding of cloud-native data solutions, experience with ETL/ELT frameworks, and a passion for building robust and cost-effective pipelines.

The ideal candidate will have a solid foundation in data engineering principles, excellent problem-solving skills, and the ability to work collaboratively with cross-functional teams. As a Junior Data Engineer at Leadtech, you will have the opportunity to work with cutting-edge technologies, develop your skills in data architecture and pipeline development, and contribute to the company's mission to drive innovation through data-driven decision making.

What You Will Do

  • Support the development and maintenance of the data platform on GCP, including data warehousing in BigQuery/Databricks and data lake storage in Google Cloud Storage.
  • Help organize data into clear layers and domain-focused Data Marts for analytics and reporting.
  • Assist with Terraform-based Infrastructure as Code to provision and manage cloud resources in a consistent way.
  • Contribute to batch and near real-time data workflows with a focus on reliability, scalability, and cost awareness.
  • Build, maintain, and improve ETL/ELT pipelines under guidance using Apache Airflow for workflow orchestration.
  • Develop and maintain dbt transformations to create clean, version-controlled data models in BigQuery.
  • Support data ingestion and processing using tools such as Google Dataflow, Apache Beam, or Pub/Sub where needed.
  • Monitor scheduled jobs, troubleshoot failures, and help ensure data is delivered on time for analytics and reporting.
  • Help implement and maintain data quality checks using Great Expectations, dbt tests, or similar tools.
  • Support documentation of datasets, metadata, lineage, and audit processes.
  • Follow security best practices, including IAM, encryption, and secure handling of sensitive data.
  • Assist in maintaining compliance with data privacy and governance requirements such as GDPR or CCPA.

What We Are Looking For

  • 1+ year of experience in data engineering or a related data role.
  • Exposure to mobile, product, or marketing data is a plus.
  • Basic hands-on experience with GCP services such as BigQuery and Google Cloud Storage.
  • Familiarity with Apache Airflow for scheduling and orchestrating data workflows.
  • Some experience with dbt or similar transformation tools.
  • Exposure to Pub/Sub, Dataflow, or other batch/streaming tools is a plus.
  • Understanding of Data Mart concepts and interest in Infrastructure as Code tools such as Terraform.
  • Good coding skills in Python; Java or Scala is a plus.
  • Ability to write scripts for automation and data processing tasks.
  • Familiarity with Docker and basic container concepts.
  • Exposure to CI/CD and version control workflows such as GitHub Actions, GitLab CI, Jenkins, or similar.

Nice to Have

  • Interest in machine learning workflows and exposure to tools such as Vertex AI or similar ML platforms.
  • Familiarity with monitoring and observability tools such as Prometheus.
  • Experience with data governance concepts such as lineage, metadata, and access control.

Benefits and Perks

  • Competitive salary and benefits package.
  • Opportunity to work with cutting-edge technologies and contribute to the company's mission.
  • Collaborative and dynamic work environment.
  • Professional development and growth opportunities.
  • Flexible working hours and remote work options.
  • Access to the latest tools and technologies.
  • Recognition and reward for outstanding performance.

How to Stand Out

  • Ensure you have a strong foundation in data engineering principles and experience with ETL/ELT frameworks.
  • Familiarize yourself with Leadtech's technology stack and be prepared to discuss your experience with similar tools.
  • Highlight your ability to work collaboratively with cross-functional teams and your passion for building robust and cost-effective pipelines.
  • Be prepared to discuss your experience with data quality checks and data governance concepts.
  • Research the company's mission and values, and be prepared to discuss how your skills and experience align with them.
  • Consider creating a portfolio that showcases your data engineering projects and achievements.
  • Prepare to discuss your experience with security best practices and handling sensitive data.

This is a remote position listed on WFA Digital, the platform for professionals who work from anywhere. Browse more remote jobs across all categories.