Senior Data Engineer (GCP/Databricks)

Job not on LinkedIn

July 30

Apply Now
Logo of Leadtech Group

Leadtech Group

B2B • SaaS • eCommerce

Leadtech Group is a company that excels in transforming ideas into successful online projects with a focus on creativity, diversity, and fast-paced development. With their headquarters in Barcelona, Spain, Leadtech specializes in numerous online services including UX design, SEO, SEM, social media management, content creation, customer service, project management, and system administration. The company is known for its dynamic and diverse team of fun-loving, hard-working professionals who aim to exceed contemporary standards by leveraging the connectivity of the internet to enhance user experiences. Leadtech is committed to providing innovative solutions and seeks to continuously grow and achieve success by embracing a pioneering mindset.

201 - 500 employees

Founded 2009

🤝 B2B

☁️ SaaS

🛍️ eCommerce

📋 Description

• We are looking for a Senior Data Engineer to design, develop, and optimize our data infrastructure on Google Cloud Platform (GCP). • You will architect scalable pipelines using Databricks, BigQuery, Google Cloud Storage, Apache Airflow, dbt, Dataflow, and Pub/Sub, ensuring high availability and performance across our ETL/ELT processes. • You will leverage great expectations to enforce data quality standards. • The role also involves building our Data Mart (Data Mach) environment, containerizing services with Docker and Kubernetes (K8s), and implementing CI/CD best practices. • A successful candidate has extensive knowledge of cloud-native data solutions, strong proficiency with ETL/ELT frameworks (including dbt), and a passion for building robust, cost-effective pipelines. • Define and implement the overall data architecture on GCP, including data warehousing in BigQuery, data lake patterns in Google Cloud Storage, and Data Mart (Data Mach) solutions. • Integrate Terraform for Infrastructure as Code to provision and manage cloud resources efficiently. • Establish both batch and real-time data processing frameworks to ensure reliability, scalability, and cost efficiency. • Design, build, and optimize ETL/ELT pipelines using Apache Airflow for workflow orchestration. • Implement dbt (Data Build Tool) transformations to maintain version-controlled data models in BigQuery, ensuring consistency and reliability across the data pipeline. • Use Google Dataflow (based on Apache Beam) and Pub/Sub for large-scale streaming/batch data processing and ingestion. • Automate job scheduling and data transformations to deliver timely insights for analytics, machine learning, and reporting. • Implement event-driven or asynchronous data workflows between microservices. • Employ Docker and Kubernetes (K8s) for containerization and orchestration, enabling flexible and efficient microservices-based data workflows. • Implement CI/CD pipelines for streamlined development, testing, and deployment of data engineering components. • Enforce data quality standards using Great Expectations or similar frameworks, defining and validating expectations for critical datasets. • Define and uphold metadata management, data lineage, and auditing standards to ensure trustworthy datasets. • Implement security best practices, including encryption at rest and in transit, Identity and Access Management (IAM), and compliance with GDPR or CCPA where applicable. • Integrate with Looker (or similar BI tools) to provide data consumers with intuitive dashboards and real-time insights. • Collaborate with Data Science, Analytics, and Product teams to ensure the data infrastructure supports advanced analytics, including machine learning initiatives. • Maintain Data Mart (Data Mach) environments that cater to specific business domains, optimizing access and performance for key stakeholders.

🎯 Requirements

• 3+ years of professional experience in data engineering, with at least 1 year in mobile data. • Proven track record building and maintaining BigQuery environments and Google Cloud Storagebased data lakes. • Deep knowledge of Apache Airflow for scheduling/orchestration and ETL/ELT design. • Experience implementing dbt for data transformations, RabbitMQ for event-driven workflows, and Pub/Sub + Dataflow for streaming/batch data pipelines. • Familiarity with designing and implementing Data Mart (Data Mach) solutions, as well as using Terraform for IaC. • Strong coding capabilities in Python, Java, or Scala, plus scripting for automation. • Experience with Docker and Kubernetes (K8s) for containerizing data-related services. • Hands-on with CI/CD pipelines and DevOps tools (e.g., Terraform, Ansible, Jenkins, GitLab CI) to manage infrastructure and deployments. • Proficiency in Great Expectations (or similar) to define and enforce data quality standards. • Expertise in designing systems for data lineage, metadata management, and compliance (GDPR, CCPA). • Strong understanding of OLTP (Online Transaction Processing) and OLAP (Online Analytical Processing) systems. • Excellent communication skills for both technical and non-technical audiences. • High level of organization, self-motivation, and problem-solving aptitude. • Will be a plus • Machine Learning (ML) Integration: Familiarity with end-to-end ML workflows and model deployment on GCP (e.g., Vertex AI). • Advanced Observability: Experience with Prometheus, Grafana, Datadog, or New Relic for system health and performance monitoring. • Security & Compliance: Advanced knowledge of compliance frameworks such as HIPAA, SOC 2, or relevant regulations. • Real-Time Data Architectures: Additional proficiency in Kafka, Spark Streaming, or other streaming solutions. • Certifications: GCP-specific certifications (e.g., Google Professional Data Engineer) are highly desirable.

🏖️ Benefits

• Growth and career development • At Leadtech, we prioritize your growth. Enjoy a flexible career path with personalized internal training and an annual budget for external learning opportunities. • Work-Life balance • Benefit from a flexible schedule with flextime (7 - 9:30 a.m. start, 3:30 - 6 p.m. end) and the option of working full remote or from our Barcelona office. Enjoy free Friday afternoons with a 7-hour workday, plus a 35-hour workweek in July and August so you can savor summer! • Comprehensive benefits • Competitive salary, full-time permanent contract, and top-tier private health insurance (including dental and psychological services). • 25 days of vacation plus your birthday off, with flexible vacation options—no blackout days! • Unique Perks • If you wish to come, in our office in Barcelona you’ll find it coplete with free coffee, fresh fruit, snacks, a game room, and a rooftop terrace with stunning Mediterranean views. • Additional benefits include ticket restaurant and nursery vouchers, paid directly from your gross salary. • Join us in an environment where you’re free to innovate, learn, and grow alongside passionate professionals. At Leadtech, you’ll tackle exciting challenges and be part of a vibrant team dedicated to delivering exceptional user experiences

Apply Now

Similar Jobs

July 11

Join Volkswagen as a Mid Data Engineer in Barcelona to enhance enterprise data quality, integrity, and accessibility.

Airflow

ETL

MySQL

Oracle

Postgres

Python

SQL

Vault

June 23

Join Plain Concepts as a Data Engineer to develop innovative data solutions remotely.

AWS

Azure

Cloud

NoSQL

Python

Scala

Spark

SQL

June 17

Volkswagen Group seeks a Senior Data Engineer for robust data solutions in a remote capacity.

Airflow

Amazon Redshift

BigQuery

Docker

ETL

Grafana

Kubernetes

MySQL

NoSQL

OpenShift

Oracle

Postgres

Prometheus

Python

Redis

Spark

SQL

June 10

Join Volkswagen Digital:Hub as a Data Engineer, focusing on data pipelines and architectures.

Airflow

Amazon Redshift

BigQuery

Docker

ETL

Grafana

Kubernetes

MySQL

NoSQL

OpenShift

Oracle

Postgres

Prometheus

Python

Redis

Spark

SQL

May 12

Drive data engineering initiatives as a Data Engineering Lead at seQura. Foster team growth and ensure data reliability in a fintech environment.

Airflow

Amazon Redshift

AWS

Azure

BigQuery

Cloud

ETL

Google Cloud Platform

Kafka

Python

SQL

Terraform

Built by Lior Neu-ner. I'd love to hear your feedback — Get in touch via DM or support@remoterocketship.com