Data Engineer – Mid-level

Job not on LinkedIn

November 27

🗣️🇧🇷🇵🇹 Portuguese Required

Apply Now
Logo of Join Creative Tech

Join Creative Tech

Software • SaaS • Compliance

Join Creative Tech is a company specializing in the conception and development of software solutions, focusing on quality and outsourcing services. They create user-centered solutions that offer innovative value while ensuring a strong return on investment through agile methodologies and rigorous testing processes. Additionally, Join Creative Tech provides qualified IT professionals to enhance service quality and productivity for their clients.

51 - 200 employees

Founded 2010

☁️ SaaS

📋 Compliance

📋 Description

• Build, test, and maintain ingestion and transformation pipelines (ETL/ELT); • Use Databricks with a focus on PySpark and SQL; • Ensure scalability and robustness for the BI team; • Optimize pipeline execution; • Perform query tuning; • Tune clusters and Delta tables to maximize performance and reduce costs; • Apply security and quality policies to pipelines; • Implement governance across Bronze → Silver → Gold layers; • Ensure quality and traceability of transformations; • Create connections and ingest data from diverse systems; • Ensure integrations are reliable and performant; • Implement automated data quality rules; • Develop gates in Databricks / Delta Live Tables (DLT); • Ensure only qualified data reaches the consumption layer; • Model structures in the Lakehouse; • Document and maintain metadata repositories and models; • Support administration of the Databricks workspace; • Manage access and enforce information security best practices.

🎯 Requirements

• Bachelor's degree in IT or any field with a postgraduate degree in IT; • Minimum of 4 years' experience in Software Development, Data Engineering and/or Data Administration; • Hands-on experience with Databricks (clusters, notebooks, Delta Lake, DLT); • Knowledge of Big Data and distributed processing; • Strong knowledge of PySpark; • Proficiency in SQL; • Experience with version control tools (Git); • Logical and physical modeling in a Lakehouse environment; • Documentation and maintenance of metadata; • Experience ingesting data from RDBMS, APIs, and Cloud Storage; • Application of security rules and governance policies; • Implementation of Data Quality Gates; • Access management in the Databricks Workspace.

🏖️ Benefits

• Infrastructure allowance; • Flexible working hours; • Birthday day off — with a surprise! • Support for training and certifications; • Access to Alura; • Partnership with FIAP; • Referral bonus; • Health insurance; • Dental plan; • Vittude — mental health platform; • Wellhub — for physical health; • New Value — discount vouchers; • SESC; • Meal and/or food allowance.

Apply Now

Similar Jobs

November 27

Data Engineer responsible for building and optimizing data pipelines to support strategic decision-making at Stone Co., a leading fintech in Brazil.

🗣️🇧🇷🇵🇹 Portuguese Required

Airflow

Cloud

ETL

Google Cloud Platform

Python

SQL

November 26

Keyrus

1001 - 5000

🤝 B2B

Senior Data Engineer developing and maintaining efficient data pipelines using AWS technologies at Keyrus. Collaborating with internal teams to ensure data integrity and solution scalability.

🗣️🇧🇷🇵🇹 Portuguese Required

Amazon Redshift

AWS

ETL

PySpark

Python

RDBMS

SQL

November 26

Senior Data Engineer at Valtech, designing and maintaining ETL/ELT pipelines for scalable data processing and supporting AI initiatives.

Airflow

Apache

Azure

BigQuery

Cassandra

Cloud

ETL

Google Cloud Platform

Java

Kafka

MongoDB

MySQL

NoSQL

Postgres

Python

Scala

Spark

SQL

November 25

Data Engineer handling data transformation and governance in AWS for a construction tech platform. Collaborating with teams to enhance data architecture and model structures.

🗣️🇧🇷🇵🇹 Portuguese Required

Airflow

AWS

PySpark

Python

Spark

SQL

November 22

Data Engineer developing and maintaining ETL pipelines and data integration services for innovative AdTech platforms. Utilizing AWS, Python, and SQL to ensure data quality and system efficiency.

AWS

DynamoDB

ETL

MongoDB

NoSQL

PySpark

Python

SQL

Built by Lior Neu-ner. I'd love to hear your feedback — Get in touch via DM or support@remoterocketship.com