
Software • SaaS • Compliance
Join Creative Tech is a company specializing in the conception and development of software solutions, focusing on quality and outsourcing services. They create user-centered solutions that offer innovative value while ensuring a strong return on investment through agile methodologies and rigorous testing processes. Additionally, Join Creative Tech provides qualified IT professionals to enhance service quality and productivity for their clients.
November 27
🗣️🇧🇷🇵🇹 Portuguese Required

Software • SaaS • Compliance
Join Creative Tech is a company specializing in the conception and development of software solutions, focusing on quality and outsourcing services. They create user-centered solutions that offer innovative value while ensuring a strong return on investment through agile methodologies and rigorous testing processes. Additionally, Join Creative Tech provides qualified IT professionals to enhance service quality and productivity for their clients.
• Build, test, and maintain ingestion and transformation pipelines (ETL/ELT); • Use Databricks with a focus on PySpark and SQL; • Ensure scalability and robustness for the BI team; • Optimize pipeline execution; • Perform query tuning; • Tune clusters and Delta tables to maximize performance and reduce costs; • Apply security and quality policies to pipelines; • Implement governance across Bronze → Silver → Gold layers; • Ensure quality and traceability of transformations; • Create connections and ingest data from diverse systems; • Ensure integrations are reliable and performant; • Implement automated data quality rules; • Develop gates in Databricks / Delta Live Tables (DLT); • Ensure only qualified data reaches the consumption layer; • Model structures in the Lakehouse; • Document and maintain metadata repositories and models; • Support administration of the Databricks workspace; • Manage access and enforce information security best practices.
• Bachelor's degree in IT or any field with a postgraduate degree in IT; • Minimum of 4 years' experience in Software Development, Data Engineering and/or Data Administration; • Hands-on experience with Databricks (clusters, notebooks, Delta Lake, DLT); • Knowledge of Big Data and distributed processing; • Strong knowledge of PySpark; • Proficiency in SQL; • Experience with version control tools (Git); • Logical and physical modeling in a Lakehouse environment; • Documentation and maintenance of metadata; • Experience ingesting data from RDBMS, APIs, and Cloud Storage; • Application of security rules and governance policies; • Implementation of Data Quality Gates; • Access management in the Databricks Workspace.
• Infrastructure allowance; • Flexible working hours; • Birthday day off — with a surprise! • Support for training and certifications; • Access to Alura; • Partnership with FIAP; • Referral bonus; • Health insurance; • Dental plan; • Vittude — mental health platform; • Wellhub — for physical health; • New Value — discount vouchers; • SESC; • Meal and/or food allowance.
Apply NowNovember 27
2 - 10
Data Engineer responsible for building and optimizing data pipelines to support strategic decision-making at Stone Co., a leading fintech in Brazil.
🗣️🇧🇷🇵🇹 Portuguese Required
Airflow
Cloud
ETL
Google Cloud Platform
Python
SQL
November 26
Senior Data Engineer developing and maintaining efficient data pipelines using AWS technologies at Keyrus. Collaborating with internal teams to ensure data integrity and solution scalability.
🗣️🇧🇷🇵🇹 Portuguese Required
Amazon Redshift
AWS
ETL
PySpark
Python
RDBMS
SQL
November 26
Senior Data Engineer at Valtech, designing and maintaining ETL/ELT pipelines for scalable data processing and supporting AI initiatives.
Airflow
Apache
Azure
BigQuery
Cassandra
Cloud
ETL
Google Cloud Platform
Java
Kafka
MongoDB
MySQL
NoSQL
Postgres
Python
Scala
Spark
SQL
November 25
Data Engineer handling data transformation and governance in AWS for a construction tech platform. Collaborating with teams to enhance data architecture and model structures.
🗣️🇧🇷🇵🇹 Portuguese Required
Airflow
AWS
PySpark
Python
Spark
SQL
November 22
Data Engineer developing and maintaining ETL pipelines and data integration services for innovative AdTech platforms. Utilizing AWS, Python, and SQL to ensure data quality and system efficiency.
AWS
DynamoDB
ETL
MongoDB
NoSQL
PySpark
Python
SQL