
HR Tech ⢠Recruitment ⢠SaaS
Remotebase is a platform that connects companies with elite, vetted software developers from around the globe. The company provides a diverse pool of developers with skills in areas like Python, JavaScript, AI, computer vision, and more, thoroughly vetted for their expertise and reliability. They streamline the hiring process by offering a curated selection of candidates, allowing companies to save time and focus on project-critical needs. Remotebase makes it easy for businesses to find and hire top talent for their tech projects by offering a simple three-step process: sharing needs, matching with developers, and interviewing to select the best fit.
51 - 200 employees
Founded 2020
đĽ HR Tech
đŻ Recruiter
âď¸ SaaS
October 1

HR Tech ⢠Recruitment ⢠SaaS
Remotebase is a platform that connects companies with elite, vetted software developers from around the globe. The company provides a diverse pool of developers with skills in areas like Python, JavaScript, AI, computer vision, and more, thoroughly vetted for their expertise and reliability. They streamline the hiring process by offering a curated selection of candidates, allowing companies to save time and focus on project-critical needs. Remotebase makes it easy for businesses to find and hire top talent for their tech projects by offering a simple three-step process: sharing needs, matching with developers, and interviewing to select the best fit.
51 - 200 employees
Founded 2020
đĽ HR Tech
đŻ Recruiter
âď¸ SaaS
⢠Design, build, and maintain scalable and resilient CI/CD pipelines for data applications and infrastructure, with a focus on Snowflake, dbt, and related data tools. ⢠Implement and manage Snowflake dbt projects for data transformation, including developing dbt models, tests, and documentation, and integrating dbt into CI/CD workflows. ⢠Develop and manage infrastructure as code (IaC) using Terraform to provision and configure cloud resources for data storage, processing, and analytics on GCP. ⢠Automate the deployment, monitoring, and management of Snowflake data warehouse environments, ensuring optimal performance, security, and cost-effectiveness. ⢠Collaborate with data engineers and data scientists to understand their requirements and provide robust, automated solutions for data ingestion, processing, and delivery. ⢠Implement and manage monitoring, logging, and alerting systems for data pipelines and infrastructure to ensure high availability and proactive issue resolution. ⢠Develop and maintain robust automation scripts and tools, primarily using Python, to streamline operational tasks, manage data pipelines, and improve efficiency; Bash scripting for system-level tasks is also required. ⢠Ensure security best practices are implemented and maintained across the data infrastructure and pipelines. ⢠Troubleshoot and resolve issues related to data infrastructure, pipelines, and deployments in a timely manner. ⢠Participate in code reviews for infrastructure code, dbt models, and automation scripts. ⢠Document system architectures, configurations, and operational procedures. ⢠Stay current with emerging DevOps technologies, data engineering tools, and cloud best practices, particularly related to Snowflake, dbt, and Terraform. ⢠Optimize data pipelines for performance, scalability, and cost. ⢠Support and contribute to data governance and data quality initiatives from an operational perspective. ⢠Help implement AI features
⢠Bachelor's degree in Computer Science, Engineering, or a related technical field expertise which is equivalent. ⢠5+ years of hands-on experience in a DevOps, SRE, or infrastructure engineering role. ⢠3+ years of experience specifically focused on automating and managing data infrastructure and pipelines. ⢠1+ years of experience enabling AI features ⢠**Others:** ⢠Strong, demonstrable experience with Infrastructure as Code tools, particularly Terraform. ⢠Strong background in DevOps principles and practices, and hands-on experience in building business intelligence solutions. ⢠Highly experienced in automation, and problem-solving skills with proficiency in cloud technologies. ⢠Ability to collaborate effectively with data engineers, analysts, and other stakeholders to ensure the reliability and performance of our data ecosystem. ⢠Proven experience with dbt for data transformation, including developing models, tests, and managing dbt projects in a production environment. ⢠Hands-on experience managing and optimizing Snowflake data warehouse environments. ⢠Demonstrable experience with data modeling techniques for ODS, dimensional modeling (Facts, Dimensions), and semantic models for analytics and BI. ⢠Strong proficiency in Python for automation, scripting, and data-related tasks. Experience with relevant Python libraries is a plus. Strong Bash scripting. ⢠Solid understanding of CI/CD principles and tools (e.g., Bitbucket Runners, Jenkins, GitLab CI, GitHub Actions, Azure DevOps). ⢠Experience with cloud platforms (GCP preferred, AWS, or Azure) and their data services. ⢠Experience with containerization technologies (e.g., Docker, Kubernetes) is a plus. ⢠Knowledge of data integration tools and ETL/ELT concepts. ⢠Familiarity with monitoring and logging tools. ⢠Strong SQL skills. ⢠Ability to work independently and as part of a collaborative team in an agile environment. ⢠Strong communication skills, with the ability to explain complex technical concepts clearly.
⢠Fully remote with office optional. You decide when you would like to work from home and when from the office. ⢠Flexible timings. You decide your work scheduled. ⢠Market competitive compensation (in $$). ⢠Insane learning and growth
Apply NowSeptember 30
Design and maintain GCP/AWS data pipelines for Leega Consultoria. Build data models, ETL/ELT processes, and ensure governance for analytics.
đŁď¸đ§đˇđľđš Portuguese Required
September 30
Senior Data Engineer leading SAS-to-GCP migrations for Leega, a data analytics consultancy. Designing BigQuery models, building ETL pipelines, and optimizing cloud infrastructure.
đŁď¸đ§đˇđľđš Portuguese Required
September 30
Build and maintain data pipelines and cloud platforms for Dadosfera's Data & AI platform; lead client projects and design scalable data architectures.
đ§đˇ Brazil â Remote
đ° $1.8M Seed Round on 2022-06
â° Full Time
đĄ Mid-level
đ Senior
đ° Data Engineer
September 28
Build data platform, APIs and MLOps for Experian Brazil; integrate batch and real-time pipelines and collaborate with business and tech teams.
đŁď¸đ§đˇđľđš Portuguese Required
September 25