Data Engineer, Databricks

November 24

Apply Now
Logo of Allata

Allata

B2B • Consulting • Technology

Allata is a global consulting firm that helps businesses achieve digital excellence by crafting unique customer experiences, identifying revenue-generating opportunities, and improving operational efficiencies. They offer a wide range of services including strategic services, technology and cloud solutions, data and insights, and artificial intelligence development. Allata's approach is experience-led, strategy-aligned, tech-enabled, and data-driven. They work across multiple industries such as automotive, financial services, health and life sciences, high tech, logistics and transportation, among others, to drive modernization, personalization, innovation, and efficiency. With a team of former IT leaders and consulting professionals, Allata guides clients through complex digital landscapes to deliver value-based outcomes and help organizations integrate digital strategies seamlessly into their overall business strategies.

201 - 500 employees

🤝 B2B

📋 Description

• Contribute to transformative enterprise data platform projects focused on developing data pipelines and logic engines to manage ingest, staging, and multi-tier data product modeling. • Data enrichment using various OEM-specific data warehouse and data lake house platform implementations. • Full life cycle design, build, deployment and optimization data products for multiple large enterprise industry vertical-specific implementations by processing datasets through a defined series of logically conformed layers, models, and views.

🎯 Requirements

• Collaborate in defining the overall architecture of the solution. This includes knowledge of modern Enterprise Data Warehouse and Data Lakehouse architectures that implement Medallion or Lamda architectures • Design, develop, test, and deploy processing modules to implement data-driven rules using SQL, Stored Procedures, and Pyspark. • Understand and own data product engineering deliverables relative to a CI-CD pipeline and standard devops practices and principles • Build and optimize data pipelines on platforms like Databricks, SQL Server, or Azure Data Fabric. • Current knowledge of modern data tools like (Databricks,FiveTran, Data Fabric and others); Core experience with data architecture, data integrations, data warehousing, and ETL/ELT processes. • Applied experience with developing and deploying custom whl and or in session notebook scripts for custom execution across parallel executor and worker nodes. • Applied experience in SQL, Stored Procedures, and Pyspark based on area of data platform specialization. • Strong knowledge of cloud and hybrid relational database systems, such as MS SQL Server, PostgresSQL, Oracle, Azure SQL, AWS RDS, Aurora or a comparable engine. • Strong experience with batch and streaming data processing techniques and file compactization strategies. • Automation experience with CICD pipelines to support deployment and integration workflows including trunk-based development using automation services such as Azure DevOps, Jenkins, Octopus. • Advanced proficiency in Pyspark for advanced data processing tasks. • Advance proficiency in spark workflow optimization and orchestration using tools such as Asset Bundles or DAG (Directed Acyclic Graph) orchestration. • Ability to identify, troubleshoot, and resolve complex data issues effectively. • Strong teamwork, communication skills and intellectual curiosity to work collaboratively and effectively with cross-functional teams. • Commitment to delivering high-quality, accurate, and reliable data products solutions. • Willingness to embrace new tools, technologies, and methodologies. • Innovative thinker with a proactive approach to overcoming challenges.

🏖️ Benefits

• At Allata, we value differences. • Allata is an equal opportunity employer. We celebrate diversity and are committed to creating an inclusive environment for all employees.

Apply Now

Similar Jobs

November 24

Senior Data Engineer at CrowdStrike designing and building scalable data integration components and automated data pipelines. Collaborating with DevOps and data scientists to enhance the data platform.

Airflow

Amazon Redshift

Apache

Cloud

ETL

Jenkins

Python

SFDC

SQL

November 24

Senior Data Engineer with strong AI experience at Quisitive building data-driven solutions. Optimizing data pipelines and exploring cutting-edge AI technologies for intelligent applications.

Azure

ETL

Neo4j

Oracle

Postgres

Python

Spark

Vault

November 24

RTX

10,000+ employees

🚀 Aerospace

Senior Data Engineer role at RTX focusing on building scalable data pipelines for aerospace challenges. Collaborating with cross-functional teams to deliver cutting-edge AI and data solutions.

AWS

Azure

Cloud

ETL

Matillion

Python

Spark

November 24

Data Engineer developing data workflows to manage and optimize data quality for Cherokee Nation Integrated Health. Key contributor in cross-functional teams handling various data processing tasks.

Amazon Redshift

AWS

DynamoDB

EC2

ETL

NoSQL

Numpy

Pandas

Postgres

Python

SQL

TypeScript

November 24

Data Architect at 3Cloud supporting Azure Data Platform solutions and mentoring junior teams. Lead client support and design data-driven architectures for various industries.

Azure

SQL

Built by Lior Neu-ner. I'd love to hear your feedback — Get in touch via DM or support@remoterocketship.com