Principal Data Architect

November 7

Apply Now
Logo of Egen

Egen

Artificial Intelligence • Healthcare • Public Sector

Egen is a company that specializes in engineering innovative solutions using platforms, data, and generative AI. They focus on harnessing the potential of data to empower organizations and individuals, providing services across various sectors including communications, healthcare, and public sector solutions. Their customizable platforms address critical challenges in urban management, customer engagement, clinical care, and operational excellence, aiming to drive impactful outcomes for clients.

501 - 1000 employees

Founded 2000

🤖 Artificial Intelligence

📋 Description

• Lead the end-to-end architecture, design, and implementation of scalable Data Lakehouse solutions on Google Cloud Platform (GCP) using BigQuery, GCS, BigLake, and Dataplex • Collaborate directly with customers to understand business goals, data challenges, and technical requirements; translate them into robust architectural blueprints and actionable plans • Design and implement data pipelines supporting both real-time and batch ingestion using modern orchestration and streaming frameworks • Establish and enforce best practices for data cataloging, metadata management, lineage, and data quality across multiple systems • Define and implement data security, access control, and governance models in compliance with enterprise and regulatory standards • Serve as the technical lead for project teams - mentoring engineers, reviewing solutions, and ensuring architectural consistency across deliverables • Balance strategic architecture discussions with hands-on solutioning, POCs, and deep dives into data pipelines or performance tuning • Partner with stakeholders, cloud architects, and delivery leads to drive solution adoption, scalability, and long-term maintainability • Represent the company as a trusted technical advisor in client engagements - clearly articulating trade-offs, best practices, and recommendations

🎯 Requirements

• 8–10 years of progressive experience in Software Engineering and Data Platform development, with 5+ years architecting data platforms on GCP and/or Databricks • Proven hands-on experience designing and deploying Data Lakehouse platforms with data products and medallion architectures • Strong understanding of data ingestion patterns (real-time and batch), ETL/ELT pipeline design, and data orchestration using tools such as Airflow, Pub/Sub, or similar frameworks • Expertise in data modeling, storage optimization, partitioning, and performance tuning for large-scale analytical workloads • Experience implementing data governance, security, and cataloging solutions (Dataplex, Data Catalog, IAM, or equivalent) • Excellent communication and presentation skills - able to confidently engage with technical and non-technical stakeholders and guide clients through solution decisions • Demonstrated ability to lead by example in mixed teams of engineers, analysts, and architects, balancing architectural vision with hands-on delivery • Nice to have: Experience with Databricks (Delta Lake, Unity Catalog) and hybrid GCP-Databricks data architectures • Strong problem-solving mindset, curiosity to explore new technologies, and ability to “zoom out” for architecture discussions and “zoom in” for code-level troubleshooting

🏖️ Benefits

• Comprehensive Health Insurance • Paid Leave (Vacation/PTO) • Paid Holidays • Sick Leave • Parental Leave • Bereavement Leave • 401 (k) Employer Match • Employee Referral Bonuses

Apply Now

Similar Jobs

November 6

Thumbtack

1001 - 5000

👥 B2C

🏪 Marketplace

Data Engineer developing and maintaining marketing datasets and data products at Thumbtack. Collaborating with product teams for data integration and quality assurance.

October 28

Brinqa

51 - 200

🔒 Cybersecurity

Principal Software Engineer responsible for building scalable data pipelines for Brinqa's Cybersecurity Knowledge Platform. Collaborating with teams to support analytics and operational needs.

🇨🇦 Canada – Remote

💰 Private Equity Round on 2021-06

⏰ Full Time

🔴 Lead

🚰 Data Engineer

October 16

NerdWallet

501 - 1000

Staff Data Engineer developing and maintaining data systems at NerdWallet. Collaborating on cross-functional data challenges to support financial decision making in a remote environment.

🇨🇦 Canada – Remote

💵 CA$162k - CA$272k / year

💰 Secondary Market on 2021-05

⏰ Full Time

🔴 Lead

🚰 Data Engineer

September 24

Docker, Inc

51 - 200

Lead data platform architecture and build petabyte-scale pipelines with Snowflake, DBT, and Airflow to enable Docker analytics.

🇨🇦 Canada – Remote

💵 $200.4k - $275.6k / year

💰 $105M Series C on 2022-03

⏰ Full Time

🔴 Lead

🚰 Data Engineer

September 13

Docker, Inc

51 - 200

Principal Data Engineer leading Docker's data platform architecture. Building scalable data infrastructure and governance to support millions of developers.

🇨🇦 Canada – Remote

💵 $208k - $286k / year

💰 $105M Series C on 2022-03

⏰ Full Time

🔴 Lead

🚰 Data Engineer

Developed by Lior Neu-ner. I'd love to hear your feedback — Get in touch via DM or support@remoterocketship.com