Gen AI Data Engineer II

November 25

Apply Now
Logo of Dynatron Software, Inc.

Dynatron Software, Inc.

Automotive • SaaS • Analytics

Dynatron Software, Inc. is a company that combines advanced analytics software with expert coaching to help automotive service departments maximize their revenue opportunities. Their product suite includes various solutions like PriceSmart for optimizing labor and part pricing, FileSmart for enhancing warranty labor rates and parts markup, SellSmart for increasing service sales to existing customers, and MarketSmart for boosting service department traffic through strategic marketing campaigns. With a focus on increasing profitability, Dynatron collaborates closely with dealerships to identify and capitalize on hidden revenue streams using their proprietary analytics and extensive repair order database.

51 - 200 employees

Founded 1999

☁️ SaaS

📋 Description

• Engineer Generative AI Data Systems • Design and maintain data pipelines for training, fine-tuning, and retrieval-augmented generation (RAG) use cases. • Build ingestion frameworks using AWS Glue, Lambda, Kinesis, and Step Functions to support large-scale AI workloads. • Develop embedding pipelines, feature stores, and vector database integrations (Pinecone, FAISS, Chroma, Amazon OpenSearch) to power semantic retrieval. • Transform unstructured data–documents, text, images, logs–into AI-ready assets for LLM applications. • Integrate & Orchestrate LLM Architectures • Build end-to-end GenAI pipelines connecting enterprise data with LLMs including Anthropic Claude, Amazon Titan, OpenAI GPT, and Llama 3. • Use LangChain, LlamaIndex, and Bedrock Agents to deliver context-rich RAG, prompt-chaining, and conversational intelligence. • Develop LLM-powered APIs enabling natural language querying, summarization, search, and generative workflows. • Optimize prompts, context windows, model evaluation, and response quality. • Scale AI Infrastructure & MLOps • Deploy, monitor, and optimize LLM workflows on AWS Bedrock and other cloud AI platforms. • Implement CI/CD pipelines for GenAI systems using Airflow, Prefect, GitHub Actions, or AWS CodePipeline. • Establish data and model observability frameworks to track drift, accuracy, latency, and performance. • Partner with Data Science and MLOps teams to streamline fine-tuning, deployment, and scalable model operations. • Champion Governance, Security & Responsible AI • Implement data lineage, access controls, encryption, and governance for AI datasets. • Enforce Responsible AI practices, ensuring transparency, risk mitigation, and ethical use of LLMs. • Maintain prompt logs, telemetry, and audit documentation supporting SOC2, GDPR, and CCPA compliance.

🎯 Requirements

• Bachelor’s or Master’s degree in Computer Science, Data Engineering, or related field. • 5+ years of data engineering experience, including 2+ years developing GenAI or LLM-based solutions. • Strong proficiency in: AWS Bedrock, SageMaker, or Vertex AI • LangChain or LlamaIndex • Snowflake, Redshift, or Databricks • Python, SQL, and API integrations • Vector databases (Pinecone, FAISS, Chroma, OpenSearch) • Proven experience building RAG pipelines, embeddings, and prompt-chaining architectures. • Deep understanding of data modeling, orchestration, and MLOps best practices. • Ability to integrate LLM capabilities into enterprise SaaS products and data platforms.

🏖️ Benefits

• Comprehensive health, vision, and dental insurance • Employer-paid short- and long-term disability and life insurance • 401(K) with competitive company match • Flexible vacation policy and 9 paid holidays • Remote-first culture

Apply Now

Similar Jobs

November 25

Data Engineer transforming healthcare data into actionable insights at Atropos Health. Collaborating with interdisciplinary teams to build and maintain data pipelines and ensure data quality.

Amazon Redshift

BigQuery

Cloud

Python

SQL

November 24

Allata

201 - 500

🤝 B2B

Data Engineer contributing to enterprise data platform projects focusing on data pipelines and logic engines. Collaborating with teams to ensure effective data management and integration.

AWS

Azure

Cloud

ETL

Jenkins

MS SQL Server

Oracle

PySpark

Spark

SQL

November 24

Senior Data Engineer at CrowdStrike designing and building scalable data integration components and automated data pipelines. Collaborating with DevOps and data scientists to enhance the data platform.

Airflow

Amazon Redshift

Apache

Cloud

ETL

Jenkins

Python

SFDC

SQL

November 24

Senior Data Engineer with strong AI experience at Quisitive building data-driven solutions. Optimizing data pipelines and exploring cutting-edge AI technologies for intelligent applications.

Azure

ETL

Neo4j

Oracle

Postgres

Python

Spark

Vault

November 24

RTX

10,000+ employees

🚀 Aerospace

Senior Data Engineer role at RTX focusing on building scalable data pipelines for aerospace challenges. Collaborating with cross-functional teams to deliver cutting-edge AI and data solutions.

AWS

Azure

Cloud

ETL

Matillion

Python

Spark

Built by Lior Neu-ner. I'd love to hear your feedback — Get in touch via DM or support@remoterocketship.com