Data Engineer (Architect) – Remote – REMOTE

Braintrust
  • Post Date: February 12, 2025
  • 45386
  • Applications 0
  • Views 2
Job Overview

logoJob Description

What You’ll Do

Snowflake & AWS: Own and optimize our Snowflake data warehouse, ensure efficient scaling on AWS, and maintain performant, reliable data stores.Build Core Data Infrastructure (0 → 1): Design and implement Neon Blue’s foundational data architecture and pipelines using tools like Airbyte/FiveTran for external data ingestion (Shopify, Klaviyo, Amplitude, etc.).Pipeline Orchestration: Develop and manage DAG-based workflows for batch and near-real-time data processing in Airflow or Dagster.Data Transformation & Modeling: Use dbt and Python to build robust, maintainable transformations that power reporting, analytics, and ML workflows.Machine Learning Platform Integration: Collaborate with Data Science teams to integrate data into cloud ML platforms (Google Vertex AI or Azure ML) for advanced analytics and modeling.Observability & Security: Implement end-to-end monitoring, alerting, and logging strategies to ensure data quality, reliability, and compliance. Champion security best practices and SOC-2 standards.Compliance & Controls: Work alongside legal and DevOps teams to meet data governance requirements, enforce data access controls, and support compliance audits.Scalability & Performance: Architect solutions for rapid, large-scale data ingestion and transformation. Design automated frameworks to handle real-time event streams.Culture & Collaboration: Be part of a small, nimble, and passionate engineering team—contributing to a collaborative, ownership-driven environment.

What We’re Looking For

Experience in 0-1 Data Systems: Proven track record of building data infrastructure from scratch at a startup or similar environment.Advanced Cloud Warehouse Skills: Expert-level Snowflake knowledge (data ingestion, optimization, security, RBAC controls, etc.).Python, dbt, Airflow/Dagster: Strong proficiency in scripting and orchestration tools for both batch and near-real-time pipelines.Cloud Services (AWS/GCP/Azure): Hands-on experience setting up and optimizing data pipelines and ML workflows on major cloud platforms.Data Observability & Security: Familiarity with logging, monitoring, alerting, and end-to-end data governance, including SOC-2 compliance.Performance & Scaling: Deep understanding of how to handle large, rapid data ingestion while ensuring reliability and performance.Strong Communication & Collaboration: Ability to work closely with cross-functional teams (Data Science, Product, Ops) to deliver data-driven insights.

Why Neon Blue?

Impact & Autonomy: Be a key architect who shapes the entire data landscape in a fast-growing startup.Experienced Founders: Work alongside a leadership team that has successfully built and scaled a $100M/year e-commerce venture.Cutting-Edge Tech: Get hands-on with modern tooling and cloud platforms for machine learning and real-time analytics.Competitive Compensation & Benefits: We offer a generous salary, equity, healthcare, and a flexible time-off policy.

If you’re excited about defining and owning the full data stack in a high-growth environment, we’d love to hear from you. Join us at Neon Blue and help us build a data platform that powers the next generation of innovative products!

Job Detail
Shortlist Never pay anyone for job application test or interview.