Maze (mazehq.com) Logo

Maze (mazehq.com)

Backend Engineer (Data Engineering)

Posted 5 Days Ago
Remote
28 Locations
Senior level
Remote
28 Locations
Senior level
As a Backend Engineer (Data Engineering), design and build scalable data pipelines and infrastructure for processing security data, ensuring reliability and performance.
The summary above was generated by AI

Summary of the Role:

As a Backend Engineer (Data Engineering) at Maze, you'll be the technical architect behind our data infrastructure, building production-grade data pipelines that process massive volumes of security data at scale. This is a unique opportunity to join as one of the early engineering team members of a well-funded startup building at the intersection of generative AI and cybersecurity. You'll design and implement the data lake architecture, streaming pipelines, and transformation systems that power our AI agents' ability to analyze and protect customer environments.

You'll take full ownership of the entire data lifecycle—from ingesting millions of records in minutes to building self-managed, distributed data systems on AWS. Your success will be measured by pipeline reliability, data processing performance, and your ability to scale our infrastructure to handle exponentially growing data volumes as we bring on enterprise customers. This role is perfect for a software engineer who specializes in building data systems at scale, thinks in terms of platform architecture, and thrives on solving the complex technical challenges that come with processing security data for the world's largest organizations.

Your Contributions to Our Journey:

  • Build Production Data Pipelines: Design, implement, and maintain scalable data pipelines that ingest gigabytes to terabytes of security data daily, processing millions of records in single-digit minutes while maintaining reliability and data quality

  • Architect Distributed Data Systems: Build and evolve our S3-based data lake infrastructure using Apache Iceberg, creating self-managed, distributed systems that enable rapid data transformations and efficient storage at massive scale

  • Own the Complete Data Lifecycle: Take end-to-end ownership from data ingestion through Kafka streams to transformation via Spark/EMR, ensuring seamless data flow from customer environments to our AI-powered analysis platform

  • Enable Platform Scalability: Build data infrastructure with platform thinking, creating systems that support current product needs while laying the foundation for future products and exponential data growth

  • Optimize for Enterprise Scale: Continuously improve data processing performance and cost efficiency as we scale from current volumes to supporting the world's largest enterprise security environments

  • Drive Technical Excellence: Establish data engineering best practices, participate in code reviews as a software engineer, and mentor team members on building robust, maintainable data systems

  • Collaborate Cross-Functionally: Work closely with infrastructure engineers, backend engineers, and product teams to ensure data systems seamlessly integrate with our AI agents and security analysis capabilities

What You Need to Be Successful:

  • Software Engineering Foundation: 7+ years of software engineering experience with at least 4+ years focused specifically on data engineering—you must be a strong software engineer who will pass our coding challenges, not just someone who transforms existing data

  • Production Data Pipeline Mastery: Proven track record building and scaling data ingestion systems that handle gigabytes to terabytes daily, with hands-on experience at companies moving massive data volumes (early Fivetran/Matillion engineers, or engineers at companies feeding Databricks/Snowflake at scale)

  • Core Technology Expertise: Deep, hands-on production experience with Python, Apache Kafka, and Apache Spark—these are your bread and butter technologies that you use daily and know intimately, not just tools you've experimented with

  • AWS Data Infrastructure: Strong expertise with AWS data services including S3, EMR, and building data lakes at scale—you intuitively design systems using AWS and understand how to architect for both performance and cost optimization

  • Data Lake Architecture: Proven experience with Apache Iceberg (critical—this is our core technology), data lakehouse concepts, and building distributed data systems that process massive datasets efficiently

  • Attention to Detail: Exceptional care and precision in data ingestion and transformation work—you understand that in security data, accuracy and reliability are non-negotiable

  • Scale Experience: Direct experience working at companies that deal with serious data scale—environments where Redshift and BigQuery aren't sufficient and you've had to architect custom solutions for petabyte-scale challenges

  • Hands-On Builder: Currently active as a developer, writing production code regularly—you're not someone who just designs systems or reviews others' work, you build and own them yourself

  • Nice to haves:

    • Experience with Temporal workflow orchestration (very important for our architecture)

    • Knowledge of Apache Hoodie, Parquet, or ORC file formats for optimized data storage

    • Background with RDS, PostgreSQL optimization, or other database performance tuning

    • Previous experience at technical security product companies or handling security-related data

    • Track record of building self-service data platforms that enable other teams to operate independently

Why Join Us:

  • Ambitious Data Challenges: We're leveraging advanced data processing at the intersection of generative AI and cybersecurity, building systems that process security data at massive scale to enable proactive threat detection. You'll architect the data infrastructure that powers breakthrough AI capabilities for security teams worldwide.

  • Expert Team: We are a team of hands-on leaders with deep experience in Big Tech and Scale-ups. Our team has been part of the leadership teams behind multiple acquisitions and an IPO.

  • Impactful Work: Cybersecurity is a force for good—helping stop cyber attacks ultimately helps deliver better outcomes for all of us. The data systems you build will directly enable security teams to protect organizations from real threats.

  • Build an AI-Native Company: We're building a new company in the AI era with the opportunity to design everything from the ground up—you'll architect the data foundation using cutting-edge technologies like Apache Iceberg and build systems with platform thinking from day one.

  • Technical Growth: Direct partnership with experienced infrastructure and engineering teams, significant equity upside, and the opportunity to own the data engineering function as we scale from startup to handling the world's largest security datasets.

Top Skills

Apache Iceberg
Apache Kafka
Spark
AWS
Emr
Python
S3

Similar Jobs

3 Hours Ago
Remote
30 Locations
Mid level
Mid level
Artificial Intelligence • Productivity • Software • Automation
As a Data Engineer at Zapier, you'll build scalable data systems, enhance product functionality through data, and collaborate with teams to improve data access and usability.
Top Skills: AWSAzureDatabricksGCPPythonSparkSQLTypescript
3 Hours Ago
Easy Apply
Remote
30 Locations
Easy Apply
Senior level
Senior level
Cloud • Security • Software • Cybersecurity • Automation
Lead the evolution of GitLab's Security Risk Management platform, architecting scalable systems for vulnerability analysis and remediation while transforming the engineering culture toward high-performance practices.
Top Skills: AIApache FlinkApache KafkaCloud SpannerDevsecopsGitlabKubernetesPostgres
Yesterday
Easy Apply
Remote
31 Locations
Easy Apply
Mid level
Mid level
Cloud • Security • Software • Cybersecurity • Automation
As a Pricing Analyst, you will analyze pricing and monetization performance using data analysis and build dashboards. You will partner with various teams to inform pricing strategies and improve monetization of AI capabilities.
Top Skills: PythonRSQLTableau

What you need to know about the Edinburgh Tech Scene

From traditional pubs and centuries-old universities to sleek shopping malls and glass-paneled office buildings, Edinburgh's architecture reflects its unique blend of history and modernity. But the fusion of past and future isn't just visible in its buildings; it's also shaping the city's economy. Named the United Kingdom's leading technology ecosystem outside of London, Edinburgh plays host to major global companies like Apple and Adobe, as well as a growing number of innovative startups in fields like cybersecurity, finance and healthcare.

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account