hero

Job Opportunities

Applications Engineer (GPU-Accelerated)

Alembic

Alembic

IT
San Francisco, CA, USA
USD 235k-258k / year + Equity
Posted on Jul 17, 2025

Location

San Francisco HQ

Employment Type

Full time

Location Type

On-site

Department

R&D

Compensation

  • IC4Estimated salary commensurate with experience. $235K • Offers Equity
  • IC5Estimated salary commensurate with experience. $258K • Offers Equity
Tier Guide

Our Compensation Philosophy:

  • Market-based: Our formula ensures new hires earn at or above current 75th percentile cash compensation benchmarks.

  • Ownership: Our generous equity program ensures new hires are owners, not just employees.

  • Transparent: We openly discuss salary expectations to avoid surprises later in the process.

  • Data-driven: We use objective data to remove bias and ensure consistency in compensation decisions.

About Alembic

Alembic is pioneering a revolution in marketing, proving the true ROI of marketing activities. The Alembic Marketing Intelligence Platform applies sophisticated algorithms and AI models to finally solve this long-standing problem. When you join the Alembic team, you’ll help build the tools that provide unprecedented visibility into how marketing drives revenue, helping a growing list of Fortune 500 companies make more confident, data-driven decisions.

About the Role

We’re looking for a Machine Learning Applications Engineer with GPU, Python, and C++ expertise to help productionize cutting-edge causal AI models. You’ll work closely with ML scientists to turn experimental research code into optimized, scalable, and well-structured software that powers Alembic’s real-time analytics and inference systems.

This is a hands-on, performance-focused role where you’ll operate at the intersection of applied ML, systems engineering, and high-performance computing.

Key Responsibilities

  • Translate early-stage ML research and prototypes into reliable, testable, and performant software components

  • Use CUDA, Triton, and Numba to optimize GPU-accelerated workloads for inference and preprocessing

  • Contribute to core libraries and performance-critical routines using modern C++ in hybrid Python/C++ environments
    Develop modular, reusable infrastructure that supports deployment of ML workloads at scale
    Collaborate with data scientists and engineers to optimize data structures, memory usage, and execution paths

  • Build interfaces and APIs to integrate ML components into Alembic’s broader platform
    Implement logging, profiling, and observability tools to track performance and model behavior

Must-Have Qualifications

  • 4–7 years of software engineering experience, including substantial time in Python and C++

  • Hands-on experience with GPU programming, including CUDA, Triton, Numba, or related frameworks
    Strong familiarity with the Python data stack (Pandas, NumPy, PyArrow) and low-level performance tuning
    Experience writing high-performance, memory-efficient code in C++

  • Demonstrated ability to work cross-functionally with researchers, platform engineers, and product teams

  • Comfort transforming research-grade ML code into maintainable, production-grade software

Nice-to-Have

  • Experience with hybrid Python/C++ or Python/CUDA extension development (e.g., Pybind11, Cython, custom ops)

  • Familiarity with ML serving or inference tools (e.g., TorchServe, ONNX Runtime, Triton Inference Server)

  • Exposure to structured data modeling, causal inference, or large-scale statistical computation

  • Background in distributed systems or parallel processing is a plus

What You’ll Get

  • A pivotal role building GPU-accelerated software at the heart of a real-world AI product

  • Collaboration with an elite team of ML scientists, engineers, and product leaders

  • The opportunity to shape performance-critical infrastructure powering enterprise decision-making

  • A culture rooted in technical rigor, curiosity, and product impact

Compensation Range: $235K - $258K