Design and build scalable data pipelines, develop internal data APIs, and ensure data governance for an AI safety startup. Collaborate with engineers and researchers to drive innovation. Experience with modern data stacks and SQL/Python required.
Key Highlights
Technical Skills Required
Benefits & Perks
Job Description
Our client is looking for a Data Engineer to architect and build data pipelines from the ground up, capable of handling petabytes of logs, events, and model traces, and to create a clean, dependable environment for production, testing, and research workloads.
This organization is an AI safety startup developing the safety, reliability, and optimization layer for advanced AI systems. A core part of the platform is a natural-language policy engine: clear rules that define what an AI model should and shouldn’t do. These policies are automatically tested, enforced, and continuously refined at scale.
Its infrastructure processes hundreds of millions of API calls every month, and the team trains and fine-tunes proprietary LLMs that outperform both open-source and commercial models on speed and cost.
This is a small, deeply focused team where your work will ship quickly, solve hard problems, and directly shape how AI safety is implemented in real systems.
What You’ll Do
- Build and maintain a clean, scalable data environment, enabling teammates to access petabytes of traces, logs, and model outputs in the formats they need—without delays or manual data wrangling.
- Develop internal data APIs, SDKs, and tooling that let engineering, product, and research teams discover, query, and use data without managing infrastructure.
- Monitor and optimize data performance—from storage layouts to query execution plans—to ensure analytics and research workflows stay fast as data volumes surge.
- Own data access and governance, defining and enforcing permissions, access controls, and security policies.
What We’re Looking For
- Experience building or scaling a modern data stack (e.g., Snowflake, ClickHouse, event streaming systems) in a startup or similarly fast-moving environment.
- Strong SQL and Python skills, with comfort working on large, noisy datasets.
- Clear communication and the ability to collaborate directly with engineers and researchers. Fluent English.
- Bonus: experience with Metabase, Tableau, or similar internal dashboarding tools.
Compensation & Benefits
- Competitive salary + equity
- 20 days of paid vacation
- Hybrid work from Paris with relocation support
- Top-tier medical insurance
- All required hardware, tools, and services provided
- Paid subscriptions for AI agents and developer tools
- Twice-yearly team off-sites
Hiring Process
- Introductory call with a team member
- Take-home exercise
- Technical interview
- Final discussion with executive leadership