We want to build agents that autonomously validate code changes. Today that looks like AI that reviews pull requests in GitHub, catching bugs and enforcing standards. We're reviewing close to 1B lines of code a month now for over 1,000 companies.
Problems we're excited about:
Coding standards can be idiosyncratic and are often poorly documented; can we build agents that learn them through osmosis like a new hire might?
Can we identify for each customer what types of PR feedback they do and don't care about, perhaps using some sample efficient RL, in order to increase signal-to-noise ratio?
Some bugs are best caught by running the code, potentially against discerning AI-generated E2E tests. Can we autonomously deploy feature branches and use agents to parallel try to break the application to detect bugs?
Commercials:
Went from 0→XM in revenue in 11 months, growing 25% a month
$30M raised from Benchmark, YC, Paul Graham and others.
Team:
Engineering: senior and staff engineers that are ex Google, LinkedIn, Rivian, best characterized by extreme intellectual curiosity and very high agency.
Head of growth was 3rd growth hire at Figma, led growth for FigJam
Enterprise sales lead launched Stripe in the UAE, before that was the first seller at HuggingFace and first few at Postman and New Relic.
Designer did this → greptile.com
Qualifications
BS in Computer Science of equivalent
Some research experience, ideally with ML/LMs/Agents
Strong programming skills and good product intuition
Responsibilities
Experiment with and apply recent advancements in agents/LMs etc. to improve the performance and capabilities of our products
Example: you might study multi-agent architectures, prototype and evaluate a multi-agent code review workflow, and then work with a
#J-18808-Ljbffr
Apply for this Job
Please use the APPLY HERE link below to view additional details and application instructions.