Our engineering leaders are at the forefront of Retool's product development, bridging the gap between engineering excellence and customer impact. We look for leaders who not only bring strong technical expertise but also the strategic vision to shape Retool's product direction, balancing day-to-day execution with long-term thinking. In this role, you'll lead Retool's Evaluations & Observability platform. You'll own setting the bar for what "good" looks like across our newly-launched Assist experience, making sure it works and works consistently, at scale. You'll build the systems, tools, and culture that let us measure, understand, and improve quality in real time, driving relentless iteration across everything we ship. You'll guide engineers focused on: Evaluation platforms: building the frameworks that let us test and compare performance across LLM providers and model versions. Quality systems: defining and enforcing rubrics, metrics, and evaluation loops that answer the hardest question in AI: "Is this actually good?" Data curation: managing the datasets that power and test our AI models, sourced from real-world usage to keep our systems grounded and relevant. Search & retrieval quality: owning the retrieval layer that underpins both AI and non-AI experiences ensuring results are relevant, accurate, and fast. Reusable AI quality infrastructure: creating the building blocks (evaluation tools, pipelines, and feedback systems) that other teams can leverage to maintain quality across Retool's AI surface area. Culture of continuous improvement: embed a data-driven approach to AI quality, where experimentation and measurement are the default as we scale our capabilities.
Stand Out From the Crowd
Upload your resume and get instant feedback on how well it matches this job.
Job Type
Full-time
Career Level
Manager
Industry
Publishing Industries
Education Level
No Education Listed
Number of Employees
251-500 employees