Machine Learning Engineer, LLM Evals & Observability
full-time
junior
Posted 1 day ago
About this role
About Glean:
Glean is the Work AI platform that helps everyone work smarter with AI. What began as the industry’s most advanced enterprise search has evolved into a full-scale Work AI ecosystem, powering intelligent Search, an AI Assistant, and scalable AI agents on one secure, open platform. With over 100 enterprise SaaS connectors, flexible LLM choice, and robust APIs, Glean gives organizations the infrastructure to govern, scale, and customize AI across their entire business - without vendor lock-in or costly implementation cycles.
At its core, Glean is redefining how enterprises find, use, and act on knowledge. Its Enterprise Graph and Personal Knowledge Graph map the relationships between people, content, and activity, delivering deeply personalized, context-aware responses for every employee. This foundation powers Glean’s agentic capabilities - AI agents that automate real work across teams by accessing the industry’s broadest range of data: enterprise and world, structured and unstructured, historical and real-time. The result: measurable business impact through faster onboarding, hours of productivity gained each week, and smarter, safer decisions at every level.
Recognized by Fast Company as one of the World’s Most Innovative Companies (Top 10, 2025), by CNBC’s Disruptor 50, Bloomberg’s AI Startups to Watch (2026), Forbes AI 50, and Gartner’s Tech Innovators in Agentic AI, Glean continues to accelerate its global impact. With customers across 50+ industries and 1,000+ employees in more than 25 countries, we’re helping the world’s largest organizations make every employee AI-fluent, and turning the superintelligent enterprise from concept into reality.
If you’re excited to shape how the world works, you’ll help build systems used daily across Microsoft Teams, Zoom, ServiceNow, Zendesk, GitHub, and many more - deeply embedded where people get things done. You’ll ship agentic capabilities on an open, extensible stack, with the craft and care required for enterprise trust, as we bring Work AI to every employee, in every company.
About the Role:
Building a great AI assistant is only half the battle – knowing whether it's actually great is the other half. Our team owns the measurement and quality layer that make Glean's Assistant and Agents reliably better over time: evaluation pipelines, quality eval-sets, LLM-powered judges, agent observability, and the tooling engineers use to understand what changed and why. It's a rare combination of infrastructure engineering, applied ML, and direct product impact. If you care deeply about quality and want to build the systems that make it measurable, this role is for you.
You will:
Design and curate evaluation datasets – sampling strategies, query diversity, and golden sets that give reliable, representative coverage of real assistant behavior.
Build and maintain large-scale evaluation pipelines that measure assistant quality across thousands of real user queries.
Build LLM-powered judges that score metrics like correctness, completeness, and response quality, and align them against human judgment.
Evaluate new models and product changes before they ship – providing the quality signal that gates launches and prevents regressions.
Build observability infrastructure for AI agents: trace enrichment, data pipelines, and dashboards that make assistant behavior inspectable.
Close the loop between quality measurement and improvement using eval results, customer feedback, and techniques like automated prompt iteration to help drive concrete gains in assistant behavior.
Collaborate with engineers across the company to make evals a first-class part of how we ship.
About you:
2+ years of software engineering experience with strong coding skills.
Strong backend fundamentals in Go and Python; comfortable with distributed data pipelines.
Experience working with LLM evaluation, reinforcement learning from human feedback, natural language processing, or other large systems involving machine learning.
Analytically rigorous – you think carefully about what offline metrics actually predict about real user experience.
Thrive in a customer-focused, tight-knit and cross-functional environment - being a team player and willing to take on whatever is most impactful for the company
You care about quality – not just in the systems you build, but in the product you're helping measure and improve.
Location:
This role is hybrid (3-4 days a week in one of our SF Bay Area offices)
Compensation & Benefits:
The standard base salary range for this position is $200,000 - $300,000 annually. Compensation offered will be determined by factors such as location, level, job-related knowledge, skills, and experience. Certain roles may be eligible for variable compensation, equity, and benefits.
We offer a comprehensive benefits package including competitive compensation, Medical, Vis
Similar Jobs
Related searches:
On-site Jobs
Junior Jobs
On-site Junior Jobs
Junior Machine LearningJunior AI Agents & RAGJunior Data EngineeringJunior NLP & Language AIJunior AI InfrastructureJunior Robotics & AutonomyJunior AI ResearchJunior Data Science
AI Jobs in Mountain View
Machine Learning in Mountain ViewAI Agents & RAG in Mountain ViewData Engineering in Mountain ViewNLP & Language AI in Mountain ViewAI Infrastructure in Mountain ViewRobotics & Autonomy in Mountain ViewAI Research in Mountain ViewData Science in Mountain View
agentsllmclouddata-pipelinenlpreinforcement-learningevaluationmachine-learning
Get jobs like this delivered weekly
Free AI jobs newsletter. No spam.