Google’s Angular team has released Web Codegen Scorer, a tool that evaluates the quality of web code generated by Large Language Models (LLMs). The tool, introduced on September 16, focuses on web code generation and comprehensive quality evaluation. It helps the Angular team create fine-tuned prompts for optimizing LLMs for the framework and integrate application features and syntax as the framework evolves. Web Codegen Scorer can be used to make evidence-based decisions about AI-generated code, allowing developers to iterate on system prompts, compare code quality from different models, and monitor generated code quality as models and agents evolve. It can be used with any web library or framework, or none at all. Specific capabilities include: Configuring evaluations with different models, frameworks, and tools; Specifying system instructions and adding MCP (Model Context Protocol) servers; Built-in checks for build success, runtime errors, accessibility, security, LLM rating, and coding best practices; Automatic attempts to repair issues detected during code generation; Viewing and comparing results with a report viewer UI.