April 10, 2025

LLMs effectiveness: key approaches to evaluate AI Large Language Models

Key approaches and the Chatbot Arena Italia benchmark

Evaluating Artificial Intelligence models, particularly large language models (LLMs), is critical to ensuring their quality and reliability in real-world applications. An advanced language model can generate texts of astonishing complexity; however, without efficacy measures, we risk deploying models with severe flaws, ranging from hallucinations to incorrect responses. LLMs produce open-ended and variable responses, making simple comparisons with an "exact solution" insufficient. Indeed, classic automatic metrics often fail to capture fundamental qualitative aspects such as textual fluency, logical coherence, or expressive creativity.

Consequently, the AI community has developed diversified and complementary evaluation approaches. These methods, from using LLMs as judges to human user feedback and standard metrics datasets, provide different perspectives on a model’s performance.

In this article, we explore in detail the main evaluation methods and delve into Chatbot Arena Italia, the platform for comparing Italian-language LLMs.

Main approaches for evaluating LLM 

Effectiveness Methods for assessing the quality of a language model can be grouped into three main categories. Each approach offers different perspectives on the quality and effectiveness of language models.

LLM-as-a-Judge

The concept of using a language model as a judge is a recent and rapidly spreading approach for evaluating AI-generated responses. In practice, an LLM, typically one of the most powerful available, is given a carefully designed prompt to assess another model's output according to predetermined criteria. For instance, when testing two virtual assistants, we could ask the same question and then request the judge model to compare the answers and determine the superior one, providing reasoning. Alternatively, the LLM judge can score a single response against quality parameters (factual correctness, relevance, style, etc.). This method is reference-free, not necessarily requiring a predefined correct response, as judgment is based on the evaluator model's linguistic experience and "common sense." Thus, it’s a flexible mechanism for approximating human judgment through AI itself.

Advantages

  1. LLM-based evaluation provides significant scalability, enabling rapid analysis of thousands of responses, which is difficult to achieve through human assessments.
  2. LLM judges ensure consistency and repeatability in evaluations, reducing typical variability in human judgments.
  3. Altering the prompt easily shifts evaluation towards language aspects like empathy or technical accuracy.
  4. Advanced models like GPT-4 detect stylistic and semantic nuances, offering textual explanations for their judgments, improving interpretability.
  5. Compared to manual evaluation, the LLM-as-a-Judge approach is notably faster and less costly, enabling more frequent development iterations.

Limitations

  1. The reliability of LLM judgments depends directly on the evaluator model used, which may harbor biases or errors affecting evaluation accuracy.
  2. LLM evaluations are inherently probabilistic and may exhibit instability based on specific prompt formulations.
  3. Clear evaluation prompt-writing is crucial, as ambiguity can lead to differing interpretations of the same output.
  4. Although advanced, LLMs still cannot entirely replicate human judgment precision across all contexts; thus, integrating human checks in critical cases is prudent.
  5. If the judge model belongs to the same "family" as the evaluated model, there is a risk of shared biases or training errors.

User feedback and human evaluations 

Involving human evaluators, either end-users or specialized annotators, is a proven method for evaluating generative AI model performance. This approach includes spontaneous feedback, such as star ratings or user likes/dislikes for virtual assistants, and structured methodologies like Reinforcement Learning from Human Feedback (RLHF). In RLHF, human judges compare model-generated responses, expressing preferences used to train a reward model. The goal is to align the model's perceived quality with human expectations, leveraging users’ intuitive assessments of qualities difficult to formalize, such as humor or contextual relevance.

Advantages

  1. Human feedback remains the gold standard for qualitative AI evaluation, as only users can confirm if the system truly meets their needs.
  2. Identifies complex qualitative aspects such as clarity and stylistic or contextual appropriateness, often missed by automated metrics.
  3. Directly aligns models with real end-user needs, continually enhancing perceived satisfaction and practical utility within business digitalization processes
  4. Approaches like RLHF have yielded remarkable results, facilitating significant model evolutions such as the transition from GPT-3 to ChatGPT.
  5. Human feedback frequently reveals unforeseen issues or unexpected model uses, enabling targeted, timely corrective actions.

Limitations

  1. Gathering and managing high-quality human feedback is expensive regarding time, financial resources, and organization, making scalability challenging.
  2. Human judgment is subjective and noisy, with substantial individual differences complicating consensus.
  3. High risk of bias; models might be optimized for the majority preferences, neglecting minorities or specific preferences.
  4. Potential "reward hacking": models optimized via human feedback might learn tricks to maximize perceived scores without genuine qualitative improvements, showing fragility in novel contexts.
  5. Intensive human feedback can reduce response diversity, causing mode collapse and limiting originality and creativity.
  6. In open environments, erroneous or malicious feedback can negatively impact training processes, necessitating additional moderation controls.
  7. Incorporating human feedback into the development cycle is often slow, causing significant delays in addressing user-identified issues.

Standard benchmarks

Using benchmark datasets is an established method for evaluating AI models. This approach involves testing models against datasets with known responses, employing specific metrics to measure how closely a model's output matches expected results. For example, question-answering models use accuracy, while translation models use metrics like BLEU. As LLMs have evolved, increasingly sophisticated benchmarks have emerged, SuperGLUE, MMLU, BIG-bench, HELLASWAG, TruthfulQA, probing competencies from logic to creativity. Benchmark-based evaluation remains crucial, providing objective, repeatable standards.

Advantages

  1. Benchmarks provide clear, quantitative metrics that facilitate direct model comparisons and identify improvements or regressions.
  2. Allow automated evaluation integrated into development pipelines, ensuring efficiency and rapid iteration that can positively influence customer experience
  3. Specific aspects like factual correctness or numerical reasoning are well represented by targeted benchmarks.
  4. Public benchmarks facilitate comparability among models, stimulating research and development innovation.
  5. Ensure models meet minimum performance requirements before real-world deployment, maintaining consistent evaluation standards.

Limitations

  1. Real-world responses often lack a single correct solution; overly rigid benchmarks may penalize correct but slightly differing answers.
  2. Classical metrics (BLEU, ROUGE) may not capture complex qualitative aspects, showing poor correlation with human judgment for lengthy or intricate content.
  3. Models risk "overfitting" benchmarks, achieving high test performance but poor real-world generalization.
  4. Standard datasets rapidly become outdated or insufficient as LLM capabilities evolve, necessitating costly, complex new benchmark production.
  5. Benchmark evaluation doesn’t reflect the end-to-end user experience, missing conversational context or response consistency over time, difficult to capture in static tests.
Try indigo.ai, in a few minutes, without installing anything.
Try a demo

Chatbot Arena Italia: A Cooperative Benchmark for the Italian Language 

Besides the general methods described, having independent platforms to openly compare models on realistic tasks is essential. Internationally, an example is LMSYS’s Chatbot Arena, introduced in 2023, where different chatbots compete blindly to rank based on user preferences. Recently, Italy has filled this gap by launching our Chatbot Arena Italia platform. It is the first benchmark dedicated exclusively to the Italian language, designed to transparently and collaboratively assess AI chatbots' linguistic capabilities.

Platform features 

Chatbot Arena Italia provides intuitive features to facilitate fair and impartial LLM evaluations. Anyone can freely access the platform online and test models in three modes.

  1. "Battle" Arena allows users to enter a prompt simultaneously submitted to two randomly selected anonymous models. Users evaluate responses without knowing the models' identities, eliminating preference biases.
  2. "Side-by-side" Arena enables expert users to explicitly select two models to compare simultaneously, viewing responses directly associated with respective model names. This mode is useful for targeted tests and specific comparisons, such as between GPT-4 and open-source Italian models.
  3. Direct Chat allows single interactions with a specific model, facilitating deep qualitative evaluation while contributing user feedback to the evaluation dataset.

Available models and fair comparison 

A distinctive feature of Chatbot Arena Italia is its variety of available models, from major international models like GPT-4 and Anthropic's Claude, to models specifically trained or adapted for Italian, including Gemma2-9B, Maestrale-chat, Modello Italia 9B, and LLaMAntino. Over thirty models currently compete on the platform, with ongoing new additions. The simultaneous presence of commercial and open-source models allows unprecedented transparent comparisons, highlighting each model's linguistic capabilities. The ranking, built from thousands of anonymous evaluations aggregated through an ELO-like rating system, is updated in real-time. This cooperative method leverages the "wisdom of crowds," ensuring a robust, representative benchmark.

Importance for Italian language evaluation

Chatbot Arena Italia represents a significant national benchmark, addressing transparency gaps regarding LLM performance in Italian. Previously, tests were primarily conducted in English, leaving uncertainties about international models' true Italian capabilities. Already in 2024, preliminary tests indicated that Anthropic's Claude 3 could outperform GPT-4 on specific Italian prompts. A dedicated platform now clearly identifies each model's strengths and weaknesses in Italian. Additionally, Chatbot Arena Italia fosters competition and continuous improvement within the local ecosystem, providing Italian developers immediate performance feedback compared to global giants.

The platform also serves an educational and democratizing purpose, allowing everyone, without costly subscriptions, to access advanced models typically available only through paid services. Students and developers thus freely access advanced tools, lowering entry barriers in LLM research and evaluation.

Finally, Chatbot Arena Italia’s crowdsourced benchmarking provides an insightful example of hybrid evaluation, combining human feedback with automated result aggregation infrastructure. Although not replacing standard benchmarks, it significantly complements them; results can guide new dataset creation targeting model weaknesses or even help develop specialized AI judges for linguistic evaluation. Chatbot Arena Italia thus lays the foundation for an active, aware Italian community in AI evaluation, ensuring conversational AI development in our language keeps pace with global standards.

Evaluating an LLM is not a one-off event but a continuous, multifaceted process. Each evaluation layer provides invaluable insights from offline to online, laboratory to crowdsourcing. Only by combining these insights can we form a comprehensive picture of a model’s capabilities and areas needing refinement. Investing in a robust evaluation ecosystem means investing in AI quality, security, and reliability, a critical step to ensuring language technologies fulfill their promises and foster trust among users and adopting sectors. Serving as both a developer's compass and user transparency criteria, evaluation remains the cornerstone for building the next generation of genuinely effective AI systems aligned with our needs.

FAQs

What is the best method to evaluate an LLM? 

No single evaluation method is universally superior, as each approach has its unique advantages and limitations. For example, human feedback is essential for identifying qualitative nuances like coherence and relevance, but lacks scalability; conversely, using an LLM as an automatic judge is fast and cost-effective, but can introduce errors or biases inherent to the model itself. Standard benchmarks offer objective comparisons but might be restrictive given the variety of possible responses. Hence, the ideal solution is to adopt a multi-level strategy that integrates automatic evaluations, direct human feedback, and benchmark tests.

Does Chatbot Arena Italia genuinely allow the evaluation of the best language model in Italian?

Chatbot Arena Italia is currently the best available tool for directly comparing the capabilities of various language models in Italian. The platform adopts a crowdsourced approach, enabling users to evaluate responses generated by models anonymously. This method transparently identifies which models perform best in Italian, providing a real-time updated ranking based on thousands of genuine votes.

Can LLM judges completely replace human evaluation for language model quality verification? 

Currently, LLM judges cannot fully replace human judgment in evaluating language model quality. While they offer scalability, cost-effectiveness, and speed, they may introduce biases and inaccuracies from their training. Thus, integrating automatic evaluations with human feedback remains critical, particularly for sensitive or critical scenarios, ensuring robust, balanced results aligned with actual user needs.

Sign up for our newsletter
Don't take our word for it
Try indigo.ai, in a few minutes, without installing anything and see if it lives up to our promises
Try a Demo
Non crederci sulla parola
This is some text inside of a div block. This is some text inside of a div block. This is some text inside of a div block. This is some text inside of a div block.

Lorem ipsum dolor sit amet, consectetur adipiscing elit. Suspendisse varius enim in eros elementum tristique. Duis cursus, mi quis viverra ornare, eros dolor interdum nulla, ut commodo diam libero vitae erat. Aenean faucibus nibh et justo cursus id rutrum lorem imperdiet. Nunc ut sem vitae risus tristique posuere.