← Back to documents
United States2024-09en

SCALABLE WORKFLOW FOR EVALUATING TRUSTWORTHINESS OF LARGE LANGUAGE MODELS

Summary

Large Language Models (LLMs) present significant trustworthiness challenges, including generating inaccurate or biased outputs and potentially misaligning with human values. Oak Ridge National Laboratory has developed a scalable workflow to comprehensively evaluate open-source LLMs across multiple dimensions of trustworthiness, such as truthfulness, safety, and fairness. This workflow has been applied to evaluate Meta Llama3.1 LLMs and aims to lay the groundwork for future development of more trustworthy AI models.

Key Facts

Available with Pro

Structured Key Facts + original PDF link + AI chat

See pricing

Source Document

https://example-government.gov/policy-document-link

AI chat is part of Pro. See pricing →