That's all according to a Thursday report from researchers at Arthur AI, a machine learning monitoring platform.
AI hallucinations occur when large language models, or LLMs, fabricate information entirely, behaving as if they are spouting facts.
Meta's Llama 2, on the other hand, hallucinates more overall than GPT-4 and Anthropic's Claude 2, researchers found.
In a second experiment, the researchers tested how much the AI models would hedge their answers with warning phrases to avoid risk (think: "As an AI model, I cannot provide opinions").
"Making sure you really understand the way the LLM performs for the way it's actually getting used is the key."
Persons:
Claude, Arthur AI, It's, Adam Wenchel, Arthur, OpenAI's, Claude 2, Wenchel, that's
Organizations:
Microsoft, Arthur, CNBC, New, Moroccan
Locations:
New York