Toggle light / dark theme

Ways to Deal With Hallucinations in LLM

Originally published on Towards AI.

One of the major challenges in using LLMs in business is that LLMs hallucinate. How can you entrust your clients to a chatbot that can go mad and tell them something inappropriate at any moment? Or how can you trust your corporate AI assistant if it makes things up randomly?

That’s a problem, especially given that an LLM can’t be fired or held accountable.