Anyone who develops an AI solution sometimes goes on a journey into the unknown. At least at the beginning, researchers and designers do not always know whether their algorithms and AI models will work as expected or whether the AI will ultimately make mistakes.
Sometimes, AI applications that work well in theory perform poorly under real-life conditions. In order to gain the trust of users, however, an AI should work reliably and correctly. This applies just as much to popular chatbots as it does to AI tools in research.
Any new AI tool has to be tested thoroughly before it is deployed in the real world. However, testing in the real world can be an expensive, or even risky endeavor. For this reason, researchers often test their algorithms in computer simulations of reality. However, since simulations are approximations of reality, testing AI solutions in this way can lead researchers to overestimate an AI’s performance.