@tchambers well this is based on datasets the LLMs are actually trained on.... It is unrealistic that users will ask the same questions. Then, depending on the context with prompt that can be completely different. Also the safety guards put by some LLMs may obfuscate the results for "publicly known datasets" in either direction. The leatherboard has the same issues as LLM: Nobody can verify that the leatherboard delivers meaningul results.
@tchambers there is no specific hallucination rate of a LLM. The errors of LLM change depending on the context and what you ask. It is impossible to measure a specific rate - it is a nonsense measure as it always depends on the context.
Interested in #data #nosql #cloud #geospatial #radio #ai and #analytics supported by #environmental #sustainable #opensource #science #software #europePosts older than 1 year are automatically deleted (except favourites) - we do not need to keep all digital content forever.Ich lasse meine Posts automatisch nach einem Jahr löschen - außer Favoriten - wir müssen nicht alle digitalen Inhalte behalten.