Physical Address
304 North Cardinal St.
Dorchester Center, MA 02124
Physical Address
304 North Cardinal St.
Dorchester Center, MA 02124
Take part in our daily and weekly newsletters to get the latest updates and exclusive content for reporting on industry -leading AI. Learn more
In the race for Enterprise AI, an obstacle blocks the way: hallucinations. These invented answers from AI systems have caused everything to honor fictional guidelines from legal sanctions against lawyers to lawyers to forced.
Organizations have tried various approaches Solution of the hallucination challenge, including fine -tuning with better data, access to augmented generation (LAG) and Guidelines. Open source Development company Oumi Now offers a new approach, albeit with a somewhat “cheese” name.
The The company’s name is an acronym for Open Universal Machine Intelligence (Oumi). It is Under the direction of ex-Apple and Google Engineers on the mission to build an unconditional open source You have a platform.
On April 2, the company published Halloumi, an open source claim for review, which is intended to solve the accuracy problem through a new approach to hallucination detection. Halloumi is of course a kind of hard cheese, but that has nothing to do with naming the model. The name is a combination of hallucination and oumi, although the time of publication near the April Fool’s joke would have suspected that the release was a joke – but it is anything but a joke; It is a solution to a very real problem.
“Hallucinations are often called one of the most critical challenges in the use of generative models,” Manos Koukoumidis, CEO of Oumi, told Venturebeat. “In the end, there is a question of trust – generative models are trained to produce products that are likely likely, but not necessarily true.”
Halloumi analyzes the content of ai-generated content of set-to-be ducks. The system accepts both a source document and an AI response and then determines whether the starting material supports any claim in the answer.
“What Halloumi does is to analyze every single sentence independently of one another,” said Koukokoumidis. “For each sentence it analyzes, you are given the specific sentences in the input document that you should check so that you do not have to read the entire document to check whether the (large language model) LLM is exactly or not.”
The model contains three key outputs for each analyzed sentence:
“We trained to be very nuanced,” said Koukoumidis. “Even for our linguists, if the model identifies something as hallucination, we first think that it looks right. If you then look at the reason, Halloumi points out exactly on the nuanced reason why it is a hallucination – why the model made a kind of acceptance or why it is inaccurate in a very nuanced manner.”
There are different ways of how Halloumi can be used and integrated in Enterprise AI today.
One way is to try the model with a somewhat manual process, even though the online Demo interface.
An API-controlled approach will be more optimal for the production and corporate ACI workflows. Manos explained that the model is completely open source and connected to existing workflows, executed locally or in the cloud and used with any LLM.
The process includes feeding the original context and the reaction of the LLM to Halloumi, which then checks the output. Companies can integrate Halloumi to add a test layer to their AI systems and help to recognize and prevent hallucinations in AI-generated content.
Oumi has released two versions: the generative 8b model that offers a detailed analysis and a classifike model that only provides a score, but greater arithmetic efficiency.
What distinguishes Halloumi from other grounding approaches is, as it adds and replaces non -existent techniques such as RAG (access to augmented generation) and at the same time offers more detailed analyzes than typical guardrails.
“The input document that you feed through the LLM could be rag,” said Koukoumidis. “In some other cases it is not exactly rag because people say:” I don’t call anything. I already have the document that is interested. I tell you that this is the document that is important to me. Take it together for me. “This way, Halloumi can apply to rags, but not just on loapping scenarios.”
This distinction is important, because while RAG aims to improve the production with a relevant context, Halloumi checks the output after the generation, regardless of how this context was preserved.
Compared to guardrails, Halloumi offers more than just a binary check. The analysis at the set level with trust and explanations gives users a detailed understanding of where and how hallucinations occur.
Halloumi contains a special form of argument in its approach.
“There was definitely a variant of thinking that we did to synthesize the data,” said Koukokoumidis. “We have led the model step by step or claim through sub -decline to think about how a greater claim or a larger sentence should classify to make the prediction.”
The model can also not only recognize accidental hallucinations, but also deliberate misinformation. In a demonstration, Koukoumidis showed how Halloumi identified as Deepseek’s model Ignored Wikipedia content and instead created propaganda-like content about the reaction of Chinas Covid-19.
For companies that want to go the way in the introduction of AI, Halloumi offers a potentially crucial instrument to safely provide generative AI systems in production environments.
“I really hope that this will unlock many scenarios,” said Koukoumidis. “Many companies cannot trust their models because existing implementations were not very ergonomic or efficient. I hope that Halloumi enables them to trust their LLMs because they now have something to convey the trust they need.”
For companies in a slower AI adoption curve, the open source nature of Halloumi means that they can now experiment with the technology, while Oumi offers commercial support options if necessary.
“If companies better adapt to their domain or have a certain commercial way you should use, we are always happy to help you develop the solution,” added Koukokoumidis.
If AI systems progress, tools like Halloumi can become standard components from Enterprise -ai stacks -Essential infrastructure to separate AI sticks from fiction.