Sunday, March 15, 2026

Does your AI application pisses users or is going beyond the script? Raindrop appears along with the AI’s rescue Platform for monitoring performance

Share


Join our daily and weekly newsletters to get the latest updates and exclusive content regarding the leading scope of artificial intelligence. Learn more


As an enterprise More and more often try to build and implement generative AI powered applications And services for internal or external exploit (employees or customers), one of the most tough questions they encounter, is to understand how well these AI tools are at gigantic.

In fact, recently A survey conducted by the consulting company McKinsey and Company It was found that only 27% of 830 respondents stated that their enterprises reviewed all the results of their generative AI systems before they went to users.

Unless the user actually does not write a complaint report on how the company must know if his product AI behaves as expected and planned?

RaindropEarlier known as Dawn AI, he is a novel startup that fights the challenge, setting up as the first observation platform specially built for artificial intelligence in production, catching errors as it happens, and explaining to enterprises, what went wrong and why. Objective? Lend a hand solve the so -called “black box problem”.

“AI products are constantly disappointing – in the way funny and terrifying” Recently wrote co -founder Ben Hyrak on X“Regular software throws exceptions. But AI products fail in silence.”

Raindrop is trying to offer any tool that defines the category similar to which observation company Sentry He does for established software.

But while established tools for tracking exceptions do not record the nuanced incorrect behavior of gigantic languages ​​models or accompanied by AI, rainy attempts to fill the hole.

“In traditional software you have tools such as Sentry and Datadog to tell you what is wrong in production,” said Venturebeat in an interview with video conversations last week. “There was nothing with AI.”

Until now – of course.

How the rain drop works

Raindrop offers a package of tools that allow teams in gigantic and miniature enterprises to detect, analyze and respond to AI problems in real time.

The platform is at the intersection of user interactions and outputs of models, analyzing patterns in hundreds of millions of daily events, but by doing so with the SOC-2 encryption enabled, protecting the data and privacy of users and a company offering AI.

“Raindrop sits where the user is,” explained Hyrak. “We analyze their messages, as well as signals, such as their thumbs up/down, build errors or whether they have implemented the output data to conclude about what is actually going wrong.”

Raindrop uses a machine learning pipeline, which combines a summary powered by LLM with smaller classifiers on request optimized for scale.

Promotional screenshot of the Raindrop dashboard. Loan: raindrop.ai

“Our ML pipeline is one of the most complex I have seen,” said Hyrak. “We use large LLM for early processing, and then training small, efficient models for a scale on hundreds of millions of events every day.”

Customers can track indicators such as user frustration, task failures, refusal and memory of memory. Raindrop uses feedback signals such as thumb down, user corrections or control behavior (such as unsuccessful implementation) to identify problems.

Co -founder and general director of Raindrop, Zubin Singh Kotich, said Venturebeat in the same interview that while many enterprises were based on evaluation, reference points and unit tests checking the reliability of their AI solutions, not much was designed to check AI during production.

“Imagine traditional coding, if you like:” Oh, my software undergoes ten unit tests. It’s great. It’s solid software. ” Of course, this is not how it works – Koticha said. “This is a similar problem that we try to solve here, where there is not much in production, what he tells you: is it very good? Is it broken or not? And we fit there.”

In the case of enterprises in highly regulated industries or for people looking for an additional level of privacy and control, Raindrop offers a notification, a fully local, first version of the platform addressed to enterprises with strict data operation requirements.

Unlike established LLM tools, the notification is made by Redaction on the client side via SDK and on the server side using semantic tools. It does not store indefinite data and maintains all processing in the customer’s infrastructure.

The Raindrop notification provides daily summaries of exploit and the surface of high signaling problems directly in work tools, such as Slack and teams-including the need to record clouds or complicated Devops configurations.

Advanced error identification and precision

Identification of errors, especially in the case of AI models, is far from straightforward.

“In this space it is difficult that each AI application is different,” said Hyrak. “One customer can build a tool for a spreadsheet, another foreign companion. What looks” broken “is very different between them.” This variability is the reason why the Raindrop system adapts to each product individually.

Each AI rainwater monitors are treated as unique. The platform learns the shape of data and behavior for each implementation, and then builds lively ontology of emissions, which evolves over time.

“Raindrop learns data patterns of each product,” Hyrak explained. “It starts with high-level ontology of typical AI-thor problems like laziness, avalanche of memory or user frustration-a then adapts them to each application.”

Regardless of whether he is a coding assistant who forgets about the variable, a companion of foreign AI, who suddenly calls himself as a man from the USA, and even chatbot, who begins to randomly develop the claims about “white genocide” in South Africa, Raindrop is aimed at resolving these problems with the context possible.

Notifications are designed to be airy and timely. The teams receive notifications of Slack or Microsoft teams when something unusual has been detected, along with suggestions on how to recreate the problem.

Over time, this allows AI programmers to repair errors, improve hints, and even identify system defects in responding their applications to users.

“We classify millions of messages a day to find problems such as broken transmission or complaints of users,” said Hyrak. “It’s about strong and sufficient top surface patterns to justify the notification.”

From the assistant to a rainy dot

The history of the company’s origin is rooted in practical experience. Hyrak, who previously worked as a designer of the human interface at Visionos in Apple and Avionics Software Engineering in SpaceX, began to explore AI after meeting GPT-3 in the first days in 2020.

“As soon as I used the GPT-3-pole ending the text-it broke my mind,” he recalled. “I immediately thought:” It will change the way of interaction with technology. ”

Together with other co -founders of Kotich and Alexis Gauba, Hylak initially built HelperVS code extension with hundreds of paying users.

But Building Sidekick revealed a deeper problem: debugging AI products in production was almost impossible thanks to the available tools.

“We started by building AI products, not infrastructure,” Hyrak explained. “But we saw quite quickly that in order to grow something serious, we needed tools to understand the behavior of AI – and this tool did not exist.”

What began as irritation quickly transformed into basic focus. The team traded by building tools to understand the behavior of AI product in the real world.

In this process they discovered that they were not alone. Many native AI companies lacked visibility in what their users actually experienced and why everything broke. After that, Raindrop was born.

Prices, differentiation and rain flexibility attracted a wide range of initial customers

Raindrop prices are intended to accommodate teams of various sizes.

The start plan is available after USD 65/month, with the price of the meter. Pro tier, which includes non -standard tracking topics, semantic search and local functions, starts from $ 350/month and requires direct commitment.

Although observable tools are not novel, most existing options were built before generative artificial intelligence.

Raindrop stands out, being AI from scratch. “Dasin-Kropka is Roda AI,” said Hyrak. “Most of the observation tools have been built for traditional software. They were not designed to cope with the unpredictability and nuance of LLM behavior in the wild.”

This specificity attracted a growing set of customers, including Clay.com, Tolen and a novel computer.

Raindrop customers include a wide range of artificial intelligence industries – from the tools of generating code to add -oncoming accompanies of AI stories – everyone requiring different lenses about what “improper behavior” looks like.

Born out of necessity

Raindrop growth illustrates how AI building tools must evolve next to the models themselves. When companies send more AI powered functions, the observation becomes necessary-not only for measuring performance, but to detect hidden failures before users escalate them.

With the words of Hylak, Raindrop does for AI, what Sentry did for online applications – except for the rates now include hallucinations, refusals and poorly even intentions. Thanks to the rebrand and expansion of the product, Raindrop bet that the next generation of software observation will be the first due to design.

Latest Posts

More News