Giskard’s open-source framework evaluates AI models before they’re pushed into production
Giskard is a French startup working on an open-source testing framework for large language models. It can alert developers of risks of biases, security holes and a model’s ability to generate harmful or toxic content. While there’s a lot of hype around AI models, ML testing systems will also quickly become a hot topic as regulation is about to be enforced in the EU with the AI Act, and in other countries. Companies that develop AI models will have to prove that they comply with a set of rules and mitigate risks so that they don’t have to pay hefty fines. Giskard is an AI startup that embraces regulation and one of the first examples of a developer tool that specifically focuses on testing in a more efficient manner. “I worked at Dataiku before, particularly on NLP model integration. And I could see that, when I was in charge of testing, there were both things that didn’t work well when you wanted to apply them to practical cases, and it was very difficult to compare …