On December 25, Ricoh announced the development of a new “Safeguard Model” designed to detect harmful information generated by large language models (LLMs). The update marks a significant evolution in generative AI safety, extending beyond traditional input filtering to also monitor and block problematic LLM outputs, creating a multi-layered guardrail architecture for enterprise AI deployments. As generative AI grows in use across industries, worries about misinformation, data leaks, discrimination, and inappropriate content have increased. Ricoh's announcement tackles these risks head-on. There’s a rising demand from companies, especially in Japan, for AI systems. They want systems that are not just powerful but also trustworthy, auditable, and in line with business and regulatory needs. From Prompt Filtering to Full-Stack AI Safety Until recently, most guardrail systems focused on identifying harmful prompts before…
Sign in to your account