Overview: We are focused on developing advanced LMMs that integrate explainable reasoning and safe generation to optimize healthcare workflows. These models are designed to process and analyze multiple data modalities—such as text, images, and structured data—enabling them to assist in complex tasks like diagnostics, treatment planning, and patient management with exceptional accuracy.
Overview: This research project is at the forefront of integrating advanced Large Language Models (LLMs) into the process of deriving actionable insights from vast and complex document repositories. This initiative focuses on creating a system that allows users to interact with the LLM, guiding its analysis and refining the results based on the user’s expertise and evolving needs.
Overview: Our research spans many areas of text understanding and text generation, with a particular emphasis on factuality checking and factually guided text generation. NEC’s products help humans safely draw conclusions from large quantities of text that they don’t have the time to read. As a leader in the FEVER fact extraction and verification competitions, we have developed systems that achieved higher evidence precision and higher robustness to adversarial attack, and pioneered the ability to pursue missing evidence through multiple retrieval steps.
Overview: The Trustworthy Generative AI Project is focused on developing advanced multimodal generative models that can create and reason with content across text, images, reports, and 3D videos. These models are designed for applications in advertisement, entertainment, law enforcement, and healthcare.