A new technique can more effectively perform a safety check on an AI chatbotA new technique can more effectively perform a safety check on an AI chatbot.
MIT News Machine learning 3:04 am on May 23, 2024
The MIT News Office announces new research where curiosity-driven red-teaming helps assess large language models against company policies, using funding from various organizations. This method ensures user expectations are met upon model deployment.
- Research: Curiosity-Driven Red-Teaming for Large Language Models
- Funding Agencies: Hyundai Motor Compan, Quanta Computer Inc, MIT-IBM Watson AI Lab, Amazon Web Services MLRA research grantee, DARPA, US Army Research Office, U.S. Navy & Air Force Research Labs.
- Purpose: To test chatbot compliance with company policies through a red-team approach.
- Collaboration: Agrawal's team emphasizes its application in AI model deployment testing.
- Impact: Ensuring language models behave as expected according to user and policy expectations, benefiting industries adopting AI.
https://news.mit.edu/2024/faster-better-way-preventing-ai-chatbot-toxic-responses-0410
< Previous Story - Next Story >