
Anthropic, Google, Microsoft, and
OpenAI have formed a new industry body to promote the safe and responsible development of artificial intelligence (AI) systems.
The goal of the group -- Frontier Model Forum -- is to advance
AI safety research, identifying best practices and standards, and facilitate information sharing among policymakers and industry.
“We’re excited to work together with other leading
companies, sharing technical expertise to promote responsible AI innovation,” Kent Walker, president of global affairs at Google and Alphabet, wrote in a post. “We're all going to need to
work together to make sure AI benefits everyone.”
Early in June, OpenAI published research on Frontier AI
Regulations.
advertisement
advertisement
The company defined Frontier AI models as advanced AI and machine-learning models that are considered dangerous enough to pose "severe risks to public safety,” and are
considered large language models (LLM) that surpass the capabilities of existing models.
To address these challenges, three building blocks for the regulation of frontier models are
needed.
Requirements include standard-setting processes to identify appropriate requirements for frontier AI developers, registration and reporting requirements to provide regulators with
visibility into frontier AI development processes, and mechanisms to ensure compliance with safety standards for the development and deployment of frontier.
Membership to the forum is open to
organizations that develop and deploy frontier models, demonstrate strong commitment to safety through technical and institutional approaches, and contribute to advancing the Forum’s efforts
like participating in joint initiatives and supporting developments.
Further work is needed on safety standards and evaluations to ensure that frontier AI models are developed and deployed
responsibly.
The Forum will become one way for industry discussions and actions on AI safety and responsibility to take place.
Initially, the forum will focus on three areas
during the coming year, including identifying best practices, advancing AI safety research and facilitating information-sharing among companies and government.
In the coming months, the
Frontier Model Forum will set up an advisory board.
The founding companies will set institutional arrangements including a charter, governance and funding with a working group and executive
board to lead these efforts.