Home/resources/Evaluation Report on “Improving Methodologies for LLM Evaluations Across Global Languages”
The Singapore Conference on AI (SCAI): International Scientific Exchange (ISE) on AI Safety saw over 100 of the best global minds from academia, industry and government, collectively identify and demonstrate consensus around technical AI safety research priorities. In shaping reliable, secure and safe AI, the outcomes of the discussion at SCAI:ISE are synthesized into the Singapore Consensus on Global AI Safety Research Priorities, a living document that continues to welcome views from the global research community.
As part of the International Network of AI Safety Institutes’ continued effort to advance the science of AI model evaluations and work towards building common best practices for testing advanced AI systems, Singapore, Japan, and the United Kingdom led the Network’s latest joint testing exercise aimed at improving the efficacy of model evaluations across different languages
IMDA, together with Humane Intelligence, completed the world's first-ever multicultural and multilingual AI Safety Red Teaming Challenge focused on Asia-Pacific in November and December 2024! Over 350 participants across 9 countries tested 4 large language models for bias stereotypes in English and regional languages. AI safety initiatives like these go towards supporting the ongoing testing work of the AISI Network.<br><br>Stay tuned for the publication of the Challenge Evaluation Report in Feb 2025!