• AiNews.com
  • Posts
  • Transforming AI Safety Discussions into Concrete Actions

Transforming AI Safety Discussions into Concrete Actions

A tech summit setting with industry leaders and government officials discussing AI safety. The scene includes a large conference room with speakers on stage, a diverse audience, and banners highlighting AI technology and safety. Background elements feature symbols representing AI, such as neural networks and digital shields. The atmosphere is serious and focused, reflecting the importance of transforming AI safety discussions into actionable measures

Transforming AI Safety Discussions into Concrete Actions

At the Asia Tech x Singapore 2024 summit, industry leaders and government officials emphasized the need to turn discussions on artificial intelligence (AI) safety into tangible actions. Despite increased awareness, there is still a lack of concrete measures.

The Need for Practical Solutions

Ieva Martinekaite, head of research and innovation at Telenor Group, stressed the importance of moving from discussions to practical steps. Speaking to ZDNET, Martinekaite, who is also involved with the Norwegian Open AI Lab and Singapore's Advisory Council on the Ethical Use of AI and Data, pointed out that while awareness is high, actionable frameworks are still missing.

Challenges in High-Level Meetings

Government ministers and industry delegates acknowledged the lack of substantial progress despite numerous meetings on AI safety. Martinekaite called for the development of playbooks, frameworks, and benchmarking tools to ensure the safe deployment of AI. She also emphasized the need for ongoing investments to support these initiatives.

Addressing the Threat of Deepfakes

AI-generated deepfakes pose significant risks, especially to critical infrastructure. Martinekaite highlighted that the technology behind deepfakes has advanced, making them harder to detect. Cybercriminals can use this technology to steal credentials and gain unauthorized access to systems. She underscored the need for training, tools, and technologies to identify and prevent AI-generated content, such as digital watermarking and media forensics.

Balancing Regulation and Innovation

Martinekaite advocates for targeted regulations that address high-risk sectors, like critical infrastructure, without stifling innovation. Regulations should focus on areas where deepfake technology has the most significant impact, such as implementing watermarking, source authentication, and data access control.

Global Cooperation and Technological Solutions

Natasha Crampton, Microsoft's chief responsible AI officer, noted an increase in deepfakes and cyberbullying. She highlighted Microsoft's efforts to monitor deceptive online content, particularly around elections. German state secretary Stefan Schnorr emphasized international cooperation to protect against misinformation and safeguard democratic processes.

Zeng Yi, director of the Brain-inspired Cognitive Intelligence Lab, suggested establishing a global observatory to monitor and exchange information on disinformation. This infrastructure could help inform the public and prevent the spread of false content.

Singapore's AI Governance Framework

Singapore has updated its governance framework for generative AI, building on its previous AI governance initiatives. The Model AI Governance Framework for GenAI includes dimensions such as incident reporting, content provenance, security, and testing. This framework aims to balance addressing AI concerns with fostering innovation.

Josephine Teo, Singapore's Minister for Communications and Information, highlighted the importance of risk mitigation and evidence-based regulations. Singapore plans to enhance its governance capabilities, focusing on malicious AI-generated content and ensuring that organizations understand both the advantages and limitations of AI.

Telenor's Approach to AI Governance

Martinekaite described Telenor's approach to AI governance, which includes monitoring new AI tools and reassessing potential risks. Telenor has created a task force to oversee responsible AI adoption, establishing principles, rulebooks, and standards for its employees and partners. This approach ensures the lawful and secure use of AI technology.

Future of AI Governance

As organizations use their own data to train AI models, discussions around data usage and management will intensify. Compliance with new laws, such as the EU AI Act, will drive conversations about data curation and tracing. Organizations will need to examine their agreements with AI developers to meet these additional requirements.


While discussions on AI safety are prevalent, concrete actions and frameworks are still lacking. The Asia Tech x Singapore 2024 summit highlighted the need for practical measures to ensure AI is deployed safely. Global cooperation, technological solutions, and evidence-based regulations are crucial to addressing the challenges posed by AI and deepfake technology.