JAKARTA: Rapid advancement in artificial intelligence (AI) brings both opportunities and risks to Indonesia.
Imagine finding your face in a viral deepfake video you never authorised or being denied a job by an AI system with hidden biases.
These scenarios are becoming a reality as AI reshapes our world.
From health care to criminal justice, AI’s influence continues to grow, often imperceptibly.
Without proper oversight, this technology could widen social divides, compromise privacy and concentrate power in the hands of a few tech giants.
Government regulation is therefore crucial to ensure AI serves the public good and that individual rights are protected in our increasingly automated society.
The European Union has taken a bold step with its new AI Act, which entered into force on Aug 1.
This comprehensive law aims to create a unified rule book for AI systems across all 27 EU countries.
It categorises AI systems based on their potential risks: Unacceptable, high, limited or minimal.
Some harmful AI practices are banned outright, while high-risk systems face strict requirements.
The law also pushes for more transparency in specific AI applications and establishes new regulatory bodies to oversee compliance and foster cooperation between member states.
Australia has also joined the regulatory efforts.
Its federal government has proposed mandatory safeguards for high-risk AI systems and a voluntary safety standard for organisations using AI.
The Australian approach outlines 10 interconnected guidelines that set clear expectations for everyone in the AI supply chain, emphasising accountability, transparency and human oversight.
For Indonesia, these global developments, coupled with our own Data Protection (PDP) Law, present significant implications.
As a rapidly growing nation with a thriving tech scene, we must closely monitor these international trends.
We can learn from the EU’s risk-based approach and Australia’s voluntary standards to encourage responsible AI development domestically.
This presents an opportunity for Indonesia to actively participate in international discussions on AI governance, ensuring our perspectives are represented.
Building local expertise in AI ethics and regulation will prove invaluable in addressing future challenges.
The need for AI regulation stems from its potential for serious harm if left unchecked.
AI systems trained on biased data risk perpetuating and amplifying societal inequalities.
The vast amounts of data these systems require raise concerns about personal information protection.
Many AI systems operate as “black boxes”, making it difficult to scrutinize their decision-making processes – a troubling lack of transparency, especially when these decisions can have profound consequences.
Workforce disruption is another significant concern as AI capabilities advance.
The concentration of advanced AI technologies within a few tech firms could lead to unprecedented monopolistic power.
Security vulnerabilities in AI systems pose risks of hacking or malicious exploitation.
The rise of generative AI also introduces the potential for widespread misinformation, blurring the line between authentic and artificially generated content.
Given these risks, several areas require regulation in the AI landscape.
High-risk AI applications in the healthcare, criminal justice, finance and employment sectors need clear guidelines.
Comprehensive rules must govern how personal data is used to train AI systems and inform decision-making processes.
The PDP Law should provide a solid foundation in this area, with its provisions on consent and data processing serving as a starting point that can be further leveraged and expanded.
Transparency is key – people have the right to know when AI is making important decisions about their lives.
This aligns with the PDP Law’s emphasis on the rights of data subjects and the obligations of data controllers and processors.
Robust safety and testing standards must be implemented to ensure reliability and fairness before deploying any AI system.
Additionally, clear accountability mechanisms must be in place when things go wrong. Regulation must, however, be balanced to avoid stifling innovation.
Basic research and development of AI technologies should have some flexibility.
Low-risk AI applications, like spam filters or video game AI, may not need heavy-handed regulation.
Overly specific technical rules should be avoided, as the field moves so quickly that such regulations could become rapidly obsolete.
A lighter touch might be appropriate for AI in creative and artistic domains.
The challenge lies in striking a fine balance between protecting public interests and fostering innovation.
This is not a task for the government alone – it requires collaboration between the government, companies and consumers.
The government’s role is to develop clear, flexible and enforceable AI regulations, building upon the foundation of the PDP Law.
It should invest in AI research and education to build national capabilities.
Creating platforms for dialogue among all stakeholders is crucial, as is ensuring that regulations can keep pace with rapid technological changes.
Companies must integrate ethical AI principles into their product development from the start, adhering to the data protection principles outlined in the PDP Law.
Investing in bias mitigation and explainable AI technologies is essential.
They should be transparent with customers and stakeholders about their AI use.
Consumers also bear responsibilities. We must educate ourselves about AI and its implications, demand transparency from companies using AI systems and participate in public consultations on AI regulation.
It is up to us to exercise our data rights and make informed choices about AI-driven services.
Indonesia should focus on developing AI literacy programmes, establishing industry standards for responsible AI development, implementing mechanisms for ongoing assessment of AI’s societal impact and fostering a culture of responsible innovation in the AI sector.
By taking this collaborative, proactive approach to AI governance, Indonesia can harness the benefits of this transformative technology while mitigating its risks.
As AI continues its rapid evolution, our governance strategies must keep pace to ensure it serves the greater good of society.
The future of AI is unfolding now, and we must shape it responsibly.
Let us ensure it is a future that benefits all.
[The writer is an associate professor in digital strategy and data science at Monash University, Indonesia.]