Australian council addresses AI Risks, touts strong data management
ICA calls for government-industry-academia collaboration.
The Insurance Council of Australia (ICA) risks associated with generative artificial intelligence (AI) but emphasises the industry's robust data management and risk mitigation frameworks.
These frameworks, supported by existing consumer protection, privacy legislation, and financial services regulation, position the industry well to leverage AI for consumer benefit whilst managing risks.
Representing approximately 89% of private sector general insurers, the ICA highlights the industry's economic contribution, employing around 60,000 people, generating AU$64.5b in gross written premiums annually, and paying out AU$147m in claims daily.
New AI-driven risks include fraud, such as the use of AI to augment photographs for inflated claims and voice spoofing, and operational risks, like the accidental sharing of proprietary information through public AI models.
Despite these challenges, AI presents opportunities to improve affordability by reducing claims processing costs, identifying risks, and moderating premium growth.
ALSO READ: Australia calls for road data sharing initiatives
AI can also enhance vulnerability detection, service delivery, and disaster response, accelerating claims processes and securing necessary resources quickly.
To foster a responsible AI industry, the ICA advocates for strengthening Australia’s sovereign AI capabilities, ensuring that AI automation frees workers for more complex tasks, thereby enhancing job satisfaction.
A national education and talent pipeline is crucial for equipping Australians with the skills to develop and manage AI systems, enabling the country to navigate AI-related threats effectively.
The ICA highlights the need for government-industry-academia collaboration to plan for national AI requirements, warning against over-reliance on foreign AI technologies, which could compromise economic autonomy and misalign with Australian values.
Building domestic AI capabilities would ensure AI systems reflect Australian values, fostering responsibility, trust, and transparency between AI deployers and users.
The 2024 Edelman Trust Barometer indicates low trust in AI and innovation, underscoring the need for increased transparency and positive use cases.
Strengthening and clearly articulating the regulatory framework governing AI use will build public trust. The ICA remains committed to assisting the committee with its inquiry.