Logo

The Data Daily

Opinion | Will artificial intelligence save humanity, or supersede it? A vanguard is carving a path on AI governance

Opinion | Will artificial intelligence save humanity, or supersede it? A vanguard is carving a path on AI governance

The question has arisen with escalating frequency in recent years, a sort of journalistic thought bubble emerging from the collective consciousness of writers. Will artificial intelligence (AI) save humanity, or supplant us?

On the one hand, we are told that AI holds the potential to solve some of the world’s biggest problems — challenges like poverty, food insecurity, inequality and climate change. On the other hand, some very smart people have issued warnings. Elon Musk called AI “humanity’s greatest threat.” Stephen Hawking said the technology could “spell the end of the human race.” That’s some major alarm-sounding.

Globally, there has been a surge in advocacy for ethical AI. The Organisation for Economic Co-operation and Development catalogues more than 60 countries with national AI strategies, but few have hard laws yet. There is proposed AI legislation in the European Union and the United States, and on March 1, China passed a sweeping law governing how online recommendations are made through algorithms. Last month, the Canadian federal government joined the party, introducing the Artificial Intelligence and Data Act.

The proposed Artificial Intelligence and Data Act is one of three regulatory components in the Digital Charter Implementation Act, 2022. If passed, it will require firms designing, developing and using high-impact AI systems to meet certain requirements aimed at identifying, assessing and mitigating bias and harm. The Artificial Intelligence and Data Act outlines seven requirements relating to anonymization of data, system impact assessment, risk management, system monitoring, record keeping, publication of system description and notification of material harm. This proposed legislation is noteworthy because it is the first example of a hard law around AI being aimed at the private sector in Canada.

It will take some time to understand, in practical terms, what companies will need to do to comply, how they will demonstrate compliance and what the federal government’s approach to enforcement will look like. As with the EU’s General Data Protection Regulation, the penalties for offences are severe and can result in fines of up to $10 million or three per cent of a company’s gross global revenues.

What does seem likely is that this type of law will induce the creation of an entire new services market — the market for AI assurance. The push to develop robust approaches to assuring the trustworthiness of AI is evident in places such as the United Kingdom, where the cultivation of a world-class AI assurance industry is seen as a critical component for harnessing the technology’s transformative potential.

There is a vanguard of practitioners carving a path with AI governance, but as yet there are few practical models. That will change with regulation. Firms and other organizations who are designing, developing and deploying AI will soon have a concrete reason to figure out what it means to practise AI governance — a function that encompasses establishing guardrails, applying specialty tools and employing services including education, conformity assessment and third-party audits.

The firms who do it well will integrate the governance of AI into their enterprise risk-management systems, identifying and mitigating risks at each stage of the machine-learning life cycle. AI assurance is both a market need and a market opportunity, and there will be a flood of new entrants to the space.

Who will enter the fray? The opportunity seems ripe for the chartered professions, with their combination of subject matter expertise, credentialing and integrity — but large institutional elements don’t always lead, as size and speed are often in tension. Perhaps we will see the emergence of a new profession, based on a composite of data science, information technology and assurance skills. Maybe the start-up community will claim this territory, with a combination of agility and expertise.

In the next five years we will see the growth of a major new market in technology assurance, and much of it will be built around AI. Disclosure will be the linchpin in this emerging field.

Images Powered by Shutterstock