By Emma Schleiger, CSIRO and Stefan Hajkowicz, Data61
The question of whether technology is good or bad depends on how it’s developed and used. Nowhere is that more topical than in technologies using artificial intelligence.
When developed and used appropriately, artificial intelligence (AI) has the potential to transform the way we live, work, communicate and travel.
New AI-enabled medical technologies are being developed to improve patient care. There are persuasive indications that autonomous vehicles will improve safety and reduce the road toll. Machine learning and automation are streamlining workflows and allowing us to work smarter.
Around the world, AI-enabled technology is increasingly being adopted by individuals, governments, organisations and institutions. But along with the vast potential to improve our quality of life, comes a risk to our basic human rights and freedoms.
Appropriate oversight, guidance and understanding of the way AI is used and developed in Australia must be prioritised.
AI gone wild may conjure images of The Terminator and Ex Machina movies, but it is much simpler, fundamental issues that need to be addressed at present. This includes:
- How data is used to develop AI;
- Whether an AI system is being used fairly; and
- In which situations should we continue to rely on human decision-making?
We have an AI ethics plan
That’s why, in partnership with government and industry, we’ve developed an ethics framework for AI in Australia.
The ethical framework looks at various case studies from around the world to discuss how AI has been used in the past and the impacts that it has had. The case studies help us understand where things went wrong and how to avoid repeating past mistakes.
We also looked at what was being done around the world to address ethical concerns about AI development and use.
Based on the core issues and impacts of AI, eight principles were identified to support the ethical use and development of AI in Australia.
1. Generates net benefits
The AI system must generate benefits for people that are greater than the costs.
2. Do no harm
Civilian AI systems must not be designed to harm or deceive people and should be implemented in ways that minimise any negative outcomes.
3. Regulatory and legal compliance
The AI system must comply with all relevant international, Australian local, state/territory and federal government obligations, regulations and laws.
4. Privacy protection
Any system, including AI systems, must ensure people’s private data is protected and kept confidential and prevent data breaches that could cause reputational, psychological, financial, professional or other types of harm.
5. Fairness
The development or use of the AI system must not result in unfair discrimination against individuals, communities or groups. This requires particular attention to ensure the “training data” is free from bias or characteristics which may cause the algorithm to behave unfairly.
6. Transparency and explainability
People must be informed when an algorithm is being used that impacts them and they should be provided with information about what information the algorithm uses to make decisions.
7. Contestability
When an algorithm impacts a person there must be an efficient process to allow that person to challenge the use or output of the algorithm.
8. Accountability
People and organisations responsible for the creation and implementation of AI algorithms should be identifiable and accountable for the impacts of that algorithm, even if the impacts are unintended.
In addition to the core principles, various toolkit items are identified in the framework that could be used to help support these principles. These include impact assessments, ongoing monitoring and public consultation.
A plan, what about action?
But principles and ethical goals can only go so far. At some point, we will need to get to work on deciding how we are going to implement and achieve them.
There are various complexities to consider when discussing the ethical use and development of AI. The vast reach of the technology has the potential to impact every facet of our lives.
AI applications are already in use across households, businesses and governments, most Australians are already being impacted by them.
There is a pressing need to examine the effects that AI has on the vulnerable and on minority groups, making sure we protect these individuals and communities from bias, discrimination and exploitation. (Remember Tay, the racist chatbot?)
There is also the fact that AI used in Australia will often be developed in other countries, so how do we ensure it adheres to Australian standards and expectations?
Your say
The framework explores these issues and forms some of Australia’s first steps on the journey towards the positive development and use of AI. But true progress needs input from stakeholders across government, business, academia and broader society.
That’s why ethical framework discussion paper is now open to public comment. You have until May 31, 2019, to have your say in Australia’s digital future.
With a proactive approach to the ethical development of AI, Australia can do more than just mitigate against any risks. If we can build AI for a fairer go, we can secure a competitive advantage as well as safeguard the rights of Australians.
This article is republished from The Conversation under a Creative Commons license. Read the original article.
COMMENTS
SmartCompany is committed to hosting lively discussions. Help us keep the conversation useful, interesting and welcoming. We aim to publish comments quickly in the interest of promoting robust conversation, but we’re a small team and we deploy filters to protect against legal risk. Occasionally your comment may be held up while it is being reviewed, but we’re working as fast as we can to keep the conversation rolling.
The SmartCompany comment section is members-only content. Please subscribe to leave a comment.
The SmartCompany comment section is members-only content. Please login to leave a comment.