THANK YOU FOR SUBSCRIBING
Be first to read the latest tech news, Industry Leader's Insights, and CIO interviews of medium and large enterprises exclusively from Education Technology Insights
THANK YOU FOR SUBSCRIBING
Tanveer Zia is a Professor and Head of Computer Science at the University of Notre Dame Australia. With 25 years of experience in the Australian tertiary education sector, he specialises in the security and privacy of computing systems. Additionally, he explores the social and ethical concerns of artificial intelligence and emerging technologies.
Previously, Zia was the Associate Director and was a founding member of the Centre of Excellence in Cybercrimes and Digital Forensics at the Naif Arab University for Security Sciences from February 2021 to 2024. Before that he served in various academic and institutional leadership roles at Charles Sturt University Wagga Wagga Campus, where Zia was the Associate Head of the School of Computing, Mathematics and Engineering.
Innovation and Ethical Responsibility
In recent years, Artificial Intelligence (AI) has become a prominent technology across various industries in both the public and private sectors. AI systems are increasingly integrated into finance, healthcare, human resources, manufacturing, automation, transportation, smart cities, security, customer service and critical infrastructure, among other areas. It is anticipated that digital assistants powered by generative AI will soon become the standard in customer service. Despite the numerous advantages offered by AI advancements, there are significant social and ethical concerns associated with its adoption. Achieving a balance between leveraging technological progress and addressing its social implications is crucial for maximising the benefits of AI.
“It is anticipated that digital assistants powered by generative AI will soon become the standard in customer service.”
Ensuring that new AI developments are inclusive and free from biases is essential. This requires integrating ethics by design into every stage of technology development, from research to deployment. Involving individuals from diverse backgrounds and experiences in the technology design process is vital to create inclusive technologies and address potential biases, especially in machine learning algorithms that rely on extensive datasets. These large datasets also raise issues related to data privacy, civil liberties, and the potential misuse of personal information.
Key questions arise: How can these risks be mitigated? Is it feasible to implement privacy by design in AI systems? How can we address concerns about deepfakes and misinformation? What if AI algorithms are used to manipulate public opinion and influence democratic processes? Additionally, as the new generation becomes increasingly dependent on technology, what impact will this have on their critical thinking and problem-solving skills? There is no doubt that Generation Z is the most educated generation of our times, do they have essential communication and social skills to excel in the workforce? Is the abundant access to generative AI influencing their cognitive development and social behaviour?
With AI mimicking human behaviour and thought processes, how will this affect human relationships? For instance, increased interaction with technology might diminish personal interactions and affect family and social lives. Technology vendors, like those behind Apple’s Siri, Amazon’s Echo, and Google’s Home, can track user interests and behaviours, raising concerns about constant surveillance and privacy breaches.
When introducing new technology, it is crucial to communicate both its potential benefits and risks. For instance, fears about AI-induced job losses are becoming a reality. Just recently, one of the big four banks in Australia has announced its intention to replace thousands of call service jobs with AI-driven chatbots. It is important to equip individuals with the knowledge and skills to adapt and use technology effectively, rather than resisting its adoption or completely replacing the human roles.
These are just a few of the social impacts associated with AI advancements. As AI technology continues to evolve, addressing these concerns and ensuring that AI is developed and utilised in ways that benefit society is essential.
Read Also
I agree We use cookies on this website to enhance your user experience. By clicking any link on this page you are giving your consent for us to set cookies. More info