Study reveals 75 percent of Australian university staff use AI in academics: What India can learn from Australia’s Gen AI framework – Times of India


Artificial intelligence (AI) is increasingly becoming part of everyday life, reshaping industries and redefining how we interact with technology. In recent years, AI’s role in education has evolved rapidly, leading to a significant intermingling of learning practices and AI tools.
Students and teachers alike have begun leveraging generative AI technologies for various purposes, from automating tasks to enhancing learning experiences.As these tools have integrated into classrooms, educators worldwide are witnessing both the benefits and challenges posed by this new era of digital learning.
A recent study sheds light on the extent of AI’s adoption in higher education. The report, focusing on Australian universities, revealed that 75% of academic staff are utilising generative AI in some capacity. This figure represents a clear indication of how AI is becoming an indispensable tool in academics, used for improving administrative functions, teaching methods, and student engagement. The survey included a diverse group of university professionals, providing a comprehensive overview of AI’s impact on the educational ecosystem.

Australian Framework for Generative Artificial Intelligence (AI) in Schools

In response to this growing integration of AI in education, Australia has proactively developed frameworks to regulate and guide its use. The Australian Department of Education has introduced the “Australian Framework for Generative Artificial Intelligence (AI) in Schools,” a structured set of guidelines aimed at ensuring the ethical, safe, and effective use of AI in educational settings. This framework supports all individuals involved in school education, including school leaders, teachers, support staff, students, parents, service providers, and policymakers.
Teaching and learning
At the heart of the Australian framework is the idea that generative AI tools should enhance teaching and learning rather than replace the human element. The framework underscores that AI tools are used in ways that empower teachers, improve student outcomes, and streamline administrative processes.
It emphasises that while AI can assist in simplifying tasks, teachers remain the subject matter experts who guide the learning process. Students, in turn, are educated about the workings of AI, including its limitations and biases, ensuring they can engage critically with the technology. Schools are encouraged to integrate AI into their learning designs in a way that fosters creativity and critical thinking, rather than constraining these human faculties.
Additionally, students are taught to use AI ethically, maintaining academic integrity through clear guidelines on how AI tools should or shouldn’t be used in their assessments and work.
Human and social wellbeing
Another pillar of the framework is ensuring that AI’s integration into schools benefits all members of the community, without harming anyone’s wellbeing. AI tools must be used in ways that respect human dignity, protect individual rights, and promote diverse perspectives. The framework makes a strong case for inclusivity, highlighting that AI should expose students to a broad range of ideas rather than reinforcing biases or perpetuating discrimination. Furthermore, the wellbeing of students, teachers, and school staff is prioritised by ensuring AI tools are used in ways that are beneficial and not harmful. Human rights and individual autonomy are protected within the school environment, creating a space where AI complements human decision-making without undermining it.
Transparency
Transparency is a critical element of the Australian framework, ensuring that all school communities understand how AI tools work and how they are being used. This means that teachers, students, staff, and parents have access to clear and comprehensive information regarding AI, allowing them to make informed decisions. Schools are required to disclose when AI is in use, particularly in situations where it directly impacts students or teachers. Additionally, AI vendors must ensure that their products are explainable, meaning that users can understand the underlying methods and potential biases that may influence the output. This transparency builds trust and enables the school community to engage more meaningfully with AI.
Fairness
The principle of fairness is fundamental to the framework, ensuring that AI tools are used equitably. This includes making sure AI technologies are accessible to all students, regardless of disability, geographic location, or socio-economic background. Schools must actively work to prevent AI from being used in discriminatory ways, fostering inclusivity within the educational environment. The framework also respects cultural rights, ensuring that generative AI tools do not infringe upon Indigenous Cultural and Intellectual Property (ICIP) rights or the cultural rights of any community.
Accountability
The framework places human accountability at the forefront of AI use in schools. Even though AI tools assist in decision-making, teachers and school leaders retain ultimate responsibility for those decisions. This ensures that AI remains a support tool rather than a decision-maker. Furthermore, schools must regularly monitor the impact of AI tools, identifying both risks and opportunities as they arise. Community members are also encouraged to question and challenge AI’s role in decision-making, ensuring that AI’s influence is kept in check by human oversight.
Privacy, security, and safety
The final aspect of the framework focuses on protecting students’ privacy and ensuring the security of AI tools used within schools. AI tools must be compliant with Australian privacy laws, with strict limits on data collection and usage. Schools are required to inform students and staff about what data is being collected and how it will be used. Students are also taught to be cautious when in putting sensitive information into AI systems to protect their privacy. Cybersecurity is another priority, with schools expected to implement robust measures to protect the integrity of AI tools and associated data. In addition, schools must be mindful of copyright obligations when using AI, ensuring that they respect intellectual property rights.
What India can learn?
India, with its rapidly expanding digital infrastructure and focus on educational reform, can draw valuable lessons from Australia’s AI framework. By integrating similar principles, emphasizing transparency, accountability, and inclusivity, India can create a robust structure that ensures AI enhances education without undermining the human element.

The Conversation





Source link

Leave a Reply

Your email address will not be published. Required fields are marked *