Artwork

Content provided by Dan Bowen and Ray Fleming, Dan Bowen, and Ray Fleming. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Dan Bowen and Ray Fleming, Dan Bowen, and Ray Fleming or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://player.fm/legal.
Player FM - Podcast App
Go offline with the Player FM app!

News & Research Roundup 28 March

33:55
 
Share
 

Manage episode 409150082 series 3361952
Content provided by Dan Bowen and Ray Fleming, Dan Bowen, and Ray Fleming. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Dan Bowen and Ray Fleming, Dan Bowen, and Ray Fleming or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://player.fm/legal.

The season-ending episode for Series 7, this is the fifteenth in the series that started on 1st November last year with the "Regeneration: Human Centred Educational AI" episode. And it's an unbelievable 87th episode for the podcast (which started in September 2019).

When we come back with Series 8 after a short break for Easter, we're going to take a deeper dive into two specific use cases for AI in Education. The first we'll discuss is Assessment, where there's both a threat and opportunity created by AI. And the second topic is AI Tutors, where there's more of a focus on how we can take advantage of the technology to help improve support for learning for students.

This episode looks at one key news announcement - the EU AI Act - and a dozen new research papers on AI in education.

News

EU AI Act https://www.europarl.europa.eu/news/en/press-room/20240308IPR19015/artificial-intelligence-act-meps-adopt-landmark-law The European Parliament approved the AI Act on 13 March and there's some stuff in here that would make good practice guidance. And if you're developing AI solutions for education, and there's a chance that one of your customers or users might be in the EU, then you're going to need to follow these laws (just like GDPR is an EU law, but effectively applies globally if you're actively offering a service to EU residents). The Act bans some uses of AI that threaten citizen's rights - such as social scoring and biometric identification at mass level (things like untargeted facial scanning of CCTV or internet content, emotion recognition in the workplace or schools, and AI built to manipulate human behaviour) - and for the rest it relies on regulation according to categories. High Risk AI systems have to be assessed before being deployed and throughout their lifecycle. In the High Risk AI category it includes critical infrastructure (like transport and energy), product safety, law enforcement, justice and democratic processes, employment decision making - and Education. So decision making using AI in education needs to do full risk assessments, maintain usage logs, be transparent and accurate - and ensure human oversight. Examples of decision making that would be covered would be things like exam scoring, student recruitment screening, or behaviour management.

General generative AI - like chatgpt or co-pilots - will not be classified as high risk, but they'll still have obligations under the Act to do things like clear labelling for AI generated image, audio and video content ; make sure there's it can't generate illegal content, and also disclose what copyright data was used for training. But, although general AI may not be classified as high risk, if you then use that to build a high risk system - like an automated exam marker for end-of-school exams, then this will be covered under the high risk category.

All of this is likely to become law by the middle of the year, and by the end of 2024 prohibited AI systems will be banned - and by mid-2025 the rules will start applying for other AI systems.

ResearchAnother huge month. I spent the weekend reviewing a list of 350 new papers published in the first two weeks of March, on Large Language Models, ChatGPT etc, to find the ones that are really interesting for the podcast

Adapting Large Language Models for Education: Foundational Capabilities, Potentials, and Challenges

arXiv:2401.08664

A Study on Large Language Models' Limitations in Multiple-Choice Question Answering

arXiv:2401.07955

Dissecting Bias of ChatGPT in College Major Recommendations

arXiv:2401.11699

Evaluating Large Language Models in Analysing Classroom Dialogue

arXiv:2402.02380

The Future of AI in Education: 13 Things We Can Do to Minimize the Damage

https://osf.io/preprints/edarxiv/372vr

Scaling the Authoring of AutoTutors with Large Language Models

https://arxiv.org/abs/2402.09216

Role-Playing Simulation Games using ChatGPT

https://arxiv.org/abs/2402.09161

Economic and Financial Learning with Artificial Intelligence: A Mixed-Methods Study on ChatGPT

https://arxiv.org/abs/2402.15278

A Study on the Vulnerability of Test Questions against ChatGPT-based Cheating

https://arxiv.org/abs/2402.14881

Incorporating Artificial Intelligence Into Athletic Training Education: Developing Case-Based Scenarios Using ChatGPT

https://meridian.allenpress.com/atej/article/19/1/42/498456

Incorporating Artificial Intelligence Into Athletic Training Education: Developing Case-Based Scenarios Using ChatGPT

https://meridian.allenpress.com/atej/article/19/1/42/498456

RECIPE4U: Student-ChatGPT Interaction Dataset in EFL Writing Education

https://arxiv.org/abs/2403.08272

Comparison of the problem-solving performance of ChatGPT-3.5, ChatGPT-4, Bing Chat, and Bard for the Korean emergency medicine board examination question bank

https://journals.lww.com/md-journal/fulltext/2024/03010/comparison_of_the_problem_solving_performance_of.48.aspx?context=latestarticles

Comparing the quality of human and ChatGPT feedback of students’ writing

https://www.sciencedirect.com/science/article/pii/S0959475224000215

  continue reading

99 episodes

Artwork
iconShare
 
Manage episode 409150082 series 3361952
Content provided by Dan Bowen and Ray Fleming, Dan Bowen, and Ray Fleming. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Dan Bowen and Ray Fleming, Dan Bowen, and Ray Fleming or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://player.fm/legal.

The season-ending episode for Series 7, this is the fifteenth in the series that started on 1st November last year with the "Regeneration: Human Centred Educational AI" episode. And it's an unbelievable 87th episode for the podcast (which started in September 2019).

When we come back with Series 8 after a short break for Easter, we're going to take a deeper dive into two specific use cases for AI in Education. The first we'll discuss is Assessment, where there's both a threat and opportunity created by AI. And the second topic is AI Tutors, where there's more of a focus on how we can take advantage of the technology to help improve support for learning for students.

This episode looks at one key news announcement - the EU AI Act - and a dozen new research papers on AI in education.

News

EU AI Act https://www.europarl.europa.eu/news/en/press-room/20240308IPR19015/artificial-intelligence-act-meps-adopt-landmark-law The European Parliament approved the AI Act on 13 March and there's some stuff in here that would make good practice guidance. And if you're developing AI solutions for education, and there's a chance that one of your customers or users might be in the EU, then you're going to need to follow these laws (just like GDPR is an EU law, but effectively applies globally if you're actively offering a service to EU residents). The Act bans some uses of AI that threaten citizen's rights - such as social scoring and biometric identification at mass level (things like untargeted facial scanning of CCTV or internet content, emotion recognition in the workplace or schools, and AI built to manipulate human behaviour) - and for the rest it relies on regulation according to categories. High Risk AI systems have to be assessed before being deployed and throughout their lifecycle. In the High Risk AI category it includes critical infrastructure (like transport and energy), product safety, law enforcement, justice and democratic processes, employment decision making - and Education. So decision making using AI in education needs to do full risk assessments, maintain usage logs, be transparent and accurate - and ensure human oversight. Examples of decision making that would be covered would be things like exam scoring, student recruitment screening, or behaviour management.

General generative AI - like chatgpt or co-pilots - will not be classified as high risk, but they'll still have obligations under the Act to do things like clear labelling for AI generated image, audio and video content ; make sure there's it can't generate illegal content, and also disclose what copyright data was used for training. But, although general AI may not be classified as high risk, if you then use that to build a high risk system - like an automated exam marker for end-of-school exams, then this will be covered under the high risk category.

All of this is likely to become law by the middle of the year, and by the end of 2024 prohibited AI systems will be banned - and by mid-2025 the rules will start applying for other AI systems.

ResearchAnother huge month. I spent the weekend reviewing a list of 350 new papers published in the first two weeks of March, on Large Language Models, ChatGPT etc, to find the ones that are really interesting for the podcast

Adapting Large Language Models for Education: Foundational Capabilities, Potentials, and Challenges

arXiv:2401.08664

A Study on Large Language Models' Limitations in Multiple-Choice Question Answering

arXiv:2401.07955

Dissecting Bias of ChatGPT in College Major Recommendations

arXiv:2401.11699

Evaluating Large Language Models in Analysing Classroom Dialogue

arXiv:2402.02380

The Future of AI in Education: 13 Things We Can Do to Minimize the Damage

https://osf.io/preprints/edarxiv/372vr

Scaling the Authoring of AutoTutors with Large Language Models

https://arxiv.org/abs/2402.09216

Role-Playing Simulation Games using ChatGPT

https://arxiv.org/abs/2402.09161

Economic and Financial Learning with Artificial Intelligence: A Mixed-Methods Study on ChatGPT

https://arxiv.org/abs/2402.15278

A Study on the Vulnerability of Test Questions against ChatGPT-based Cheating

https://arxiv.org/abs/2402.14881

Incorporating Artificial Intelligence Into Athletic Training Education: Developing Case-Based Scenarios Using ChatGPT

https://meridian.allenpress.com/atej/article/19/1/42/498456

Incorporating Artificial Intelligence Into Athletic Training Education: Developing Case-Based Scenarios Using ChatGPT

https://meridian.allenpress.com/atej/article/19/1/42/498456

RECIPE4U: Student-ChatGPT Interaction Dataset in EFL Writing Education

https://arxiv.org/abs/2403.08272

Comparison of the problem-solving performance of ChatGPT-3.5, ChatGPT-4, Bing Chat, and Bard for the Korean emergency medicine board examination question bank

https://journals.lww.com/md-journal/fulltext/2024/03010/comparison_of_the_problem_solving_performance_of.48.aspx?context=latestarticles

Comparing the quality of human and ChatGPT feedback of students’ writing

https://www.sciencedirect.com/science/article/pii/S0959475224000215

  continue reading

99 episodes

All episodes

×
 
Loading …

Welcome to Player FM!

Player FM is scanning the web for high-quality podcasts for you to enjoy right now. It's the best podcast app and works on Android, iPhone, and the web. Signup to sync subscriptions across devices.

 

Quick Reference Guide