Purpose:
The All-Party Parliamentary Group on Artificial Intelligence (APPG AI) was set up in January 2017, with the aim to explore the impact and implications of Artificial Intelligence.
All-Party Parliamentary Groups (APPGs) are informal all-party groups in UK Parliament. They are run by and for Members of the Commons and Lords. The APPG AI functions as the permanent authoritative voice within UK Parliament (House of Commons and House of Lords) on all AI-related matters.
Further Details:
We focus on Machine Learning, Facial Recognition, National Security, Cybersecurity, Digital & Physical Infrastructure, Data Policy & Governance, Market Intelligence, Education (skills, jobs & the future of work), Health (diagnostics, COVID, telehealth), AI in the Boardroom, Fin-tech Automation, Innovation & Entrepreneurship, Autonomous Weapons, Sustainability, Policy and Regulation Landscaping + much more.
We bring evidence, use cases and future policy scenarios into the UK Parliament while considering the economic, social, ethical and business model implications of developing and deploying AI. Our method of working is research, round tables, webinars, showcasing, taskforces, and events.
Keywords:
Associate
Citizen Participation in AI: Navigating Disinformation and Deep Fakes - Safeguarding Democratic Processes and Responsible AI Innovation
Citizen Participation in AI:
Navigating Disinformation and Deep Fakes - Safeguarding Democratic Processes and Responsible AI Innovation
On behalf of the APPG AI Chairs, Stephen Metcalfe MP and Lord Clement-Jones CBE, and the Secretariat at Big Innovation Centre, we would like to invite you to our Evidence Meeting: Citizen Participation in AI: Safeguarding Democratic Processes and Responsible AI Innovation - Navigating Disinformation and Deep Fakes
Questions to inspire:
- Navigating Deep Fake Advancements: Differentiating Genuine Information from Disinformation - How can the public differentiate between genuine information and disinformation, especially considering the rapid advancement of deep fake technology?
- Safeguarding Elections: Ensuring Responsible Use Amidst Deepfake Threats - Given the significant threats posed by the misuse of generative AI and deepfake technology in disseminating disinformation, how can we ensure responsible national elections?
- Protecting Democracy: Balancing Innovation with Defense Against Disinformation - What practical measures and policies can or should be implemented to strengthen democratic processes and protect societal well-being against the harmful effects of disinformation and deep fakes while still fostering innovation and technological advancement?
- Combating Disinformation: Enhancing Collaboration Across Governments, Tech Firms, Civil Society Organisations and Citizens - What roles do governments, tech firms, civil society organisations and citizens play in combating disinformation, and how can collaboration be improved?
Confirmed Speakers:
Sophie Murphy Byrne, Senior Manager, Government Affairs (EU&UK), Logically
Aled Lloyd Owen, Global Policy Director, Onfido
Professor Gina Neff, Executive Director, Minderoo Centre for Technology and Democracy, University of Cambridge
Markus Anderljung, Head of Policy, Centre for the Governance of AI
Format:
- There will be 4 to 8 formal Evidence Givers.
- There will also be an opportunity to share your opinion and raise practical questions.
- Press invited.
- This event will be filmed or recorded, and pictures will be taken.
Feedback on the Government Response to the White Paper Consultation
Meeting with Saqib Bhatti – Parliamentary Under-Secretary of State for Tech and the Digital Economy
Evidence to the Science and Technology Committee 2022
Enquiry into the Governance of Artificial Intelligence