Regulation of artificial intelligence

From Wikipedia, the free encyclopedia

The regulation of artificial intelligence is the development of public sector policies and laws for promoting and regulating artificial intelligence (AI); it is therefore related to the broader regulation of algorithms. The regulatory and policy landscape for AI is an emerging issue in jurisdictions globally, including in the European Union and supra-national bodies like IEEE, OECD and others. Between 2016-2019 a wave of AI Ethics Guidelines were published in order to maintain social control over the technology.[1] Regulation is considered necessary to both encourage AI and manage associated risks. Regulation of AI through mechanisms such as review boards can also be seen as social means to approach the AI control problem.[2]

Background[]

In 2017 Elon Musk called for regulation of AI development.[3] According to NPR, the Tesla CEO was "clearly not thrilled" to be advocating for government scrutiny that could impact his own industry, but believed the risks of going completely without oversight are too high: "Normally the way regulations are set up is when a bunch of bad things happen, there's a public outcry, and after many years a regulatory agency is set up to regulate that industry. It takes forever. That, in the past, has been bad but not something which represented a fundamental risk to the existence of civilization."[3] In response, some politicians expressed skepticism about the wisdom of regulating a technology that is still in development.[4] Responding both to Musk and to February 2017 proposals by European Union lawmakers to regulate AI and robotics, Intel CEO Brian Krzanich has argued that AI is in its infancy and that it is too early to regulate the technology.[5] Instead of trying to regulate the technology itself, some scholars suggested developing common norms including requirements for the testing and transparency of algorithms, possibly in combination with some form of warranty.[6]

Perspectives[]

The regulation of artificial intelligence is the development of public sector policies and laws for promoting and regulating AI.[7] Regulation is now generally considered necessary to both encourage AI and manage associated risks.[8][9][10] Public administration and policy considerations generally focus on the technical and economic implications and on trustworthy and human-centered AI systems,[11] although regulation of artificial superintelligences is also considered.[12] The basic approach to regulation focuses on the risks and biases of AI's underlying technology, i.e., machine-learning algorithms, at the level of the input data, algorithm testing, and the decision model, as well as whether explanations of biases in the code can be understandable for prospective recipients of the technology, and technically feasible for producers to convey.[13]

AI regulation could derive from basic principles. A 2020 Berkman Klein Center for Internet & Society meta-review of existing sets of principles, such as the Asilomar Principles and the Beijing Principles, identified eights such basic principles: privacy, accountability, safety and security, transparency and explainability, fairness and non-discrimination, human control of technology, professional responsibility, and respect for human values.[14] AI law and regulations have been divided into three main topics, namely governance of autonomous intelligence systems, responsibility and accountability for the systems, and privacy and safety issues.[8] A public administration approach sees a relationship between AI law and regulation, the ethics of AI, and 'AI society', defined as workforce substitution and transformation, social acceptance and trust in AI, and the transformation of human to machine interaction.[15] The development of public sector strategies for management and regulation of AI is deemed necessary at the local, national,[16] and international levels[17] and in a variety of fields, from public service management[18] and accountability[19] to law enforcement,[17][20] healthcare (especially the concept of a Human Guarantee),[21][22][23][24][25] the financial sector,[16] robotics,[26][27] autonomous vehicles,[26] the military[28] and national security,[29] and international law.[30][31]

As a response to the AI control problem[]

Regulation of AI can be seen as positive social means to manage the AI control problem, i.e., the need to insure long-term beneficial AI, with other social responses such as doing nothing or banning being seen as impractical, and approaches such as enhancing human capabilities through transhumanism approaches such as brain-computer interfaces being seen as potentially complementary.[32][33] Regulation of research into artificial general intelligence (AGI) focuses on the role of review boards, from university or corporation to international levels, and on encouraging research into safe AI,[33] together with the possibility of differential intellectual progress (prioritizing risk-reducing strategies over risk-taking strategies in AI development) or conducting international mass surveillance to perform AGI arms control.[32] For instance, the 'AGI Nanny' is a proposed strategy, potentially under the control of humanity, for preventing the creation of a dangerous superintelligence as well as addressing other major threats to human well-being, such as subversion of the global financial system, until a superintelligence can be safely created. It entails the creation of a smarter-than-human, but not superintelligent, artificial general intelligence system connected to a large surveillance network, with the goal of monitoring humanity and protecting it from danger."[32] Regulation of conscious, ethically aware AGIs focuses on integrating them with existing human society and can be divided into considerations of their legal standing and of their moral rights.[32] Regulation of AI has been seen as restrictive, with a risk of preventing the development of AGI.[26]

Global guidance[]

The development of a global governance board to regulate AI development was suggested at least as early as 2017.[34] In December 2018, Canada and France announced plans for a G7-backed International Panel on Artificial Intelligence, modeled on the International Panel on Climate Change, to study the global effects of AI on people and economies and to steer AI development.[35] In 2019 the Panel was renamed the Global Partnership on AI, but it is yet to be endorsed by the United States.[36][37]

The OECD Recommendations on AI[38] were adopted in May 2019, and the G20 AI Principles in June 2019.[37][39][40] In September 2019 the World Economic Forum issued ten 'AI Government Procurement Guidelines'.[41] In February 2020, the European Union published its draft strategy paper for promoting and regulating AI.[17]

At the United Nations, several entities have begun to promote and discuss aspects of AI regulation and policy, including the UNICRI Centre for AI and Robotics.[29] At UNESCO’s Scientific 40th session in November 2019, the organization commenced a two year process to achieve a "global standard-setting instrument on ethics of artificial intelligence". In pursuit of this goal, UNESCO forums and conferences on AI have taken place to gather stakeholder views. The most recent draft text of a recommendation on the ethics of AI of the UNESCO Ad Hoc Expert Group was issued in September 2020 and includes a call for legislative gaps to be filled.[42]

Regional and national regulation[]

Timeline of strategies, action plans and policy papers setting defining national, regional and international approaches to AI[43]

The regulatory and policy landscape for AI is an emerging issue in jurisdictions globally, including in the European Union[44] and Russia.[45] Since early 2016, many national, regional and international authorities have begun adopting strategies, actions plans and policy papers on AI.[46][47] These documents cover a wide range of topics such as regulation and governance, as well as industrial strategy, research, talent and infrastructure.[11][48]

China[]

The regulation of AI in China is mainly governed by the State Council of the PRC's July 8, 2017 "A Next Generation Artificial Intelligence Development Plan" (State Council Document No. 35), in which the Central Committee of the Communist Party of China and the State Council of the People's Republic of China urged the governing bodies of China to promote the development of AI. Regulation of the issues of ethical and legal support for the development of AI is nascent, but policy ensures state control of Chinese companies and over valuable data, including storage of data on Chinese users within the country and the mandatory use of People's Republic of China's national standards for AI, including over big data, cloud computing, and industrial software.[citation needed]

Council of Europe[]

The Council of Europe (CoE) is an international organization which promotes human rights democracy and the rule of law and comprises 47 member states, including all 29 Signatories of the European Union's 2018 Declaration of Cooperation on Artificial Intelligence. The CoE has created a common legal space in which the members have a legal obligation to guarantee rights as set out in the European Convention on Human Rights. Specifically in relation to AI, "The Council of Europe’s aim is to identify intersecting areas between AI and our standards on human rights, democracy and rule of law, and to develop relevant standard setting or capacity-building solutions". The large number of relevant documents identified by the CoE include guidelines, charters, papers, reports and strategies.[49] The authoring bodies of these AI regulation documents are not confined to one sector of society and include organizations, companies, bodies and nation-states.[42]

European Union[]

Most European Union (EU) countries have their own national strategies towards regulating AI, but these are largely convergent.[42] The European Union is guided by a European Strategy on Artificial Intelligence,[50] supported by a High-Level Expert Group on Artificial Intelligence.[51][52] In April 2019, the European Commission published its Ethics Guidelines for Trustworthy Artificial Intelligence (AI),[53][54] following this with its Policy and investment recommendations for trustworthy Artificial Intelligence in June 2019.[55] The EU Commission’s High Level Expert Group on Artificial Intelligence carries out work on Trustworthy AI, and the Commission has issued reports on the Safety and Liability Aspects of AI and on the Ethics of Automated Vehicles. In 2020 the EU Commission sought views on a proposal for AI specific legislation, and that process is ongoing.[42]

On February 2, 2020, the European Commission published its White Paper on Artificial Intelligence - A European approach to excellence and trust.[56] The White Paper consists of two main building blocks, an ‘ecosystem of excellence’ and a ‘ecosystem of trust’. The latter outlines the EU's approach for a regulatory framework for AI. In its proposed approach, the Commission differentiates between 'high-risk' and 'non-high-risk' AI applications. Only the former should be in the scope of a future EU regulatory framework. Whether this would be the case could in principle be determined by two cumulative criteria, concerning critical sectors and critical use. Following key requirements are considered for high-risk AI applications: requirements for training data; data and record-keeping; informational duties; requirements for robustness and accuracy; human oversight; and specific requirements for specific AI applications, such as those used for purposes of remote biometric identification. AI applications that do not qualify as ‘high-risk’ could be governed by a voluntary labeling scheme. As regards compliance and enforcement, the Commission considers prior conformity assessments which could include 'procedures for testing, inspection or certification' and/or 'checks of the algorithms and of the data sets used in the development phase'. A European governance structure on AI in the form of a framework for cooperation of national competent authorities could facilitate the implementation of the regulatory framework.[56] A January 2021 draft was leaked online on April 14, 2021,[57] before the Commission ultimately presented their official "Proposal for a Regulation laying down harmonised rules on artificial intelligence (Artificial Intelligence Act)" a week later.[58] Shortly after, the Artificial Intelligence Act was formally proposed. This includes a fine-tune of the 2020 risk-based approach.

United Kingdom[]

The UK supported the application and development of AI in business via the Digital Economy Strategy 2015-2018, introduced at the beginning of 2015 by Innovate UK as part of the UK Digital Strategy.[citation needed] In the public sector, guidance has been provided by the Department for Digital, Culture, Media and Sport, on data ethics[59] and the Alan Turing Institute, on responsible design and implementation of AI systems.[60] In terms of cyber security, the National Cyber Security Centre has issued guidance on ‘Intelligent Security Tools’.[29][61]

United States[]

Discussions on regulation of AI in the United States have included topics such as the timeliness of regulating AI, the nature of the federal regulatory framework to govern and promote AI, including what agency should lead, the regulatory and governing powers of that agency, and how to update regulations in the face of rapidly changing technology, as well as the roles of state governments and courts.[62]

As early as 2016, the Obama administration had begun to focus on the risks and regulations for artificial intelligence. In a report titled Preparing For the Future of Artificial Intelligence, the National Science and Technology Council set a precedent to allow researchers to continue to develop new AI technologies with few restrictions. It is stated within the report that "the approach to regulation of AI-enabled products to protect public safety should be informed by assessment of the aspects of risk....".[63] These risks would be the principal reason to create any form of regulation, granted that any existing regulation would not apply to AI technology.

The first main report was the National Strategic Research and Development Plan for Artificial Intelligence. On August 13, 2018, Section 1051 of the Fiscal Year 2019 John S. McCain National Defense Authorization Act (P.L. 115-232) established the National Security Commission on Artificial Intelligence "to consider the methods and means necessary to advance the development of artificial intelligence, machine learning, and associated technologies to comprehensively address the national security and defense needs of the United States."[64] Steering on regulating security-related AI is provided by the National Security Commission on Artificial Intelligence.[65] The Artificial Intelligence Initiative Act (S.1558) is a proposed bill that would establish a federal initiative designed to accelerate research and development on AI for, inter alia, the economic and national security of the United States.[66][67]

On January 7, 2019, following an Executive Order on Maintaining American Leadership in Artificial Intelligence, the White House’s Office of Science and Technology Policy released a draft Guidance for Regulation of Artificial Intelligence Applications, which includes ten principles for United States agencies when deciding whether and how to regulate AI.[68] In response, the National Institute of Standards and Technology has released a position paper,[69] the National Security Commission on Artificial Intelligence has published a report,[70] and the Defense Innovation Board has issued recommendations on the ethical use of AI.[71] A year later, the administration called for comments on regulation in another draft of its Guidance for Regulation of Artificial Intelligence Applications.[72]

Other specific agencies working on the regulation of AI include the Food and Drug Administration, which has created pathways to regulate the incorporation of AI in medical imaging.[22]

Regulation of fully autonomous weapons[]

Legal questions related to lethal autonomous weapons systems (LAWS), in particular compliance with the laws of armed conflict, have been under discussion at the United Nations since 2013, within the context of the Convention on Certain Conventional Weapons.[73] Notably, informal meetings of experts took place in 2014, 2015 and 2016 and a Group of Governmental Experts (GGE) was appointed to further deliberate on the issue in 2016. A set of guiding principles on LAWS affirmed by the GGE on LAWS were adopted in 2018.[74]

In 2016, China published a position paper questioning the adequacy of existing international law to address the eventuality of fully autonomous weapons, becoming the first permanent member of the U.N. Security Council to broach the issue,[30] and leading to proposals for global regulation.[75] The possibility of a moratorium or preemptive ban of the development and use of LAWS has also been raised on several occasions by other national delegations to the Convention on Certain Conventional Weapons and is strongly advocated for by the Campaign to Stop Killer Robots - a coalition of non-governmental organizations.[76] The US government maintains that current international humanitarian law is capable of regulating the development or use of LAWS, and the US military does not have LAWS in its inventory.[77]

See also[]

References[]

  1. ^ Héder, M (2020). "A criticism of AI ethics guidelines". Információs Társadalom. 20 (4): 57–73. doi:10.22503/inftars.XX.2020.4.5.
  2. ^ "An Ethical Approach to AI is an Absolute Imperative, Andreas Kaplan". Retrieved 26 April 2021.
  3. ^ Jump up to: a b "Elon Musk Warns Governors: Artificial Intelligence Poses 'Existential Risk'". NPR.org. Retrieved 27 November 2017.
  4. ^ Gibbs, Samuel (17 July 2017). "Elon Musk: regulate AI to combat 'existential threat' before it's too late". The Guardian. Retrieved 27 November 2017.
  5. ^ Kharpal, Arjun (7 November 2017). "A.I. is in its 'infancy' and it's too early to regulate it, Intel CEO Brian Krzanich says". CNBC. Retrieved 27 November 2017.
  6. ^ Kaplan, Andreas; Haenlein, Michael (2019). "Siri, Siri, in my hand: Who's the fairest in the land? On the interpretations, illustrations, and implications of artificial intelligence". Business Horizons. 62: 15–25. doi:10.1016/j.bushor.2018.08.004.
  7. ^ Research handbook on the law of artificial intelligence. Barfield, Woodrow,, Pagallo, Ugo. Cheltenham, UK. 2018. ISBN 978-1-78643-904-8. OCLC 1039480085.CS1 maint: others (link)
  8. ^ Jump up to: a b Wirtz, Bernd W.; Weyerer, Jan C.; Geyer, Carolin (2018-07-24). "Artificial Intelligence and the Public Sector—Applications and Challenges". International Journal of Public Administration. 42 (7): 596–615. doi:10.1080/01900692.2018.1498103. ISSN 0190-0692. S2CID 158829602.
  9. ^ Buiten, Miriam C (2019). "Towards Intelligent Regulation of Artificial Intelligence". European Journal of Risk Regulation. 10 (1): 41–59. doi:10.1017/err.2019.8. ISSN 1867-299X.
  10. ^ Mantelero, Alessandro; Esposito, Maria Samantha (2021). "An evidence-based methodology for human rights impact assessment (HRIA) in the development of AI data-intensive systems". Computer Law & Security Review. 41: 105561. doi:10.1016/j.clsr.2021.105561.
  11. ^ Jump up to: a b Artificial intelligence in society. Organisation for Economic Co-operation and Development. Paris. 11 June 2019. ISBN 978-92-64-54519-9. OCLC 1105926611.CS1 maint: others (link)
  12. ^ Kamyshansky, Vladimir P.; Rudenko, Evgenia Y.; Kolomiets, Evgeniy A.; Kripakova, Dina R. (2020), "Revisiting the Place of Artificial Intelligence in Society and the State", Artificial Intelligence: Anthropogenic Nature vs. Social Origin, Cham: Springer International Publishing, pp. 359–364, doi:10.1007/978-3-030-39319-9_41, ISBN 978-3-030-39318-2, retrieved 2021-03-27
  13. ^ BUITEN, Miriam C (2019). "Towards Intelligent Regulation of Artificial Intelligence". European Journal of Risk Regulation. 10 (1): 41–59. doi:10.1017/err.2019.8. ISSN 1867-299X. S2CID 155138874.
  14. ^ Fjeld, Jessica; Achten, Nele; Hilligoss, Hannah; Nagy, Adam; Srikumar, Madhu (2020-01-15). "Principled Artificial Intelligence: Mapping Consensus in Ethical and Rights-based Approaches to Principles for AI". Cite journal requires |journal= (help)
  15. ^ Wirtz, Bernd W.; Weyerer, Jan C.; Sturm, Benjamin J. (2020-04-15). "The Dark Sides of Artificial Intelligence: An Integrated AI Governance Framework for Public Administration". International Journal of Public Administration. 43 (9): 818–829. doi:10.1080/01900692.2020.1749851. ISSN 0190-0692. S2CID 218807452.
  16. ^ Jump up to: a b Bredt, Stephan (2019-10-04). "Artificial Intelligence (AI) in the Financial Sector—Potential and Public Strategies". Frontiers in Artificial Intelligence. 2: 16. doi:10.3389/frai.2019.00016. ISSN 2624-8212. PMC 7861258. PMID 33733105.
  17. ^ Jump up to: a b c White Paper: On Artificial Intelligence – A European approach to excellence and trust (PDF). Brussels: European Commission. 2020. p. 1.
  18. ^ Wirtz, Bernd W.; Müller, Wilhelm M. (2018-12-03). "An integrated artificial intelligence framework for public management". Public Management Review. 21 (7): 1076–1100. doi:10.1080/14719037.2018.1549268. ISSN 1471-9037. S2CID 158267709.
  19. ^ Reisman, Dillon; Schultz, Jason; Crawford, Kate; Whittaker, Meredith (2018). Algorithmic impact assessments: A practical framework for public agency accountability (PDF). New York: AI Now Institute.
  20. ^ "Towards Responsible Artificial Intelligence Innovation" (PDF). UNICRI. 2020. Retrieved 2020-08-09.
  21. ^ Gruson, David (2019). "Éthique et intelligence artificielle en santé, pour une régulation positive". Soins (in French). 64 (832): 54–57. doi:10.1016/j.soin.2018.12.015. PMID 30771853.
  22. ^ Jump up to: a b Kohli, Ajay; Mahajan, Vidur; Seals, Kevin; Kohli, Ajit; Jha, Saurabh (2019). "Concepts in U.S. Food and Drug Administration Regulation of Artificial Intelligence for Medical Imaging". American Journal of Roentgenology. 213 (4): 886–888. doi:10.2214/ajr.18.20410. ISSN 0361-803X. PMID 31166758.
  23. ^ Hwang, Thomas J.; Kesselheim, Aaron S.; Vokinger, Kerstin N. (2019-12-17). "Lifecycle Regulation of Artificial Intelligence– and Machine Learning–Based Software Devices in Medicine". JAMA. 322 (23): 2285. doi:10.1001/jama.2019.16842. ISSN 0098-7484. PMID 31755907.
  24. ^ Sharma, Kavita; Manchikanti, Padmavati (2020-10-01). "Regulation of Artificial Intelligence in Drug Discovery and Health Care". Biotechnology Law Report. 39 (5): 371–380. doi:10.1089/blr.2020.29183.ks. ISSN 0730-031X.
  25. ^ Petkus, Haroldas; Hoogewerf, Jan; Wyatt, Jeremy C (2020). "What do senior physicians think about AI and clinical decision support systems: Quantitative and qualitative analysis of data from specialty societies". Clinical Medicine. 20 (3): 324–328. doi:10.7861/clinmed.2019-0317. ISSN 1470-2118. PMC 7354034. PMID 32414724.
  26. ^ Jump up to: a b c Gurkaynak, Gonenc; Yilmaz, Ilay; Haksever, Gunes (2016). "Stifling artificial intelligence: Human perils". Computer Law & Security Review. 32 (5): 749–758. doi:10.1016/j.clsr.2016.05.003. ISSN 0267-3649.
  27. ^ Iphofen, Ron; Kritikos, Mihalis (2019-01-03). "Regulating artificial intelligence and robotics: ethics by design in a digital society". Contemporary Social Science. 16 (2): 170–184. doi:10.1080/21582041.2018.1563803. ISSN 2158-2041. S2CID 59298502.
  28. ^ AI principles: Recommendations on the ethical use of artificial intelligence by the Department of Defense (PDF). Washington, DC: United States Defense Innovation Board. 2019. OCLC 1126650738.
  29. ^ Jump up to: a b c Babuta, Alexander; Oswald, Marion; Janjeva, Ardi (2020). Artificial Intelligence and UK National Security: Policy Considerations (PDF). London: Royal United Services Institute.
  30. ^ Jump up to: a b "Robots with Guns: The Rise of Autonomous Weapons Systems". Snopes.com. 21 April 2017. Retrieved 24 December 2017.
  31. ^ Bento, Lucas (2017). "No Mere Deodands: Human Responsibilities in the Use of Violent Intelligent Systems Under Public International Law". Harvard Scholarship Depository. Retrieved 2019-09-14.
  32. ^ Jump up to: a b c d Sotala, Kaj; Yampolskiy, Roman V (2014-12-19). "Responses to catastrophic AGI risk: a survey". Physica Scripta. 90 (1): 018001. doi:10.1088/0031-8949/90/1/018001. ISSN 0031-8949.
  33. ^ Jump up to: a b Barrett, Anthony M.; Baum, Seth D. (2016-05-23). "A model of pathways to artificial superintelligence catastrophe for risk and decision analysis". Journal of Experimental & Theoretical Artificial Intelligence. 29 (2): 397–414. arXiv:1607.07730. doi:10.1080/0952813x.2016.1186228. ISSN 0952-813X. S2CID 928824.
  34. ^ Boyd, Matthew; Wilson, Nick (2017-11-01). "Rapid developments in Artificial Intelligence: how might the New Zealand government respond?". Policy Quarterly. 13 (4). doi:10.26686/pq.v13i4.4619. ISSN 2324-1101.
  35. ^ Innovation, Science and Economic Development Canada (2019-05-16). "Declaration of the International Panel on Artificial Intelligence". gcnws. Retrieved 2020-03-29.
  36. ^ "The world has a plan to rein in AI—but the US doesn't like it". Wired. 2020-01-08. Retrieved 2020-03-29.
  37. ^ Jump up to: a b "AI Regulation: Has the Time Arrived?". InformationWeek. Retrieved 2020-03-29.
  38. ^ "OECD Principles on Artificial Intelligence - Organisation for Economic Co-operation and Development". www.oecd.org. Retrieved 2020-03-29.
  39. ^ G20 Ministerial Statement on Trade and Digital Economy (PDF). Tsukuba City, Japan: G20. 2019.
  40. ^ "International AI ethics panel must be independent". Nature. 572 (7770): 415. 2019-08-21. Bibcode:2019Natur.572R.415.. doi:10.1038/d41586-019-02491-x. PMID 31435065.
  41. ^ Guidelines for AI Procurement (PDF). Cologny/Geneva: World Economic Forum. 2019.
  42. ^ Jump up to: a b c d NíFhaoláin, Labhaoise; Hines, Andrew; Nallur, Vivek (2020). Assessing the Appetite for Trustworthiness and the Regulation of Artificial Intelligence in Europe (PDF). Dublin: Technological University Dublin, School of Computer Science, Dublin. pp. 1–12. This article incorporates text available under the CC BY 4.0 license. (The CC BY 4.0 licence means that everyone have the right to reuse the text that is quoted here, or other parts of the original article itself, if they credit the authors. More info: Creative Commons license) Changes were made as follows: citations removed and minor grammatical amendments.
  43. ^ "UNICRI :: United Nations Interregional Crime and Justice Research Institute". www.unicri.it. Retrieved 2020-08-08.
  44. ^ Law Library of Congress (U.S.). Global Legal Research Directorate, issuing body. Regulation of artificial intelligence in selected jurisdictions. LCCN 2019668143. OCLC 1110727808.
  45. ^ Popova, Anna V.; Gorokhova, Svetlana S.; Abramova, Marianna G.; Balashkina, Irina V. (2021), "The System of Law and Artificial Intelligence in Modern Russia: Goals and Instruments of Digital Modernization", Studies in Systems, Decision and Control, Cham: Springer International Publishing, pp. 89–96, doi:10.1007/978-3-030-56433-9_11, ISBN 978-3-030-56432-2, retrieved 2021-03-27
  46. ^ "OECD Observatory of Public Sector Innovation - Ai Strategies and Public Sector Components". Retrieved 2020-05-04.
  47. ^ Berryhill, Jamie; Heang, Kévin Kok; Clogher, Rob; McBride, Keegan (2019). Hello, World: Artificial Intelligence and its Use in the Public Sector (PDF). Paris: OECD Observatory of Public Sector Innovation.
  48. ^ Campbell, Thomas A. (2019). Artificial Intelligence: An Overview of State Initiatives (PDF). Evergreen, CO: FutureGrasp, LLC.
  49. ^ "Council of Europe and Artificial Intelligence". Artificial Intelligence. Retrieved 2021-07-29.
  50. ^ Anonymous (2018-04-25). "Communication Artificial Intelligence for Europe". Shaping Europe’s digital future - European Commission. Retrieved 2020-05-05.
  51. ^ smuhana (2018-06-14). "High-Level Expert Group on Artificial Intelligence". Shaping Europe’s digital future - European Commission. Retrieved 2020-05-05.
  52. ^ Andraško, Jozef; Mesarčík, Matúš; Hamuľák, Ondrej (2021-01-02). "The regulatory intersections between artificial intelligence, data protection and cyber security: challenges and opportunities for the EU legal framework". AI & Society. doi:10.1007/s00146-020-01125-5. ISSN 0951-5666. S2CID 230109912.
  53. ^ "Ethics Guidelines For Trustworthy AI" (PDF).
  54. ^ Weiser, Stephanie (2019-04-03). "Building trust in human-centric AI". FUTURIUM - European Commission. Retrieved 2020-05-05.
  55. ^ Anonymous (2019-06-26). "Policy and investment recommendations for trustworthy Artificial Intelligence". Shaping Europe’s digital future - European Commission. Retrieved 2020-05-05.
  56. ^ Jump up to: a b European Commission. (2020). White paper on artificial intelligence: a European approach to excellence and trust. OCLC 1141850140.
  57. ^ Heikkilä, Melissa (2021-04-14). "POLITICO AI: Decoded: The EU’s AI rules — Finland talks to machines — Facebook’s fairness project" (newsletter). POLITICO. Retrieved 2021-05-14.
  58. ^ European Commission (2021-04-21). Europe fit for the Digital Age: Commission proposes new rules and actions for excellence and trust in Artificial Intelligence (press release). Retrieved 2021-05-14.
  59. ^ Data Ethics Framework (PDF). London: Department for Digital, Culture, Media and Sport. 2018.
  60. ^ Leslie, David (2019-06-11). "Understanding artificial intelligence ethics and safety: A guide for the responsible design and implementation of AI systems in the public sector". doi:10.5281/zenodo.3240529. S2CID 189762499. Cite journal requires |journal= (help)
  61. ^ "Intelligent security tools". www.ncsc.gov.uk. Retrieved 2020-04-28.
  62. ^ Weaver, John Frank (2018-12-28). "Regulation of artificial intelligence in the United States". Research Handbook on the Law of Artificial Intelligence: 155–212. doi:10.4337/9781786439055.00018. ISBN 9781786439055.
  63. ^ National Science and Technology Council Committee on Technology (October 2016). "Preparing for the Future of Artificial Intelligence". whitehouse.gov – via National Archives.
  64. ^ "About". National Security Commission on Artificial Intelligence. Retrieved 2020-06-29.
  65. ^ Stefanik, Elise M. (2018-05-22). "H.R.5356 – 115th Congress (2017–2018): National Security Commission Artificial Intelligence Act of 2018". www.congress.gov. Retrieved 2020-03-13.
  66. ^ Heinrich, Martin (2019-05-21). "Text - S.1558 - 116th Congress (2019-2020): Artificial Intelligence Initiative Act". www.congress.gov. Retrieved 2020-03-29.
  67. ^ Scherer, Matthew U. (2015). "Regulating Artificial Intelligence Systems: Risks, Challenges, Competencies, and Strategies". SSRN Working Paper Series. doi:10.2139/ssrn.2609777. ISSN 1556-5068.
  68. ^ "AI Update: White House Issues 10 Principles for Artificial Intelligence Regulation". Inside Tech Media. 2020-01-14. Retrieved 2020-03-25.
  69. ^ U.S. Leadership in AI: A Plan for Federal Engagement in Developing Technical Standards and Related Tools (PDF). National Institute of Science and Technology. 2019.
  70. ^ NSCAI Final Report (PDF). Washington, DC: The National Security Commission on Artificial Intelligence. 2021.
  71. ^ AI Principles: Recommendations on the Ethical Use of Artificial Intelligence by the Department of Defense (PDF). Washington, DC: Defense Innovation Board. 2020.
  72. ^ "Request for Comments on a Draft Memorandum to the Heads of Executive Departments and Agencies, "Guidance for Regulation of Artificial Intelligence Applications"". Federal Register. 2020-01-13. Retrieved 2020-11-28.
  73. ^ "Background on Lethal Autonomous Weapons Systems in the CCW". United Nations Geneva.
  74. ^ "Guiding Principles affirmed by the Group of Governmental Experts on Emerging Technologies in the Area of Lethal Autonomous Weapons System" (PDF). United Nations Geneva.
  75. ^ Baum, Seth (2018-09-30). "Countering Superintelligence Misinformation". Information. 9 (10): 244. doi:10.3390/info9100244. ISSN 2078-2489.
  76. ^ "Country Views on Killer Robots" (PDF). The Campaign to Stop Killer Robots.
  77. ^ Sayler, Kelley (2020). Artificial Intelligence and National Security: Updated November 10, 2020 (PDF). Washington, DC: Congressional Research Service.
Retrieved from ""