Academia.edu no longer supports Internet Explorer.
To browse Academia.edu and the wider internet faster and more securely, please take a few seconds to upgrade your browser.
2020, Artificial General Intelligence
The complex socio-technological debate underlying safetycritical and ethically relevant issues pertaining to AI development and deployment extends across heterogeneous research subfields and involves in part conflicting positions. In this context, it seems expedient to generate a minimalistic joint transdisciplinary basis disambiguating the references to specific subtypes of AI properties and risks for an error-correction in the transmission of ideas. In this paper, we introduce a high-level transdisciplinary system clustering of ethical distinction between antithetical clusters of Type I and Type II systems which extends a cybersecurityoriented AI safety taxonomy with considerations from psychology. Moreover, we review relevant Type I AI risks, reflect upon possible epistemological origins of hypothetical Type II AI from a cognitive sciences perspective and discuss the related human moral perception. Strikingly, our nuanced transdisciplinary analysis yields the figurative formulation of the so-called AI safety paradox identifying AI control and value alignment as conjugate requirements in AI safety. Against this backdrop, we craft versatile multidisciplinary recommendations with ethical dimensions tailored to Type II AI safety. Overall, we suggest proactive and importantly corrective instead of prohibitive methods as common basis for both Type I and Type II AI safety.
Philosophy and Technology, 2024
Risks connected with AI systems have become a recurrent topic in public and academic debates, and the European proposal for the AI Act explicitly adopts a riskbased tiered approach that associates different levels of regulation with different levels of risk. However, a comprehensive and general framework to think about AI-related risk is still lacking. In this work, we aim to provide an epistemological analysis of such risk building upon the existing literature on disaster risk analysis and reduction. We show how a multi-component analysis of risk, that distinguishes between the dimensions of hazard, exposure, and vulnerability, allows us to better understand the sources of AI-related risks and effectively intervene to mitigate them. This multi-component analysis also turns out to be particularly useful in the case of general-purpose and experimental AI systems, for which it is often hard to perform both ex-ante and ex-post risk analyses.
Nowadays, there is a serious anxiety on the existence of dangerous intelligent systems and it is not just a science-fiction idea of evil machines like the ones in well-known Terminator movie or any other movies including intelligent robots – machines threatening the existence of humankind. So, there is a great interest in some alternative research works under the topics of Machine Ethics, Artificial Intelligence Safety and the associated research topics like Future of Artificial Intelligence and Existential Risks. The objective of this study is to provide a general discussion about the expressed research topics and try to find some answers to the question of 'Are we safe enough in the future of Artificial Intelligence?'. In detail, the discussion includes a comprehensive focus on 'dystopic' scenarios, enables interested researchers to think about some 'moral dilemmas' and finally have some ethical outputs that are considerable for developing good intelligent systems. From a general perspective, the discussion taken here is a good opportunity to improve awareness on the mentioned, remarkable research topics associated with not only Artificial Intelligence but also many other natural and social sciences taking role in the humankind.
2018
Nowadays, there is a serious anxiety on the existence of dangerous intelligent systems and it is not just a science-fiction idea of evil machines like the ones in well-known Terminator movie or any other movies including intelligent robots – machines threatening the existence of humankind. So, there is a great interest in some alternative research works under the topics of Machine Ethics, Artificial Intelligence Safety and the associated research topics like Future of Artificial Intelligence and Existential Risks. The objective of this study is to provide a general discussion about the expressed research topics and try to find some answers to the question of ‘Are we safe enough in the future of Artificial Intelligence?’. In detail, the discussion includes a comprehensive focus on ‘dystopic’ scenarios, enables interested researchers to think about some ‘moral dilemmas’ and finally have some ethical outputs that are considerable for developing good intelligent systems. F...
AI and Ethics, 2025
Continuing the digital revolution, AI is capable to transform our world. Thanks to its novelty, we can define how we, as a society, envision this fascinating technology to integrate with existing processes. The EU AI Act follows a risk-based approach, and we argue that addressing the human influence, which poses risks along the AI lifecycle is crucial to ensure the desired quality of the model's transition from research to reality. Therefore, we propose a holistic approach that aims to continuously guide the involved stakeholders' mindset, namely developers and domain experts, among others towards Responsible AI (RAI) lifecycle management. Focusing on the development view with regard to regulation, our proposed four pillars comprise the well-known concepts of Generalizability, Adaptability and Translationality. In addition, we introduce Transversality (Welsch in Vernunft: Die Zeitgenössische Vernunftkritik Und Das Konzept der Transversalen Vernunft, Suhrkamp, Frankfurt am Main, 1995), aiming to capture the multifaceted concept of bias, and base the four pillars on Education, and Research. Overall, we aim to provide an application-oriented summary of RAI. Our goal is to distill RAI-related principles into a concise set of concepts that emphasize implementation quality. Concluding, we introduce the ethical foundation's transition to an applicable ethos for RAI projects as part of ongoing research.
Proceedings of the 40th Annual ARCOM Conference, 2024
Artificial intelligence (AI) is in need for a framework that balances the opportunities it represents with its risks. But while there is a broad consensus on this, and public regulative initiatives are taken; there is far less knowledge about how these dilemmas/opportunities/risks are played out in practice. The interest into ethics in organisation driven by a discourse on "Trustworthy AI"; makes us wonder whether an ethical approach to AI in organisation is purposeful; or needs modification. We investigate this by viewing the development and use of AI as structuration of practices. The empirical material is our own development of an AI system. Using studies of ethics in moral engineering design; AI is a question of structuration processers with unintended consequences. It is a "slide" from ethics of virtue to ethics of benefit as corroborated by engineers/designers referring ethical dilemmas to managers and politicians. The EU framework of Trustworthy AI for designing and using more accountable AI systems-considering ethics; human autonomy; harm prevention; fairness etc., conflicts with contemporary construction organisations. We propose an extension of the EU guidelines.
2011
Machine ethics and robot rights are quickly becoming hot topics in artificial intelligence/robotics communities. We will argue that the attempts to allow machines to make ethical decisions or to have rights are misguided. Instead we propose a new science of safety engineering for intelligent artificial agents. In particular we issue a challenge to the scientific community to develop intelligent systems capable of proving that they are in fact safe even under recursive self-improvement.
Futures
Artificial Intelligence (AI) is one of the most transformative technologies of the 21st century. The extent and scope of future AI capabilities remain a key uncertainty, with widespread disagreement on timelines and potential impacts. As nations and technology companies race toward greater complexity and autonomy in AI systems, there are concerns over the extent of integration and oversight of opaque AI decision processes. This is especially true in the subfield of machine learning (ML), where systems learn to optimize objectives without human assistance. Objectives can be imperfectly specified or executed in an unexpected or potentially harmful way. This becomes more concerning as systems increase in power and autonomy, where an abrupt capability jump could result in unexpected shifts in power dynamics or even catastrophic failures. This study presents a hierarchical complex systems framework to model AI risk and provide a template for alternative futures analysis. Survey data were collected from domain experts in the public and private sectors to classify AI impact and likelihood. The results show increased uncertainty over the powerful AI agent scenario, confidence in multiagent environments, and increased concern over AI alignment failures and influence-seeking behavior.
Inteligencia artificial, 2024
In the paper, an analysis is conducted on the intricate relationship between ethics, artificial intelligence, and cybersecurity. The ethical principles that govern the advancement of AI are examined, alongside the security issues that arise from its implementation. The ethical utilization of artificial intelligence in the realms of cybersecurity and hacking is explored. Emphasis is placed on the significance of AI ethics, particularly in terms of transparency, accountability, and fairness. Additionally, the paper delves into the security challenges that emerge as AI is adopted, such as safeguarding user privacy and ensuring equitable access to the technology.
Texila International Journal of Academic Research, 2019
This paper will present and analyze reported failures of artificially intelligent systems and extrapolate our analysis to future AIs. I suggest that both the frequency and the seriousness of future AI failures will steadily increase. AI Safety can be improved based on ideas developed by cybersecurity experts. For narrow AIs safety failures are at the same, moderate, level of criticality as in cybersecurity, however for general AI, failures have a fundamentally different impact. A single failure of a super intelligent system may cause a catastrophic event without a chance for recovery. The goal of cybersecurity is to reduce the number of successful attacks on the system; the goal of AI Safety is to make sure zero attacks succeed in bypassing the safety mechanisms. Unfortunately, such a level of performance is unachievable. Every security system will eventually fail; there is no such thing as a 100% secure system. Future generations may look back at our time and identify it as one of intense change. In a few short decades, we have morphed from a machine-based society to an information-based society, and as this Information Age continues to mature, society has been forced to develop a new and intimate familiarity with data-driven and algorithmic systems. Artificial agents to refer to devices and decision-making aids that rely on automated, data- driven, or algorithmic learning procedures. Such agents are becoming an intrinsic part of our regular decision-making processes. Their emergence and adoption lead to a bevy of related policy questions. Keywords: AI Safety, Cybersecurity, Failures, Super intelligence, Algorithms, Advanced Persistent Threats (APT).
2024
Artificial intelligence (AI) assurance is an umbrella term describing many approaches-such as impact assessment, audit, and certification procedures-used to provide evidence that an AI system is legal, ethical, and technically robust. AI assurance approaches largely focus on two overlapping categories of harms: deployment harms that emerge at, or after, the point of use, and individual harms that directly impact a person as an individual. Current approaches generally overlook upstream collective and societal harms associated with the development of systems, such as resource extraction and processing, exploitative labour practices and energy intensive model training. Thus, the scope of current AI assurance practice is insufficient for ensuring that AI is ethical in a holistic sense, i.e., in ways that are legally permissible, socially acceptable, economically viable and environmentally sustainable. This article addresses this shortcoming by arguing for a broader approach to AI assurance that is sensitive to the full scope of AI development and deployment harms. To do so, the article maps harms related to AI and highlights three examples of harmful practices that occur upstream in the AI supply chain and impact the environment, labour, and data exploitation. It then reviews assurance mechanisms used in adjacent industries to mitigate similar harms, evaluating their strengths, weaknesses, and how effectively they are being applied to AI. Finally, it provides recommendations as to how a broader approach to AI assurance can be implemented to mitigate harms more effectively across the whole AI supply chain.
The humankind is currently experiencing a life supported often with intelligent systems designed and developed based on the foundations of Artificial Intelligence. It is clear that this scientific field is one of key elements for shaping better future for us. But there are also some anxieties regarding possible ethical and safety related issues that may arise because of intense use of powerful Artificial Intelligence oriented systems. In this context, objective of this paper is to provide a look at to some remarkable issues about ethics and safety within the future of Artificial Intelligence. After focusing on currently wide-discussed issues, the paper also comes with some possible solution suggestions for achieving a better Artificial Intelligence supported future with no or less issues on ethics and safety.
AI advances represent a great technological opportunity, but also possible perils. This paper undertakes an ethical and systematic evaluation of those risks in a pragmatic analytical form of questions, which we term 'Conceptual AI Risk analysis'. We then look at a topical case example in an actual industrial setting and apply that methodology in outline. The case involves Deep Learning Black-Boxes and their risk issues in an environment that requires compliance with legal rules and industry best practices. We examine a technological means to attempt to solve the Black-box problem for this case, referred to as "Really Useful Machine Learning" ( RUML SM ). DARPA has identified such cases as being the "Third Wave of AI." Conclusions to its efficacy are drawn. Martin Ciupa is the CTO of calvIO Inc., a company (associated with the Calvary Robotics group of companies) focused on simplifying the cybernetic interaction between man and machine in the industrial setting. Martin has had a career in both technology, general management and commercial roles at senior levels in North America, Europe and Asia. He has an academic background in Physics and Cybernetics. He has applied AI and Machine learning systems to applications in decision support for Telco, Manufacturing and Financial services sectors and published technical articles in Software, Robotics, AI and related disciplines.
The Alan Turing Institute, 2024
Project teams frequently engage in tasks pertaining to the technical safety and sustainability of their AI projects. In doing so, they need to ensure that their resultant models are reproducible, robust, interpretable, reliable, performant, and secure. The issue of AI safety is of paramount importance, because possible failures have the potential to produce harmful outcomes and undermine public trust. This work of building safe AI outputs is an ongoing process requiring reflexivity and foresight. To aid teams in this, the workbook introduces the core components of AI Safety (reliability, performance, robustness, and security), and helps teams develop anticipatory and reflective skills which are needed to responsibly apply these in practice. This workbook is part of the AI Ethics and Governance in Practice series (https://aiethics.turing.ac.uk) co-developed by researchers at The Alan Turing Institute in partnership with key public sector stakeholders.
arXiv (Cornell University), 2024
The integration of advanced artificial intelligence (AI) across contemporary sectors and industries is not just a technological upgrade but a transformation with the potential to have profound implications. This paper explores the concept of structural risks associated with the rapid integration of advanced AI across social, economic, and political systems. This framework challenges conventional perspectives that primarily focus on direct AI threats such as accidents and misuse and suggests that these more proximate risks influence and are influenced by larger sociotechnical dynamics. By analyzing the complex interactions between technological advancements and social dynamics, this study identifies three primary categories of structural risks: antecedent structural causes, antecedent AI system causes, and deleterious feedback loops. We present a comprehensive framework to understand the causal chains that drive these risks, highlighting the interdependence between structural forces and the more proximate risks of misuse, system failures, and the diffusion of misaligned systems. The paper articulates how unchecked AI advancement can reshape power dynamics, trust, and incentive structures, with the potential for profound and unpredictable societal shifts. We introduce a methodological research agenda for mapping, simulating, and gaming these dynamics aimed at preparing policymakers and national security professionals for the challenges posed by next-generation AI technologies. The paper concludes with policy recommendations to incorporate a more nuanced understanding of the sociotechnical nexus into international governance and strategy.
AI advances represent a great technological opportunity, but also possible perils. This paper undertakes an ethical and systematic evaluation of those risks in a pragmatic analytical form of questions, which we term ‘Conceptual AI Risk analysis’. We then look at a topical case example in an actual industrial setting and apply that methodology in outline. The case involves Deep Learning Black-Boxes and their risk issues in an environment that requires compliance with legal rules and industry best practices. We examine a technological means to attempt to solve the Black-box problem for this case, referred to as “Really Useful Machine Learning” ( RUMLSM ). DARPA has identified such cases as being the “Third Wave of AI.” Conclusions to its efficacy are drawn.
Front. Comput. Sci., 2024
Introduction: The rapid evolution of Artificial Intelligence (AI) has introduced transformative potential across various sectors, while simultaneously posing significant cybersecurity risks. Methods: The aim of this paper is to examine the debates on AI-related cybersecurity risks through the lens of Beck's theory of the risk society. Utilizing thematic content analysis, we explored public discourse on AI and cybersecurity as presented in articles published by WIRED. Results: Our analysis identified several key themes: the global nature of AI risks, their pervasive influence across multiple sectors, the alteration of public trust, the individualization of risk, and the uneven distribution of AI risks and benefits. Discussion: The editorial choices in WIRED predominantly favor a functionalist and solutionist perspective on AI cybersecurity risks, often marginalizing the opinions of ordinary individuals and non-Western voices. This editorial bias tends to limit diversity and underrepresent key opposing viewpoints, potentially hindering a more comprehensive and nuanced debate on AI and cybersecurity issues.
2023
Artificial Intelligence (AI) is a rapidly advancing technology that permeates human life at various levels. It evokes hopes for a better, easier, and more exciting life, while also instilling fears about the future without humans. AI has become part of our daily lives, supporting fields such as medicine, customer service, finance, and justice systems; providing entertainment, and driving innovation across diverse fields of knowledge. Some even argue that we have entered the “AI era.” However, AI is not solely a matter of technological progress. We already witness its positive and negative impact on individuals and societies. Hence, it is crucial to examine the primary challenges posed by AI, which is the subject of AI ethics. In this paper, I present the key challenges that emerged in the literature and require ethical reflection. These include the issues of data privacy and security, the problem of AI biases resulting from social, technical, or socio-technical factors, and the challenges associated with using AI for prediction of human behavior (particularly in the context of the justice system). I also discuss existing approaches to AI ethics within the framework of technological regulations and policymaking, presenting concrete ways in which ethics can be implemented in practice. Drawing on the functioning of other scientific and technological fields, such as gene editing, the development of automobile and aviation industries, I highlight the lessons we can learn from how they function to later apply it to how AI is introduced in societies. In the final part of the paper, I analyze two case studies to illustrate the ethical challenges related to recruitment algorithms and risk assessment tools in the criminal justice system. The objective of this work is to contribute to the sustainable development of AI by promoting human-centered, societal, and ethical approaches to its advancement. Such approach seeks to maximize the benefits derived from AI while simultaneously mitigating its diverse negative consequences.
Oxford University Press eBooks, 2022
This chapter formulates seven lessons for preventing harm in artificial intelligence (AI) systems based on insights from the field of system safety for software-based automation in safety-critical domains. New applications of AI across societal domains and public organizations and infrastructures come with new hazards, which lead to new forms of harm, both grave and pernicious. The text addresses the lack of consensus for diagnosing and eliminating new AI system hazards. For decades, the field of system safety has dealt with accidents and harm in safety-critical systems governed by varying degrees of software-based automation and decision-making. This field embraces the core assumption of systems and control that AI systems cannot be safeguarded by technical design choices on the model or algorithm alone, instead requiring an end-to-end hazard analysis and design frame that includes the context of use, impacted stakeholders and the formal and informal institutional environment in which the system operates. Safety and other values are then inherently socio-technical and emergent system properties that require design and control measures to instantiate these across the technical, social and institutional components of a system. This chapter honors system safety pioneer Nancy Leveson, by situating her core lessons for today's AI system safety challenges. For every lesson, concrete tools are offered for rethinking and reorganizing the safety management of AI systems, both in design and governance. This history tells us that effective AI safety management requires transdisciplinary approaches and a shared language that allows involvement of all levels of society.
Loading Preview
Sorry, preview is currently unavailable. You can download the paper by clicking the button above.