Turing’s Vision: Navigating the Landscape of Ethical and Safe AI

First published 2022; revised 2023

In the dawn of the artificial intelligence era, there is an imperative need to navigate the complexities of AI ethics and safety. Ensuring that AI systems are both safe and ethically sound is no longer just a theoretical concern but a pressing practical issue that affects the global threads of industry, governance, and society at large. Drawing insights from Leslie, D. (2019) in “Understanding artificial intelligence ethics and safety: A guide for the responsible design and implementation of AI systems in the public sector”, published by The Alan Turing Institute, this essay explores the varied dimensions of AI’s responsible design and implementation.

The Alan Turing Institute forges its position as an aspirational, world-leading hub that examines the technical intricacies that underpin safe, ethical, and trustworthy AI. Committed to fostering responsible innovation and pioneering research breakthroughs, the Institute aims to go beyond mere theoretical discourses. It envisions a future where AI not only advances in capabilities but also upholds the core values of transparency, fairness, robustness, and human-centered design. Such an ambition necessitates a commitment to advancing AI transparency, ensuring the fairness of algorithmic systems, forging robust systems resilient against external threats, and cultivating AI-human collaborations that maintain human control.

However, the quest to realise this vision is not an isolated endeavour. It requires broad, interdisciplinary collaborations, connecting the dots between technical experts, industry leaders, policy architects, and the public. Aligning with the UK government’s Industrial Strategy and meeting the burgeoning global demand for informed guidance in AI ethics, the Institute’s strategy serves as a blueprint for those committed to the responsible growth of AI. However, it’s essential to remember that the responsible evolution of AI is not just about mastering the technology but understanding its implications for the broader context of our society.

The dawn of the information age has been marked by an extraordinary convergence of factors: the expansive availability of big data, the unparalleled speed and reach of cloud computing platforms, and the maturation of intricate machine learning algorithms. This synergy has propelled us into an era of unmatched human potential, characterised by a digitally interwoven world where the power of AI stands as a beacon of societal improvement.

Already, we witness the profound impact of AI across various sectors. Essential social domains such as healthcare, education, transportation, food supply, energy, and environmental management have all been beneficiaries of AI-driven innovations. These accomplishments, however significant they may appear now, are perhaps only the tip of the iceberg. AI’s very nature, its inherent capability to evolve and refine itself with increased access to data and surging computing power, guarantees its continuous ascent in efficacy and utility. As we navigate further into the information age, it’s conceivable that AI will soon stand at the forefront, guiding the progression of critical public interests and shaping the contours of sustainable human development.

Such a vision, where AI aids humanity in addressing its most pressing challenges, is undeniably exhilarating. Yet, like any frontier technology that’s rapidly evolving, AI’s journey is fraught with pitfalls. A steep learning trajectory ensures that errors, misjudgments, and unintended consequences are not just possible but inevitable. AI, despite its immense promise, is not immune to these challenges.

Addressing these challenges is not a mere recommendation but a necessity. It is imperative to prioritise AI ethics and safety to ensure its responsible evolution and to maximise its public benefit. This means an in-depth integration of social and ethical considerations into every facet of AI deployment. It calls for a harmonised effort, requiring data scientists, product managers, data engineers, domain experts, and delivery managers to work in unison. Their collective goal? To align AI’s development with ethical values and principles that not only prevent harm but actively enhance the well-being of communities that come under its influence.

The emergence of the field of AI ethics is a testament to this necessity. Born out of a growing recognition of the potential individual and societal harms stemming from AI’s misuse, poor design, or unforeseen repercussions, AI ethics seeks to provide a compass by which we navigate the AI-driven future responsibly.

Understanding the evolution of AI and its implications requires us to first recognise the genesis of AI ethics. The eminent cognitive scientist and AI trailblazer, Marvin Minsky, once described AI as the art of enabling computers to perform tasks that, when done by humans, necessitate intelligence. This fundamental definition highlights a crucial aspect of the discourse surrounding AI: humans, when undertaking tasks necessitating intelligence, are held to standards of reliability, accuracy, and sound reasoning. We expect them to justify their decisions, and to act with fairness, equity, and reasonableness in their interactions.

However, the rise and spread of AI technologies have reshaped this landscape. As AI systems take over myriad cognitive functions, they introduce a conundrum. Unlike humans, these algorithmic processes aren’t directly accountable for their actions, nor can they be held morally responsible for the outcomes they produce. Essentially, while AI systems exhibit a form of ‘smart agency’, they lack inherent moral responsibility, creating a discernible ethical void.

Addressing this void has become paramount, giving birth to a host of frameworks within AI ethics. One such framework is the FAST Track Principles, which stands for Fairness, Accountability, Sustainability, and Transparency. These principles are designed to bridge the gap between AI’s capabilities and its intrinsic moral void. To foster an environment conducive to responsible AI development, it is vital that every stakeholder, from data scientists to policy experts, familiarises themselves with the FAST Track Principles. These principles should guide actions and decisions throughout the AI project lifecycle, underscoring the idea that creating ethical AI is a collective endeavor.

Delving deeper into the principle of fairness, one must remember that while AI systems might project a veneer of neutrality, they are ultimately products of human design. Humans, with all their inherent biases and contextual limitations, play a pivotal role in AI’s creation. At any stage of an AI project, from data extraction to model building, the spectres of human error, prejudice, and misjudgment can introduce biases. Moreover, AI systems often derive their accuracy by analysing data that might encapsulate age-old societal biases and discriminations, further complicating the fairness equation.

Addressing fairness in AI is far from straightforward. There isn’t a singular, foolproof method to eliminate biases or ensure fairness. However, by adopting best practices that focus on fairness-aware design and implementation, there’s potential to create systems that yield just and equitable outcomes. One foundational approach to fairness is the principle of discriminatory non-harm. It mandates that AI innovations should not result in harm due to biased or discriminatory outcomes. This principle, while seemingly basic, serves as a cornerstone, directing the development and deployment of AI systems towards a more equitable and fair future.

The Principle of Discriminatory Non-Harm sets forth that AI system designers and users should be deeply committed to reducing biases and preventing discriminatory outputs, especially when dealing with social or demographic data. This implies a few specific obligations. First, AI systems should be built upon data that is representative, accurate, and generalisable, ensuring “Data Fairness.” Second, the systems’ design should not include any variables, features, or processes that are morally objectionable or unjustifiable – this is “Design Fairness.” The systems should also be crafted to avoid producing discriminatory effects on individuals or groups – ensuring “Outcome Fairness.” Lastly, the onus is on the users to be adequately trained to use AI systems responsibly, embodying “Implementation Fairness.”

When considering the concept of Accountability in AI, the best practices for data processing as mentioned in Principle 6 of the Data Ethics Framework come to mind. However, the ever-evolving AI landscape brings forward distinct challenges, especially in public sector accountability. Two major challenges emerge: the “accountability gap” and the multifaceted nature of AI production processes. Automated decisions, inherently, are not self-explanatory. Unlike human agents, statistical models and AI’s underlying infrastructure don’t bear moral responsibility, creating a void in accountability. Coupled with this is the intricate nature of AI project deliveries involving a myriad of stakeholders, making it a daunting task to pinpoint responsibility if an AI system’s implementation has adverse consequences.

To address these challenges, it’s imperative to adopt a comprehensive approach to accountability that encompasses both Answerability and Auditability. Answerability stresses that human creators and users of AI systems should take full responsibility for the algorithmically-driven decisions. They should be ready to provide clear, coherent, and non-technical explanations for these decisions, ensuring that every stage of the AI process is accountable. Auditability, on the other hand, focuses on how to hold these AI system designers and implementers accountable. It emphasises the demonstration of both responsible design and use practices, and the justifiability of the outcomes.

Another critical pillar is Sustainability. AI system designers and users must be continually attuned to the long-term and transformative effects their technologies might have on individuals and society at large. This proactive awareness ensures that the systems not only address the immediate needs but also consider the long-term societal impacts.

In tandem with sustainability is Safety. Besides considering the broader social ramifications of an AI system, it’s essential to address its technical sustainability and safety. Given that AI operates in an unpredictable environment, achieving technical safety becomes a challenging task. However, the importance of building a safe and reliable AI system cannot be overstated, especially when potential failures could result in harmful consequences and erode public trust. To achieve this, emphasis must be placed on the core technical objectives of accuracy, reliability, security, and robustness. This involves rigorous testing, consistent validation, and frequent reassessment of the system. Moreover, effective oversight mechanisms need to be integrated into the system’s real-world operation to ensure that it functions safely and as intended.

The intrinsic challenges of accuracy in artificial intelligence systems can be linked to the inherent complexities and unpredictability of the real world. When trying to model this chaotic reality, it’s a significant task to ensure that an AI system’s predictions or classifications are precise. Data noise, which is unavoidable, combined with the potential that a model might not capture all aspects of the underlying patterns and changes in data over time, can all contribute to these challenges.

On the other hand, the reliability of an AI system rests on its ability to consistently function in line with its intended design and purpose. This means that if a system is deemed reliable, users can trust that its operations will adhere to its set specifications, bolstering user confidence in the safety and predictability of its outcomes.

AI systems also face threats on the security front. Security is not just about safeguarding an AI system from potential external threats but also ensuring that the system’s architecture remains uncompromised and that any data or information within it remains confidential. This integrity is paramount, especially when considering the potential adversarial threats that AI systems might face.

Robustness in AI, meanwhile, centres on an AI system’s ability to function effectively even under less than ideal conditions. Whether these conditions arise from intentional adversarial actions, human errors, or misalignments in automated learning objectives, the system’s ability to maintain its integrity is a testament to its robustness.

One of the more nuanced challenges that machine learning models face is the phenomenon of concept drift. When the historical data, which informs the model’s understanding, becomes outdated or misaligned with current realities, the model’s accuracy and reliability can suffer. Therefore, staying attuned to changes in the underlying data distribution is vital. Ensuring that the technical team is aware of the latest research on detecting and managing concept drift will be crucial to the continued success of AI projects.

Another pressing concern in the realm of AI is adversarial attacks. These attacks cleverly manipulate input data, causing AI models to make grossly incorrect predictions or classifications. The subtle nature of these perturbations can lead to significant ramifications, especially in critical systems like medical imaging or autonomous vehicles. Recognising these vulnerabilities, there has been a surge in research in the domain of adversarial machine learning, aiming to safeguard AI systems from these subtle yet disruptive inputs.

Equally concerning is the threat of data poisoning, where the very data that trains an AI system is tampered with, causing the system to generate inaccurate or harmful outputs. This kind of attack can be especially sinister as it might incorporate ‘backdoors’ into the system, which when triggered, can cause malfunctions. Therefore, beyond technical solutions, it becomes imperative to source data responsibly and ensure its integrity throughout the data handling process. The emphasis should be on responsible data management practices to ensure data quality throughout the system’s lifecycle.

In the world of artificial intelligence, the term “transparency” has taken on a nuanced and specialised meaning. While the everyday usage of the term typically evokes notions of clarity, openness, and straightforwardness, in AI ethics, transparency becomes even more multifaceted. One aspect of this is the capacity for AI systems to be interpretable. That is, those interacting with an AI system should be able to decipher how and why the system made a particular decision or acted in a certain way. This kind of transparency is about shedding light on the internal workings of the often enigmatic AI mechanisms, allowing for greater understanding and trust.

Furthermore, transparency isn’t limited to merely understanding the “how” and “why” of AI decisions. It also encompasses the ethical considerations behind both the design and deployment of AI systems. When AI systems are said to be transparent, it implies that they can be justified as ethical, unbiased, trustworthy, and safety-oriented both in their creation and their outcomes. This dual focus on process and product is vital.

In developing AI, teams are tasked with several responsibilities to ensure this two-tiered transparency. First, from a process perspective, there is a need to assure all stakeholders that the entire journey of creating the AI system was ethically sound, unbiased, and instilled with measures ensuring trust and safety. This includes not just designing with these values in mind but also ensuring auditability at every stage.

Secondly, when it comes to the outcome or product of AI, there’s the obligation to make sure that any decision made by the AI system is elucidated in ways that are understandable to non-experts. The explanations shouldn’t merely regurgitate the mathematical or technical jargon but should be phrased in relatable terms, reflecting societal contexts. Furthermore, the results or behaviors of the AI should be defensible, fitting within parameters of fairness, trustworthiness, and ethical appropriateness.

In addition to these tasks, there’s a broader need for professional and institutional transparency. Every individual involved in the AI’s development and deployment should adhere to stringent standards that emphasise values like integrity, honesty, and neutrality. Their primary allegiance should be to the public’s best interests, superseding other considerations.

Moreover, throughout the AI development process, there should be an open channel for public oversight. Of course, certain information may need to remain confidential for valid reasons, like ensuring bad actors can’t exploit the system. But, by and large, the emphasis should be on openness.

Transitioning into the structural aspects of AI development, a Process-Based Governance (PBG) Framework emerges as a crucial tool. Such a framework is pivotal for integrating ethical considerations and best practices seamlessly into the actual development process. The guide might delve into specifics like the CRISP-DM, but it’s worth noting that the principles of responsible AI development can be incorporated into other workflow models, including KDD and SEMMA. Adopting such a framework helps ensure that the values underpinning ethical AI are not just theoretical but find active expression in every phase of the AI’s life cycle.

Alan Turing’s simple sketch in 1936 was nothing short of revolutionary. With just a linear tape, symbols, and a set of rules, he demystified the very essence of calculations, giving birth to the conceptual foundation of the modern computer. His Turing machine wasn’t just a solution to the enigma of effective calculations, it was the conceptual forerunner of the digital revolution we live in today. This innovative leap, stemming from a quiet room at Kings College, Cambridge, is foundational to our digital landscape.

Fast forward to our present day, and we find ourselves immersed in a world where the lines between the physical and digital blur. The seamless interplay of connected devices, sophisticated algorithms, and vast cloud computing platforms is redefining our very existence. Technologies like the Internet of Things and edge computing are not just changing the way we live and work; they’re reshaping the very fabric of our society. AI is becoming more than just a tool or a technology; it is rapidly emerging as the fulcrum upon which our future balances. The possibilities it presents, both optimistic and cautionary, are monumental. It’s essential to realise that the trajectory of AI’s impact lies in our hands. The decisions we make today will shape the society of tomorrow, and the implications of these choices weigh heavily on our collective conscience.

It’s paramount to see that artificial intelligence isn’t just about codes and algorithms. It’s about humanity, our aspirations, our values, and our shared vision for the future. In many ways, the guide on AI ethics and safety serves as a compass, echoing Turing’s ethos by emphasising that the realm of AI, at its core, remains a profoundly human domain. Every line of code, every algorithmic model, every deployment carries with it a piece of human intention, purpose, and responsibility.

In essence, understanding the ethics and safety of AI isn’t just about mitigating risks or optimising outputs. It’s about introspection and realising that behind every technological advancement lie human choices. Responsible innovation isn’t just a catchphrase; it’s a call to action. Only by staying grounded in our shared ethical values and purpose-driven intentions can we truly harness AI’s potential. Let’s not just be passive recipients of technology’s gifts. Instead, let’s actively shape its direction, ensuring that our collective digital future resonates with our shared vision of humanity’s greatest aspirations.

Links

https://www.turing.ac.uk/news/publications/understanding-artificial-intelligence-ethics-and-safety

https://www.turing.ac.uk/sites/default/files/2019-06/understanding_artificial_intelligence_ethics_and_safety.pdf

Did Galen Help or Hinder Progress in Medicine?

First published 2021

Galen, a prolific Greek author of antiquity, played a foundational role in medical education throughout the Byzantine empire and Europe for numerous centuries. However, Galen primarily considered himself a medical practitioner rather than an anatomist, theorist, or researcher. He dedicated many decades of his life, spanning from approximately 162 CE until his death after 203, to treating patients primarily in the city of Rome within the Roman empire. His writings now offer us the most direct glimpse into the medical practices of highly skilled ancient doctors.

Galen’s patients ranged from some of the empire’s most influential figures, including Emperor Marcus Aurelius and his household, to ordinary individuals. He provided medical care not only to the wealthy, treating their spouses, children, and slaves, but also to peasants he encountered in rural areas, acquaintances of acquaintances he met on the streets, and various patients who visited the clinic in his home. In a lost treatise called “That the Best Men Profit from their Enemies,” preserved in Arabic, Galen even claimed never to have accepted payment for his services. The numerous anecdotes Galen shared about his patients, over 300 of which survive in his works, do not portray him as a detached courtier distanced from everyday society. Instead, they illustrate Galen as a public figure, widely known and recognized, approached on the streets, invited to engage in debates, and constantly accompanied by a retinue of friends, supporters, students, household staff, and professional assistants.

However, the question of whether he helped or hindered progress in medicine is a matter of debate. It is necessary to explore both sides of the argument, shedding light on additional aspects of his influence.

On one hand, Galen’s contributions to medicine were substantial and influential. He was a prolific writer, and his works became the standard reference in medical education for centuries. Galen’s emphasis on systematic observation and dissection of animals, particularly pigs and monkeys, contributed to a deeper understanding of anatomy. His descriptions of the circulatory system, nerves, and muscles, although not entirely accurate by modern standards, were groundbreaking for his time and laid the foundation for future anatomical studies.

Furthermore, Galen’s writings on pathology and therapeutics were influential. He developed a system of humoral medicine, which proposed that health depended on the balance of four bodily humors (blood, phlegm, black bile, and yellow bile). While this theory may seem outdated today, it guided medical practice for centuries, shaping the way physicians diagnosed and treated diseases.

However, there are also significant arguments to suggest that Galen hindered progress in medicine. First, his teachings were dogmatic and dominated medical thought for over a millennium. This rigid adherence to Galenic principles stifled innovation and discouraged the questioning of established ideas, potentially impeding the development of more accurate medical theories and treatments.

Second, Galen’s reliance on animal dissection limited his understanding of human anatomy. He made several errors in his anatomical descriptions due to the anatomical differences between humans and the animals he studied. This hindered the advancement of human anatomy and delayed the correction of these errors until the Renaissance when human dissection became more accepted.

Furthermore, the dominance of Galenic medicine discouraged the pursuit of alternative medical approaches and hindered the development of a more diverse and comprehensive medical knowledge base. Only with the gradual decline of Galenism in the Renaissance did medical science begin to explore new avenues of understanding and treatment.

In conclusion, Galen’s contributions to medicine were undeniably significant, especially in terms of his anatomical observations and descriptions. However, his dogmatic influence and the limitations of his theories had the effect of hindering progress in medicine to some extent. While Galen’s work laid important foundations, it also held back the development of more accurate and innovative medical practices. Ultimately, the impact of Galen on the progress of medicine is a complex and multifaceted issue, and the answer to whether he helped or hindered remains a matter of historical interpretation.

Links

https://www.thelancet.com/journals/lancet/article/PIIS0140-6736(11)61240-3/fulltext