This section of the text scrutinizes the risks and limitations associated with the use of predictive and generative artificial intelligence, highlighting the intrinsic dangers that come with their utilization.
Narayanan and Kapoor emphasize that despite its widespread use by corporations and governmental entities, predictive AI often falls short of fulfilling its expected advantages. The accuracy of predictive AI can be compromised by the difficulty in anticipating human behavior, along with issues stemming from inadequate data, inherent biases, and overreliance.
Arvind Narayanan and Sayash Kapoor highlight a significant issue with AI systems that forecast outcomes: the decisions stemming from these models may be flawed, even though they can produce precise predictions based on past information. They examine an instance in which a predictive algorithm inaccurately deemed asthma patients to have a lower risk of pneumonia, which occurred because the algorithm's training data came from a medical protocol that ensured swift and comprehensive care for these individuals. Should the model replace the existing system and mistakenly release patients with asthma, the outcomes could be disastrous. AI predictions often overlook the impact of their determinations. Decisions might be misguided when they overlook the fundamental causative elements in the system and the consequences of altering these elements, even though they are grounded in reliable forecasts.
The authors present a different example of a predictive model claiming to determine the likelihood of hypertension with remarkable precision. The model's performance was gauged by examining the records of individuals previously diagnosed and treated for hypertension. The algorithm factored in whether a patient had been prescribed medication for hypertension, indicating a pre-existing medical condition. In the healthcare sector, there is a wide recognition of the importance of collecting new information, particularly for evaluating the effects of introducing novel medications or immunizations. Narayanan and Kapoor highlight that the dependability of assertions from companies specializing in predictive AI often falls short because of their inadequate abilities to gather sufficient data.
Context
- Effective predictions often require insights from multiple disciplines, such as sociology, psychology, or economics, to understand the full impact of decisions, which purely data-driven models might overlook.
- In complex systems, such as healthcare, altering one element can have cascading effects. For example, changing treatment protocols based on AI predictions without understanding the interconnectedness of patient care can lead to unintended health outcomes.
- Legal and regulatory constraints can impact the ability to gather and use data, affecting the comprehensiveness and accuracy of AI models.
- Many environments are dynamic and change over time. AI models trained on static data may fail to adapt to new conditions, leading to outdated or incorrect decisions.
The opaqueness of predictive AI systems, which obscures the decision-making process from users, fosters efforts to game the system. The book explores how artificial intelligence is employed in hiring processes, examining the way automated systems assess candidates based on criteria that lack clarity. Candidates modify their strategies by embedding notable university names in concealed text within their resumes and using complex terminology during video interviews that are assessed by artificial intelligence systems.
Narayanan and Kapoor examine a report that critically assesses Retorio, an employment tool that evaluates applicants through their behavior in recorded job interviews. The inquiry revealed that slight changes in an individual's visual representation, backdrop, or the arrangement of their resume had a substantial impact on the assessments made by artificial intelligence systems, underscoring their susceptibility to changes that bear no relation to a person's job performance capabilities. This trend in gaming behavior casts doubt on the precision professed by companies regarding their systems and reveals that non-transparent algorithms cause candidates to prioritize surface-level tactics over genuine skill enhancement.
Context
- Predictive AI systems often operate as "black boxes," meaning their internal workings are not visible or understandable to users. This lack of transparency can make it difficult for users to trust or verify the decisions made by these systems.
- AI systems might inadvertently hinder diversity efforts if they favor certain profiles based on historical data that does not reflect a diverse workforce.
- Many companies use AI to streamline the hiring process by automating the initial screening of resumes and conducting preliminary interviews. These systems often rely on algorithms to identify keywords and assess candidate responses.
- Retorio uses AI algorithms to analyze non-verbal cues, such as facial expressions, tone of voice, and body language, during video interviews to assess candidates' suitability for a job.
- AI systems often rely on computer vision algorithms that can be overly sensitive to visual inputs. This means that even minor changes in lighting, camera angle, or background can alter the system's perception and evaluation of a candidate.
- The use of non-transparent algorithms raises ethical questions about fairness and...
Unlock the full book summary of AI Snake Oil by signing up for Shortform.
Shortform summaries help you learn 10x better by:
Here's a preview of the rest of Shortform's AI Snake Oil summary:
This passage discusses the challenges involved in overseeing and controlling digital content on the internet by utilizing artificial intelligence. Narayanan and Kapoor argue that the complexities of online communication, along with the continuously changing terrain of societal conversation, inherently limit what artificial intelligence can achieve in monitoring conversations on the internet.
Narayanan and Kapoor argue that the subtleties and context within online discussions often elude AI, leading to errors in content moderation. They narrate the incident of Mark, a father, who was denied access to his Google account after the company's automated systems erroneously flagged photographs of his son's rash-afflicted private areas, which were intended for a doctor's review, as child exploitation material. This example, along with other instances where an account faced suspension for disseminating an illustration featuring Captain America in combat with a Nazi, and the situation where the complete collection of video material from Cornell University was taken...
This section delves into the broader social, economic, and organizational forces that shape the development and application of artificial intelligence, highlighting the importance of hyperbole, lack of transparency, and the overreliance on artificial intelligence to solve complex social problems.
The authors argue that the current trajectory in the field of AI research, particularly the focus on outperforming established standards and the strong ties to industry, creates a climate that overstates AI capabilities while obscuring its limitations.
Narayanan and Kapoor emphasize the growing influence of corporate funding on the advancement of AI studies, with companies like OpenAI, Google, and Meta leading the charge in developing advanced AI technologies. The authors highlight that corporate funding can lead to a bias, as it often prioritizes the development of products for the market rather than ensuring the research is comprehensive and...
AI Snake Oil
This is the best summary of How to Win Friends and Influence People I've ever read. The way you explained the ideas and connected them to other books was amazing.