Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
SlideShare a Scribd company logo
Quantum Fax Machine
QFM024: Irresponsible AI Reading
List June 2024
quantumfaxmachine.com
1
QFM024: Irresponsible AI Reading
List June 2024
The June edition of the Irresponsible AI Reading List starts with a critique of
AI's potential for exaggerated promises in science and academia then heads
over to look at the societal implications of AI-driven job automation and privacy
concerns. An ongoing theme emerges for the urgent need for critical, informed
engagement with AI technologies.
We then take a look at the role that AI plays in amplifying societal biases,
privacy rights concerning AI profiling, and the mental health impacts of AI-
induced job insecurity. The discussion on Meta's use of facial recognition for age
verification highlights the tension between technological advancements and
privacy rights, while Citigroup's report on AI's potential to automate over half of
banking jobs illustrates the profound economic shifts underway in the job
market.
ChatBug: Tricking AI Models into Harmful Responses looks at how AI model
vulnerabilities can lead to harmful exploitation, pointing to the critical need for
even more robust cybersecurity measures when working with AI.
As always, the Quantum Fax Machine Propellor Hat Key will guide your
browsing. Enjoy!
Key:
: Nothing to do with AI, but interesting nonetheless
: Mentions AI
: Talks about irresponsible AI
: Talks about irresponsible AI in a real-world failure scenario
: Talks about technical details of irresponsible AI
: Discusses technical details and mitigation of irresponsible AI
Source: Photo by Thomas Kinto on Unsplash
2
Scientists should use AI as a tool, not an
oracle (aisnakeoil.com): This article
discusses the hype surrounding AI and its
misuse in scientific research, leading to
flawed studies and reproducibility issues. It
emphasises that AI should be used as a
tool to aid human understanding rather
than an infallible oracle, advocating for a
cultural shift towards more critical and
careful use of AI in scientific disciplines to
ensure research integrity and reliability.
#AI #Science #ResearchIntegrity
#Reproducibility #MachineLearning
3
Critique of Forrester's Foundation Model
Assessment (linkedin.com): Peter Gostev
critiques Forrester's Foundation Model
assessment, highlighting issues with the
weightings and scores of various AI
models, pointing out inconsistencies and
lack of specificity in the evaluation process.
#AI #MachineLearning
#ForresterReport #ModelAssessment
#AIevaluation
4
AI Is a False God -- (thewalrus.ca): Navneet
Alang critiques the overhyped promises of
AI, arguing that while it can enhance data
processing and efficiency, it lacks the moral
and subjective capacities necessary to
solve deeply human problems, often
exacerbating existing biases and societal
issues.
#AI #TechCritique
#ArtificialIntelligence
#EthicsInAI #TechnologyImpact
5
What is the biggest challenge in our
industry? (thrownewexception.com): The
biggest challenge in the tech industry is the
anxiety caused by layoffs and the fear of AI
replacing jobs, leading to mental health
issues like burnout. Leaders can help by
fostering open communication, leading
positively, leveraging new technologies,
investing in continuous learning, and
collaborating with HR to support their
teams.
#TechIndustry #AI #MentalHealth
#Leadership #Layoffs
6
The Right Not to Be Subjected to AI Profiling
Based on Publicly Available Data—Privacy and
the Exceptionalism of AI Profiling: This article
argues for the new legal right to protect
individuals from AI profiling based on publicly
available data without their explicit informed
consent. It develops three primary arguments
dealing with social control, stigmatisation, and
the unique threat posed by AI profiling
compared to other data processing methods.
The article suggests that existing GDPR
regulations are not sufficient and calls for
explicit regulation with a sui generis right for
protection.
#AI #Privacy #Profiling #TechLaw
#DataProtection
7
Don't Expect Juniors to Teach Senior Professionals
to Use Generative AI: Emerging Technology Risks
and Novice AI Risk Mitigation Tactics: A Harvard
Business Working Paper explores the challenges
and limitations of expecting junior professionals to
guide senior professionals in the use of emerging
technologies like generative AI. The study,
conducted with Boston Consulting Group, included
interviews with junior consultants using GPT-4 and
found that they often lack deep understanding and
experience, making them ineffective in mitigating AI
risks at a senior level. Insights suggest the need for
more seasoned strategies and mitigation tactics
focusing on system design and ecosystem-level
changes.
#AI #EmergingTech #GenerativeAI
#TechRisks #Innovation
8
How a Single ChatGPT Mistake Cost Us
$10,000+: In the early days of a startup, a
critical mistake involving ChatGPT-
generated code cost over $10,000 in lost
sales. The problem stemmed from a single
hardcoded ID that caused unique ID
collisions, preventing new users from
subscribing. This story highlights the
importance of robust testing and the perils
of copy-pasting code in production
environments.
#startup #ChatGPT #codingmistakes
#techerrors #softwaredevelopment
9
Meta Uses Facial Recognition for Age
Verification Amid Political Pressure: Meta is
now using facial recognition to verify the
age of some users on Facebook and
Instagram. This move comes amid growing
political pressure to protect children's
mental health, with both major Australian
political parties expressing support for
stricter age verification laws.
#Facebook #Meta #AgeVerification
#ChildSafety #TechPolitics
10
Is Silicon Valley Building Universe 25?: This
article revisits John B. Calhoun's 1968
experiment known as Universe 25, where a
perfect society was created for mice.
Despite optimal conditions, the mice
society collapsed due to social
dysfunctions such as narcissism,
aggression, and disengagement. The
author draws parallels to modern human
society and warns about the implications
of tech-driven utopias created by Silicon
Valley.
#SiliconValley #Universe25 #AI
#Technology #Society
11
Facebook, Instagram are using your data –
and you can’t opt out: If you're using
Facebook or Instagram, Meta is employing
your data to enhance its AI models without
giving you the choice to opt out. While EU
users have the option to opt out due to
stricter privacy laws, Australian users don't
have this privilege. This has sparked
backlash and calls for stronger privacy laws
in Australia.
#Privacy #AI #DataProtection #Meta
#Facebook
12
What Policy Makers Need to Know About
AI (and What Goes Wrong if They Don’t):
Policy makers must grasp how AI works to
effectively regulate it. The article uses SB
1047 as a case study, highlighting the
differences between deployment and
release of AI models. It emphasizes that
regulating deployment, rather than release,
would avoid stifling open source innovation
and better align with safety goals.
#AI #Policy #Opensource
#TechRegulation #AIsafety
13
Google Gemini tried to kill me: After
attempting to infuse garlic into olive oil
without heating, a user discovered that tiny
carbonation bubbles indicated the growth
of a botulism culture, highlighting the
potential danger of this method. Prompt
with care and verify information, as this
process can be hazardous.
#foodSafety #botulism #garlic
#oliveOil #healthAlert
14
London premiere of movie with AI-
generated script cancelled after backlash:
The Prince Charles Cinema in London
cancelled the premiere of 'The Last
Screenwriter,' a film with a script generated
by ChatGPT 4.0, following a backlash from
their audience. The filmmakers intended it
as a contribution to the conversation
surrounding AI in scriptwriting, but received
200 complaints. Despite the cancellation, a
private screening for the cast and crew will
go ahead.
#AI #Film #ChatGPT #Screenwriting
#Cinema
15
Safe Superintelligence Inc.: Safe
Superintelligence Inc. (SSI) has been
established as the first lab dedicated solely
to developing safe superintelligence. The
company focuses on advancing
superintelligent capabilities while ensuring
safety measures remain a step ahead.
Located in Palo Alto and Tel Aviv, SSI is
recruiting top engineers and researchers to
tackle this monumental challenge.
#Superintelligence #AI
#TechInnovation #SafetyFirst
#FutureTech
16
Ask HN: Could AI be a dot com sized bubble?: The
recent enthusiasm for AI has drawn comparisons
to the dotcom bubble, with inflated stock prices
and hype around AI technologies driving
substantial investments. Some argue that while
AI’s long-term potential is significant, current
market behaviors resemble the speculative
frenzy of the dotcom era. Notably, Nvidia and
other tech giants are at the forefront but
concerns persist about the sustainability of
these high valuations and the possibility of
market corrections if near-term expectations
aren’t met. The discussion highlights both the
promise and potential pitfalls of today's AI boom.
#AI #TechBubble #Investment
#Speculation #Nvidia
17
Top 10 Generative AI Models Mimic
Russian Disinformation: A NewsGuard
audit found that top generative AI models,
including ChatGPT-4 and Google's Gemini,
repeated Russian disinformation narratives
32% of the time. These bots often cited
fake local news sites as authoritative
sources. The findings, amid the first AI-
influenced election year, reveal how easily
AI platforms can spread false information
despite efforts to prevent this misuse.
#AI #Disinformation
#RussianPropaganda #CyberSecurity
#Elections
18
Researchers describe how to tell if ChatGPT
is confabulating: Researchers from the
University of Oxford have developed a
method to identify when large language
models (LLMs) like ChatGPT are
confabulating, or providing false answers
with confidence. The approach, which
analyzes statistical uncertainty in responses,
could help mitigate the issue of AI giving
confidently incorrect answers by determining
when the AI is unsure of the correct answer
versus unsure of how to phrase it.
#AI #ChatGPT #Research
#Confabulation #LLM
19
What policy makers need to know about AI
(and what goes wrong if they don’t):
Understanding AI is crucial for policy makers
to avoid ineffective legislation. SB 1047,
currently being considered in California, aims
to regulate AI for safety but lacks the
necessary technical definitions, causing
potential issues. By focusing on the
difference between model 'release' and
'deployment', the article explains how current
legislative language could negatively impact
the open-source AI community and suggests
ways to improve it.
#AI #Policy #OpenSource #Legislation
#AIFuture
20
Citigroup says AI could replace more than half of
jobs in banking: Citigroup has issued a report
revealing that AI could automate over half of the
jobs in the banking sector, significantly transforming
consumer finance and enhancing productivity. The
bank notes that around 54% of banking roles have a
high likelihood of automation, with another 12%
potentially being augmented by AI technology.
The report underscores the growing
experimentation with AI by the world's largest
banks, driven by the potential to improve staff
productivity and reduce costs. This highlights a
major shift within the banking industry towards AI-
driven operations.
#AI #Banking #Automation #Finance
#Technology
21
LinkedIn's Cookie Consent Notice: Peter
Gostev, Head of AI at Moonpig, critiques
Forrester's Foundation Model assessment
for its confusing weightings and scores,
especially questioning the logic behind the
rankings for core capabilities and specific
models, such as IBM's and Anthropics'. He
highlights the overlap in categories and the
odd results for well-regarded open-source
models like Mistral.
#AI #Forrester #MachineLearning
#AIModels #TechAssessment
22
ChatBug: Tricking AI Models into Harmful
Responses: A recent research paper from the
University of Washington and the Allen Institute
for AI has highlighted a critical vulnerability in
Large Language Models (LLMs), including GPT,
Llama, and Claude. The study reveals that chat
templates used in instruction tuning can be
exploited through attacks like format mismatch
and message overflow, leading the models to
produce harmful responses. This vulnerability,
named ChatBug, was tested on several state-
of-the-art LLMs, revealing high susceptibility
and a need for improved safety measures.
#AI #LLM #CyberSecurity #Research
#TechNews
23
What is the biggest challenge in our
industry?: In this article, the author
addresses a crucial question regarding the
major challenges faced by the tech
industry. They highlight the anxiety and
mental health issues stemming from
layoffs and fears of AI replacing
programmers. The piece advises how
leaders can mitigate these anxieties
through open communication, positivity,
and leveraging new technologies
effectively.
#TechAnxiety #AI #Layoffs
#Leadership #MentalHealth
24
Thank you!
hello@matthewsinclair.com
matthewsinclair.com
masto.ai/@matthewsinclair
medium.com/@matthewsinclair
twitter.com/@matthewsinclair
Originally published on
quantumfaxmachine.com and Medium.
If you’d like to sign up for this content as an
email, click here to join the mailing list.
Image: Photo by Austin Chan on Unsplash
25

More Related Content

20240705 QFM024 Irresponsible AI Reading List June 2024

  • 1. Quantum Fax Machine QFM024: Irresponsible AI Reading List June 2024 quantumfaxmachine.com 1
  • 2. QFM024: Irresponsible AI Reading List June 2024 The June edition of the Irresponsible AI Reading List starts with a critique of AI's potential for exaggerated promises in science and academia then heads over to look at the societal implications of AI-driven job automation and privacy concerns. An ongoing theme emerges for the urgent need for critical, informed engagement with AI technologies. We then take a look at the role that AI plays in amplifying societal biases, privacy rights concerning AI profiling, and the mental health impacts of AI- induced job insecurity. The discussion on Meta's use of facial recognition for age verification highlights the tension between technological advancements and privacy rights, while Citigroup's report on AI's potential to automate over half of banking jobs illustrates the profound economic shifts underway in the job market. ChatBug: Tricking AI Models into Harmful Responses looks at how AI model vulnerabilities can lead to harmful exploitation, pointing to the critical need for even more robust cybersecurity measures when working with AI. As always, the Quantum Fax Machine Propellor Hat Key will guide your browsing. Enjoy! Key: : Nothing to do with AI, but interesting nonetheless : Mentions AI : Talks about irresponsible AI : Talks about irresponsible AI in a real-world failure scenario : Talks about technical details of irresponsible AI : Discusses technical details and mitigation of irresponsible AI Source: Photo by Thomas Kinto on Unsplash 2
  • 3. Scientists should use AI as a tool, not an oracle (aisnakeoil.com): This article discusses the hype surrounding AI and its misuse in scientific research, leading to flawed studies and reproducibility issues. It emphasises that AI should be used as a tool to aid human understanding rather than an infallible oracle, advocating for a cultural shift towards more critical and careful use of AI in scientific disciplines to ensure research integrity and reliability. #AI #Science #ResearchIntegrity #Reproducibility #MachineLearning 3
  • 4. Critique of Forrester's Foundation Model Assessment (linkedin.com): Peter Gostev critiques Forrester's Foundation Model assessment, highlighting issues with the weightings and scores of various AI models, pointing out inconsistencies and lack of specificity in the evaluation process. #AI #MachineLearning #ForresterReport #ModelAssessment #AIevaluation 4
  • 5. AI Is a False God -- (thewalrus.ca): Navneet Alang critiques the overhyped promises of AI, arguing that while it can enhance data processing and efficiency, it lacks the moral and subjective capacities necessary to solve deeply human problems, often exacerbating existing biases and societal issues. #AI #TechCritique #ArtificialIntelligence #EthicsInAI #TechnologyImpact 5
  • 6. What is the biggest challenge in our industry? (thrownewexception.com): The biggest challenge in the tech industry is the anxiety caused by layoffs and the fear of AI replacing jobs, leading to mental health issues like burnout. Leaders can help by fostering open communication, leading positively, leveraging new technologies, investing in continuous learning, and collaborating with HR to support their teams. #TechIndustry #AI #MentalHealth #Leadership #Layoffs 6
  • 7. The Right Not to Be Subjected to AI Profiling Based on Publicly Available Data—Privacy and the Exceptionalism of AI Profiling: This article argues for the new legal right to protect individuals from AI profiling based on publicly available data without their explicit informed consent. It develops three primary arguments dealing with social control, stigmatisation, and the unique threat posed by AI profiling compared to other data processing methods. The article suggests that existing GDPR regulations are not sufficient and calls for explicit regulation with a sui generis right for protection. #AI #Privacy #Profiling #TechLaw #DataProtection 7
  • 8. Don't Expect Juniors to Teach Senior Professionals to Use Generative AI: Emerging Technology Risks and Novice AI Risk Mitigation Tactics: A Harvard Business Working Paper explores the challenges and limitations of expecting junior professionals to guide senior professionals in the use of emerging technologies like generative AI. The study, conducted with Boston Consulting Group, included interviews with junior consultants using GPT-4 and found that they often lack deep understanding and experience, making them ineffective in mitigating AI risks at a senior level. Insights suggest the need for more seasoned strategies and mitigation tactics focusing on system design and ecosystem-level changes. #AI #EmergingTech #GenerativeAI #TechRisks #Innovation 8
  • 9. How a Single ChatGPT Mistake Cost Us $10,000+: In the early days of a startup, a critical mistake involving ChatGPT- generated code cost over $10,000 in lost sales. The problem stemmed from a single hardcoded ID that caused unique ID collisions, preventing new users from subscribing. This story highlights the importance of robust testing and the perils of copy-pasting code in production environments. #startup #ChatGPT #codingmistakes #techerrors #softwaredevelopment 9
  • 10. Meta Uses Facial Recognition for Age Verification Amid Political Pressure: Meta is now using facial recognition to verify the age of some users on Facebook and Instagram. This move comes amid growing political pressure to protect children's mental health, with both major Australian political parties expressing support for stricter age verification laws. #Facebook #Meta #AgeVerification #ChildSafety #TechPolitics 10
  • 11. Is Silicon Valley Building Universe 25?: This article revisits John B. Calhoun's 1968 experiment known as Universe 25, where a perfect society was created for mice. Despite optimal conditions, the mice society collapsed due to social dysfunctions such as narcissism, aggression, and disengagement. The author draws parallels to modern human society and warns about the implications of tech-driven utopias created by Silicon Valley. #SiliconValley #Universe25 #AI #Technology #Society 11
  • 12. Facebook, Instagram are using your data – and you can’t opt out: If you're using Facebook or Instagram, Meta is employing your data to enhance its AI models without giving you the choice to opt out. While EU users have the option to opt out due to stricter privacy laws, Australian users don't have this privilege. This has sparked backlash and calls for stronger privacy laws in Australia. #Privacy #AI #DataProtection #Meta #Facebook 12
  • 13. What Policy Makers Need to Know About AI (and What Goes Wrong if They Don’t): Policy makers must grasp how AI works to effectively regulate it. The article uses SB 1047 as a case study, highlighting the differences between deployment and release of AI models. It emphasizes that regulating deployment, rather than release, would avoid stifling open source innovation and better align with safety goals. #AI #Policy #Opensource #TechRegulation #AIsafety 13
  • 14. Google Gemini tried to kill me: After attempting to infuse garlic into olive oil without heating, a user discovered that tiny carbonation bubbles indicated the growth of a botulism culture, highlighting the potential danger of this method. Prompt with care and verify information, as this process can be hazardous. #foodSafety #botulism #garlic #oliveOil #healthAlert 14
  • 15. London premiere of movie with AI- generated script cancelled after backlash: The Prince Charles Cinema in London cancelled the premiere of 'The Last Screenwriter,' a film with a script generated by ChatGPT 4.0, following a backlash from their audience. The filmmakers intended it as a contribution to the conversation surrounding AI in scriptwriting, but received 200 complaints. Despite the cancellation, a private screening for the cast and crew will go ahead. #AI #Film #ChatGPT #Screenwriting #Cinema 15
  • 16. Safe Superintelligence Inc.: Safe Superintelligence Inc. (SSI) has been established as the first lab dedicated solely to developing safe superintelligence. The company focuses on advancing superintelligent capabilities while ensuring safety measures remain a step ahead. Located in Palo Alto and Tel Aviv, SSI is recruiting top engineers and researchers to tackle this monumental challenge. #Superintelligence #AI #TechInnovation #SafetyFirst #FutureTech 16
  • 17. Ask HN: Could AI be a dot com sized bubble?: The recent enthusiasm for AI has drawn comparisons to the dotcom bubble, with inflated stock prices and hype around AI technologies driving substantial investments. Some argue that while AI’s long-term potential is significant, current market behaviors resemble the speculative frenzy of the dotcom era. Notably, Nvidia and other tech giants are at the forefront but concerns persist about the sustainability of these high valuations and the possibility of market corrections if near-term expectations aren’t met. The discussion highlights both the promise and potential pitfalls of today's AI boom. #AI #TechBubble #Investment #Speculation #Nvidia 17
  • 18. Top 10 Generative AI Models Mimic Russian Disinformation: A NewsGuard audit found that top generative AI models, including ChatGPT-4 and Google's Gemini, repeated Russian disinformation narratives 32% of the time. These bots often cited fake local news sites as authoritative sources. The findings, amid the first AI- influenced election year, reveal how easily AI platforms can spread false information despite efforts to prevent this misuse. #AI #Disinformation #RussianPropaganda #CyberSecurity #Elections 18
  • 19. Researchers describe how to tell if ChatGPT is confabulating: Researchers from the University of Oxford have developed a method to identify when large language models (LLMs) like ChatGPT are confabulating, or providing false answers with confidence. The approach, which analyzes statistical uncertainty in responses, could help mitigate the issue of AI giving confidently incorrect answers by determining when the AI is unsure of the correct answer versus unsure of how to phrase it. #AI #ChatGPT #Research #Confabulation #LLM 19
  • 20. What policy makers need to know about AI (and what goes wrong if they don’t): Understanding AI is crucial for policy makers to avoid ineffective legislation. SB 1047, currently being considered in California, aims to regulate AI for safety but lacks the necessary technical definitions, causing potential issues. By focusing on the difference between model 'release' and 'deployment', the article explains how current legislative language could negatively impact the open-source AI community and suggests ways to improve it. #AI #Policy #OpenSource #Legislation #AIFuture 20
  • 21. Citigroup says AI could replace more than half of jobs in banking: Citigroup has issued a report revealing that AI could automate over half of the jobs in the banking sector, significantly transforming consumer finance and enhancing productivity. The bank notes that around 54% of banking roles have a high likelihood of automation, with another 12% potentially being augmented by AI technology. The report underscores the growing experimentation with AI by the world's largest banks, driven by the potential to improve staff productivity and reduce costs. This highlights a major shift within the banking industry towards AI- driven operations. #AI #Banking #Automation #Finance #Technology 21
  • 22. LinkedIn's Cookie Consent Notice: Peter Gostev, Head of AI at Moonpig, critiques Forrester's Foundation Model assessment for its confusing weightings and scores, especially questioning the logic behind the rankings for core capabilities and specific models, such as IBM's and Anthropics'. He highlights the overlap in categories and the odd results for well-regarded open-source models like Mistral. #AI #Forrester #MachineLearning #AIModels #TechAssessment 22
  • 23. ChatBug: Tricking AI Models into Harmful Responses: A recent research paper from the University of Washington and the Allen Institute for AI has highlighted a critical vulnerability in Large Language Models (LLMs), including GPT, Llama, and Claude. The study reveals that chat templates used in instruction tuning can be exploited through attacks like format mismatch and message overflow, leading the models to produce harmful responses. This vulnerability, named ChatBug, was tested on several state- of-the-art LLMs, revealing high susceptibility and a need for improved safety measures. #AI #LLM #CyberSecurity #Research #TechNews 23
  • 24. What is the biggest challenge in our industry?: In this article, the author addresses a crucial question regarding the major challenges faced by the tech industry. They highlight the anxiety and mental health issues stemming from layoffs and fears of AI replacing programmers. The piece advises how leaders can mitigate these anxieties through open communication, positivity, and leveraging new technologies effectively. #TechAnxiety #AI #Layoffs #Leadership #MentalHealth 24
  • 25. Thank you! hello@matthewsinclair.com matthewsinclair.com masto.ai/@matthewsinclair medium.com/@matthewsinclair twitter.com/@matthewsinclair Originally published on quantumfaxmachine.com and Medium. If you’d like to sign up for this content as an email, click here to join the mailing list. Image: Photo by Austin Chan on Unsplash 25