
AI-SOP'S FABLES
ANDREA M. MATWYSHYN
MIRANDA K. MOWBRAY
Introduction
CHAPTER 1: GRIM(M) TALES OF AI
-
What is AI?
-
Margaret Boden taxonomy
-
The Current Models
-
SuperAGI and AGI
-
GAI
-
Narrow AI
-
-
-
Three Core Problems
1. old – security generally and tech debt
2. aggravated – transparency and data quality, attack automation
3. new(ish) – speed, degree of credible impersonation, AI cluster(ducks)
-
“Hallucination”
-
“Intelligence”
-
Theories of intelligence
-
Instrumentalist limited readings
-
3.Three Core Questions
1. Functions
1. is the particular system seeking to replace human labor?
2. is the particular system seeking to replace some human thinking?
3. is the particular system seeking to replace humans?
2. Supervision
1. Supervised
2. Unsupervised
3. Reinforcement
4. Combination
3. Applications
CHAPTER 2: ONCE UPON A(I) TIME...
-
The Terrible Toads: Context Specificity
-
Recursive Hamsters
-
Rabbit Fence
-
Data quality
-
-
The Digesting Duck: Harm
-
The Great Emu War
-
Defining “Autonomy”
-
-
The Argle-Bargle Bears: Intent
-
Goldilocks from three perspectives
-
Bench bear v. Lasagna bear v. roof bear v. revenge bear v. cocaine bear
-
Berenstain Bears
-
Degree of automation and reversal/ unlearning ability
-
CHAPTER 3:
-
Lessons for Future AI
-
Context
-
Suitability is fact-specific, e.g. comedy
-
Feedback problems
-
Mushroom murder books
-
Spy Whales and dolphins, predictive octopi, bomb/mine sniffing rats
-
Beavers and inconvenient dams
-
-
Harm
-
Autonomy issues in cars – hardware, training data (incorrect, inadequate), wrong generalization, faulty distribution, fear, screaming, loss of trust. Disconnect with normies, ignoring DOT
-
Tay – prompt engineering: abusing systems but as built – a problematic repurposing, not a criminal act of computer intrusion even though the poor build can be a security issue
-
Security issues –
-
Training data manipulation by attackers or insiders
-
Lack of model archiving and logging
-
Lack of inspection ability
-
Lack of capabilities for “unlearning”
-
Lack of transparency on which training data (and which version) is included in proprietary products
-
Data quality deficits leading to “spicy” prompt injection opportunities
-
Sampling
-
Labeling
-
Proxy variable issues
-
Bias magnification
-
Secret whitelists/blacklists
-
NLP problems
-
Sensor data problems from hardware and/or design
-
Dependency/Relying on third parties: There was recently, a particularly problematic discovery of a rogue insider in an open source application in the last couple of weeks that set the computer security world ablaze with discussion and concern because of its impact on Linux systems. And of course, some machine learning algorithms use linux in some component of their design operation or deployment. So we are living in a very interconnected world, where the public sector and the private sector interrelate, and we have flaws both into the technologies as built, but also in the governance that controls them.
-
-
Same problems of internal controls and supervision that previously existed but even more complex, e.g. supply chain inclusion
-
Legally problematic use
-
And more…
-
-
Evolutionary/emergent nature of harm and diagnosing it
-
Dogs and orangutans driving etc. – self-preservation instinct overrides
-
-
Intent
-
Goal specification (can be hybrid): money versus psych (ego/human replacement/ curiosity) versus malice
-
Anthropomorphism of animals, e.g. love
-
Ambiguity in line between malice and incompetence
-
Corvids – disabling anticorvid tech – messy intent
-
Squirrels and categories of adversarial attacks on AI, e.g. perturbation – don’t require high “intelligence” just persistence
-
-
CHAPTER 4:
-
The Path Forward
-
Whose viewpoint? Animal reality versus our imagined versions of them wo understanding the reality of their limitation and skills
-
Cats – interaction effects re: anthropomorphism and obsession but see cat that brought phone to man who fell on bathroom floor. Exceptional cases happen so needs to be fact-specific analysis
-
Bomb/land mine rats - Actual superiority – versus humans versus any current tech
-
Turing v. Lovelace
-
Humility
-
Avoiding shallow definitions uninformed by humanities and all fields of psychology - not just neuropsych.
-
Humans beat AI in human-machine symbiosis condition – DARPA grand challenge, Go experiment
-
Human-machine symbiosis requires trustworthiness backstops in law
-
-
-
Context Sensitivity
-
Dogs
-
Capabilities that can exceed human
-
Special purpose v. general risks/ benefits
-
Spectrum of kinds
-
-
Harm
-
Look to existing categories of human harms
-
Failure of a duty to warn/ disclosure
-
Failure to monitor
-
-
Failure to mitigate ongoing harm/ auditor-identified problems ß bolster clarity of this duty
-
Replicability and replication as core to security correction / stochastic choice
-
-
-
Intent/Knowledge
-
Mental state attributions are questions of fact – don’t need one to apply to the machine
-
Was it under your control?
-
Audit requirements – willful blindness not ok
-
Don’t need a machine theory of mind
-
Don’t need a machine theory of consciousness
-
Definitely don’t need machine personhood legal theory
-
CHAPTER 5:
-
Remedies: a set of human persons must always be accountable/responsible
-
Tags – labeling of AI to signal categories of errors and set baseline expectations of fallibility
-
Mowbray’s “Shevchenko Machine” – interrogated code must not lie about itself
-
VW defeat devices fraud enforcement and additional mandatory disclosure
-
-
Pedigrees – data provenance and data quality disclosures
-
Training and testing – audit; threat modeling
-
Leashes – don’t use humans as guinea pigs; lab control
-
Muzzles – oversight and restraint
-
Neutering and capture – unlearning; board and individual consequences
-
Animal Control – a new Bureau of Technology Safety to address emergent problems etc.
-
Conclusion