My name isJiahao Chen, and I’m thrilled to announce theformation ofResponsible Artificial Intelligence LLC, my new consultancy focused on practical, yet responsible AI solutions for the evolving needs of businesses and regulators that is informed by the very latest academic research.
About this newsletter.The research and practice of responsible AI is evolving rapidly. Every month, new laws are passed, existing regulations are updated to cover AI/ML systems, and new policy and AI research continues to be published at an exponential rate. I’m pleased to offer this newsletter summarizing my reading of the latest developments in academic research, government policy, and community events.
Solicitation. Please get in touch if you’re interested in consultancy services for designing and operating responsible AI solutions, algorithmic auditing services, or expert witness testimony.
Branding note.The logo and cover image were generated withBen Firschman’s fork of Stable Diffusion, which runs locally on a M1 Mac. Interest in generative models has exploded in this past month, with the open source release of Stable Diffusion (GitHub;HuggingFace) byStability.AI(CVPR’22,arXiv:2112.10752), and a comprehensive survey preprint of diffusion models (arXiv:2209.00796).
Generative models have shown sensitivity of the output to the specific text, even down to the level of word order and punctuation. Some recent work even proposes that the problem of finding suitable text to input to Stable Diffusion itself is amenable to machine learning, resulting in work onmeta-prompting(arXiv:2209.11486;arXiv:2209.11711). My attempts to generate usable images occasionally resulted in rickrolling in the form of still frames from the music video.
At the same time, Stable Diffusion harbors the risk of amplifying NSFW content, stereotypes, and slurs (arXiv:2110.01963). The training data set was subsetted fromLAION-5B, consisting of 5.8 billion annotated images. A previous version of this data set,LAION-400M, was found to include not only copyrighted materials, but also problematic content like pornography, sexual violence, racial stereotypes, and ethnic slurs (arXiv:2110.01963). Such problematic content was not removed in LAION-5B, but purports to be tagged as such (NeurIPS’22). It’s unclear if Stable Diffusion’s training subset has such problematic content removed.
☛ Are you part of Stable Diffusion’s training data? Search the LAION-5B data set on haveibeentrained.com. (CW: NSFW, violence)
☛ Have a Mac with Apple Silicon? Try Stable Diffusion on your local machine with Diffusion Bee.
The pseudotext in the logo image is a poorly understood phenomenonstudied previouslyinDALL·E 2(arXiv:2204.06125,arXiv:2206.00169). What do you think the text means?
It’s impossible for any single person to give a comprehensive review of a rapidly evolving area like responsible AI, with multiple developments in governments worldwide and new research being published every day. Instead, here’s a summary of what I’ve been reading in the past month which may be of interest to readers.
This newsletter may be truncated by your e-mail provider, in which case you can simply read it online.
TheDigital Regulation Cooperation Forum (DRCF)have updatedtheir response to public submissionson their two discussion papers on algorithmic processing and algorithmic auditing. This forum between four different UK regulatory bodiesaimsto support improvements in algorithmic transparency while enabling innovation in their regulated industries. They continue to welcome feedback on the two discussion papers.
☛ Send the Forum your feedback at drcf.algorithms@cma.gov.uk.
The public comment period for theBank of England’s consultation paper,Model risk management principles for banks(CP6/22), ends on October 21. In particular, they are interested to know if these prudential regulations sufficiently “identify, manage, monitor, and control the risks associated with AI or ML models” (¶¶ 2.18-20).
☛ Send the Bank of England your feedback at CP6_22@bankofengland.co.uk.
TheU.S. Congressreleased their new beta API(api.congress.gov), enabling developers to retrieve legislative documents like bills and committee reports programmatically, and followdiscussions of AI in the current sessionsuch as theAI Training Act, theAlgorithmic Accountability Act, and theAdvancing American AI Innovation Act.
☛ Sign up for a Congress.gov API key here.
[FR] The Federal Communications Commission’s Technical Advisory Committeeheld a meeting on September 15which was videorecorded and publicly viewable. Starting at 2:20 of the recording, the AI/ML Working Group presented their vision of AI/ML as deeply imbedded into the new 6G network infrastructure across its entire lifecycle, especially in design (digital twins and resource allocation) and operations (NLP and MLOps). Natural language processing was of particular interest for its implications for existing laws on wiretapping and robocalling.
[FR] Notice of meeting. TheNational Artificial Intelligence Advisory Committee (NAIAC)will be holding a public meeting on October 12-13 (Stanford; hybrid). The notice claims the agenda will be published on theNAIAC website.
☛ Submit comments and questions by October 5, 5pm PDTto Melissa Banner at melissa.banner@nist.gov.
[FR] Request for Information. TheDepartment of Transportationis seeking input on the use of AI in autonomous vehicles and its implications for pedestrian and bicyclist safety.
[FR] Request for Comment. The Department of Commerce “is requesting public comments to gain insight on the current global AI market and stakeholder concerns regarding international AI policies, regulations, and other measures which may impact U.S. exports of AI technologies.”
The U.S.National Institute of Standards and Technology (NIST)has published a second version of itsAI risk management framework(pdf), ahead of itsthird workshop on October 18-19(virtual;free registration).
☛ Register for the NIST workshop for freehere.
[FR] Advance notice of proposed rulemaking. The U.S. Federal Trade Commission (FTC) is requesting “public comment on the prevalence of commercial surveillance and data security practices that harm consumers.” Commissioner Noah Joshua Phillips is particularly concerned with whether AI can amplify disparate outcomes, and if so, if the FTC “shouldput the kibosh on the development of artificial intelligence.” (emphasis mine)
New York City’s Department of Consumer and Worker Protectionhas published itsProposed Rulesfor compliance with the City’s law onautomated employment decision tools (AEDT;Int. No. 1894-2020-A), which comes into effect in the new year. Prominent law firm Debevoise & Plimpton have summarizedtheir reading of the Proposed Rules, highlighting the intersectionality concerns of testing beyond race, gender, and age in isolation, and the remaining gaps such as the limitations of metrics and data availability. DCWP is soliciting public comments ahead thepublic hearing onOctober 24at 11am EDT(Zoom link).
☛ Provide your comments on AEDT to NYC here.
Disclosure: Responsible AI LLC was not paid to advertise these events.
Credo AIis organizing aGlobal Responsible AI Summiton October 27 (online), with a speaker list representing a mix of UK and US companies, policymakers and researchers.
A new pirate-themedBias Bountywill be announced later this month atCAMLIS’22(October 21-22 in Arlington, VA;$200 paid registration). The team behind this bias bounty include several people who were previously behindTwitter’s algorithmic bias bounty challengein 2021.
☛ Sign up for the bias bounty here.
Here’s a briefly annotated reading list of new papers and preprints that were published or updated in the past month.
[arXiv:2203.16432;blog; AIES’22] Nil-Jana Akpinaret al.,Long-term Dynamics of Fairness Intervention in Connection Recommender Systems. Using an urn model for recommender systems, the authors prove that interventions must be dynamically applied over time to achieve a desired fairness outcome, and that remediation at a single point in time is insufficient in general.
[arXiv:2206.01254; NeurIPS’22;Twitter thread 1and2] Tessa Han, Suraj Srinivas, and Himabindu Lakkaraju,Which Explanation Should I Choose? A Function Approximation Perspective to Characterizing Post hoc Explanations. Introduces a “no free lunch”-type theorem which quantifies the ultimate limitation of post hoc explanation methods that attempt to build best local approximations to the decision function.
[arXiv:2206.11104; NeurIPS’22;OpenReview;project site] Chirag Agarwalet al.,OpenXAI: Towards a Transparent Evaluation of Model Explanations. This Python package provides a uniform evaluation environment for explanation methods, using metrics based on faithfulness, stability and fairness. They also provide several synthetic data generators for evaluating whether methods correctly capture ground truth, which is almost always unknown in real data sets.
[arXiv:2207.04154;GitHub;Twitter thread] Dylan Slacket al.,TalkToModel: Explaining Machine Learning Models with Interactive Natural Language Conversations. Describes a natural language system to query models for explanations of its predictions using plain English, which was favorably received by medical professionals.
[arXiv:2107.07511;Twitter thread] Anastasios N. Angelopoulos and Stephen Bates,A Gentle Introduction to Conformal Prediction and Distribution-Free Uncertainty Quantification. A tutorial to conformal prediction, a method for refining heuristic error bars on the output of AI systems into more statistically rigorous uncertainty bounds. Conformal prediction has seen renewed interest, judging fromrecent ICLR submissions.
[doi:10.1145/3514094.3534167, AIES’22;preprint;Twitter thread] Max Langenkamp and Daniel N. Yue,How Open Source Machine Learning Software Shapes AI. Estimates that open source software in ML produces $30 billion/year of value, albeit with a steep upfront development cost of $120-180 million/year for TensorFlow and $180-300 million/year for PyTorch, which leaves development of non-deep learning software severely underfunded by comparison. User interviews highlight traditional benefits of open sources, such as community engagement, while noting a convergence in APIs and capabilities.
[arXiv:2209.02836; ECCV’22;project page] Vongani H. Malulekeet al.,Studying Bias in GANs through the Lens of Race. Shows that truncation, a technique used to control output quality in GANs, reduces racial diversity in generated images by triggering mode collapse. In my opinion, this phenomenon is emblematic of a general issue of diversity loss resulting from compressing a learned representation.
[arXiv:2209.07572; EAAMO’22] Abeba Birhaneet al.,Power to the People? Opportunities and Challenges for Participatory AI.Reviews mass participation efforts in AI, focusing on case studies ofMasakhane(NLP for African languages),Te Hiku Media(NLP dataMāoriwith a community-focusedlicense), and theData Cards Playbook(for developing data-centric documentation).
[arXiv:2209.09125;Twitter thread] Shreya Shankaret al.,Operationalizing Machine Learning: An Interview Study. Highlights the experimental nature of ML development as a recurring pain point in industry, leading to a tension between fast development and deployment on one hand, and validating correctness of results on the other.
[arXiv:2209.09780] David Gray Widder and Dawn Nafus,Dislocated Accountabilities in the AI Supply Chain: Modularity and Developers' Notions of Responsibility. Interviews with 27 AI practitioners reveals that modularity in tech design, while useful for separation of concerns as a best practice for software engineering, results in a disconnect in their accountability for algorithmic harms by reducing visibility of how their AI work is used in the larger system.
[SSRN:4217148]Mehtab Khan and Alex Hanna,The Subjects and Stages of AI Dataset Development: A Framework for Dataset Accountability(September 13, 2022). Comprehensive survey of legal concerns (such as copyright, privacy, and algorithmic harms) that surround the collection of training data for foundational models.
[report] Alec Radfordet al.,Robust Speech Recognition via Large-Scale Weak Supervision.This report announces OpenAI’s new Transformer-basedWhispermodel for multilingual transcription and translation. See also the summary Twitter threadshere. Preliminary reports on Twitter document some limitations, such as underperformance relative to other ASR modelson several metrics, as well assevere transcription errors in under-resourced languages like Bengali and Kannada.
☛ Try out Whisper on your own audio data using Google Colab notebooks such asthis one.
☛ Do you have a recommendation for what to read? Comment below to be considered for the next newsletter.
Solicitation. Please get in touch if you’re interested in consultancy services for designing and operating responsible AI solutions, algorithmic auditing services, or expert witness testimony.