Quickly, if you apply for a constructing allow, name a few missed trash pickup or ask a query about your utility invoice, a synthetic intelligence system could assist resolve how rapidly you get a solution and whether or not your request counts as “pressing.” That’s the long run sketched out in Seattle’s 2025-26 Artificial Intelligence Plan, geared toward making Seattle a “accountable AI” chief.
The coverage does embrace safeguards. It’s making an attempt to ban a number of the riskiest makes use of of AI, requires “human within the loop” overview of automated selections and dictates that giant quantities of AI‑generated textual content in public paperwork ought to be disclosed.
However for the individuals who truly reside underneath these methods — the renters, small‑enterprise homeowners, metropolis employees and other people navigating housing or social providers — one factor is lacking: actual say over how AI is used of their names.
Together with the “Accountable AI” program, town established a brand new hackathon collection meant to carry group voices into the dialog. A weekend occasion co‑hosted with a expertise vendor, nevertheless, shouldn’t be the identical as giving the general public an ongoing say during which AI instruments are deployed, how they’re evaluated and when they need to be shut off.
Ask the typical Seattle resident, “The place is town already utilizing AI on me?” Most of us would don’t know. The plan mentions pilots round allow overview, buyer‑response instruments and accessibility options, however these particulars are buried in PDFs and information releases, not introduced in plain language the place folks truly reside: on payments, notices and straightforward‑to‑discover metropolis internet pages.
If Seattle is critical about accountable AI, step one ought to be complete transparency. We’d like a transparent, public “AI registry” that lists each system town is utilizing or piloting: what it does, what knowledge it makes use of, who oversees it and the way residents can choose out or attraction.
Second, we’d like governance that extends past Metropolis Corridor. The present plan leans closely on inner committees, employees coaching and steerage paperwork. Seattle ought to create a everlasting, impartial oversight physique, one thing nearer to a civilian overview board than a tech advisory panel, with members drawn from the communities most definitely to be affected by automated methods.
Third, the guarantees of “human within the loop” and “accountable AI” have to imply one thing concrete for an individual on the opposite finish of a choice. If an AI system helps type your allow utility, housing placement or eligibility for a program, you ought to be informed in plain language that AI was concerned, given a easy approach to ask for a human overview and entitled to an evidence you don’t want a pc‑science diploma to interpret.
There may be additionally the query of who bears the dangers. Seattle’s plan talks about innovation and effectivity. However when AI instruments misfire, it won’t be tech distributors or the mayor’s workplace ready on maintain or dropping a day’s wages to repair a mistake. It will likely be the grocery employee whose profit utility will get misplaced in an automatic queue; the small‑enterprise proprietor whose allow falls right into a black field; the elder whose criticism will get caught in chatbot limbo.
Seattle has an opportunity to set a nationwide normal and the Metropolis Council ought to strengthen this plan earlier than it hardens into the brand new regular. Create a public AI registry. Set up an impartial, resident‑led oversight physique. Assure clear rights to notification, clarification and human attraction. And report every year not what number of AI pilots have been launched, however how they affected actual folks.
If AI goes to assist run Seattle, the general public deserves greater than a seat at a hackathon; they deserve an actual say in the way it’s designed. And each time we hear “accountable AI,” we should always ask: accountable to whom?
