- Last week, the New York City Council passed a bill designed to tackle “algorithmic inequity” in city government and address if tech systems appear to discriminate against people based on a variety of statuses including age, race, religion, gender or sexual orientation. The bill now awaits a signature from Mayor Bill de Blasio.
- The bill summary notes it would “require the creation of a task force that provides recommendations on how information on agency automated decision systems may be shared with the public and how agencies may address instances where people are harmed by agency automated decision systems.”
- ProPublica reports that Council Member James Vacca, the bill’s sponsor, was inspired to introduce the bill after a May 2016 investigation into machine bias.
“Biases in AI systems can be introduced during the programming itself, as a system learns or during application from a variety of sources. A lack of diversity among the programmers making the system and biased data sets can cause these problems,” CIO Dive reports.
Additionally, a recent story of failed iPhone X facial recognition in China sparked the question of diversity in tech talent, and called on tech developers to ensure equal representation of all backgrounds when writing code or developing functionalities. While unrelated to this specific case study, New York’s bill presents an opportunity for cities to be proactive about diversity in tech and to design algorithms that are beneficial for all city residents, despite their race, gender, sexual orientation or other status.
If New York’s bill is passed into law, it could set a precedent for cities across the U.S. to prioritize appropriate development and monitoring of AI systems — a pressing concern, considering the influx of AI-controlled devices and systems being rolled out nationwide.
Recommended Reading:
The New York City Council Int 1696-2017
ProPublicaNew York City Moves to Create Accountability for Algorithms
The algorithms that play increasingly central roles in our lives often emanate from Silicon Valley, but the effort to hold them accountable may have another epicenter: New York City. Last week, the New York City Council unanimously passed a bill to tackle algorithmic discrimination — the first measure of its kind in the country.
The algorithmic accountability bill, waiting to be signed into law by Mayor Bill de Blasio, establishes a task force that will study how city agencies use algorithms to make decisions that affect New Yorkers’ lives, and whether any of the systems appear to discriminate against people based on age, race, religion, gender, sexual orientation or citizenship status. The task force’s report will also explore how to make these decision-making processes understandable to the public.
The bill’s sponsor, Council Member James Vacca, said he was inspired by ProPublica’s investigation into racially biased algorithms used to assess the criminal risk of defendants.
“My ambition here is transparency, as well as accountability,” Vacca said.
A previous, more sweeping version of the bill had mandated that city agencies publish the source code of all algorithms being used for “targeting services” or “imposing penalties upon persons or policing” and to make them available for “self-testing” by the public. At a hearing at City Hall in October, representatives from the mayor’s office expressed concerns that this mandate would threaten New Yorkers’ privacy and the government’s cybersecurity.
The bill was one of two moves the City Council made last week concerning algorithms. On Thursday, the committees on health and public safety held a hearing on the city’s forensic methods, including controversial tools that the chief medical examiner’s office crime lab has used for difficult-to-analyze samples of DNA.
As a ProPublica/New York Times investigation detailed in September, an algorithm created by the lab for complex DNA samples has been called into question by scientific experts and former crime lab employees.
The software, called the Forensic Statistical Tool, or FST, has never been adopted by any other lab in the country.
Council Member Corey Johnson, chair of the health committee, quoted two key findings of our investigation: that FST’s inventors had acknowledged a margin of error of 30 percent for one key input of the program, and that the program could not take into consideration that family members might share DNA.
New York City no longer uses the tool for new cases. But officials at the hearing said they saw no need to revisit the thousands of criminal cases that relied on the technique in years past.
“Would you be open to reviewing cases in which testing was done on very small mixtures, or do you feel totally confident in all of the methods and science that were used on every case that’s come through your lab?” Johnson asked the officials.
“We are totally confident,” answered Dr. Barbara Sampson, the city’s chief medical examiner.
The algorithm’s source code was a closely held secret for years until a federal judge granted a motion filed by ProPublica to lift a protective order on it in October. We then published the code.

Defense attorneys testified at the hearing, criticizing the medical examiner’s office for what they saw as a dangerous lack of transparency in the development of its DNA tools.
Some had joined together to write to the state’s inspector general in September, demanding an investigation into the lab and a review of past cases. The inspector general, Catherine Leahy Scott, has not yet indicated whether she will pursue it. Meanwhile, the New York State Commission on Forensic Science, which oversees the use of forensic methods in the state’s labs, has discussed the criticisms in executive session meetings. Those sessions are closed to the public and commission members are prohibited from speaking about them.
After the hearing, Johnson said he was concerned by the discrepancies between the medical examiner’s testimony and that of advocates and intended to explore it further.
“This is a very, very, very important issue, and we have to ensure that methods that are used are scientifically sound, validated in appropriate ways, transparent to the public and to defense counsels, and ensure greater trust in the justice system,” said Johnson. “And I think that is what, hopefully, we can achieve, through asking more questions — and potentially thinking about legislation in the future.”