Should artificial intelligence be used to improve decision-making in the court of law? According to a new working paper posted to the arXiv preprint server, not only does one example of an AI algorithm fail to improve the accuracy of judicial calls, on its own the technology fares worse than humans.
“A lot of researchers have focused on whether the algorithm has a bias or AI has a bias,” noted co-author Kosuke Imai, professor of government and statistics. “What they haven’t really looked at is how the use of AI affects the human decision.”
While several sectors, including criminal justice, medicine, and even business, use AI recommendations, humans are typically the final decision-makers. The researchers took this into account by comparing criminal bail decisions made by a single judge with recommendations generated by an AI system. The researchers specifically analyzed AI’s influence on whether cash bail should be imposed.
The randomized controlled trial was conducted in Dane County, Wisconsin, focusing on whether arrestees were released on their own recognizance or subjected to cash bail. The researchers—led by Imai and Jim Greiner, the Honorable S. William Green Professor of Public Law at Harvard Law School—set their sights on hearings held by a single judge over a 30-month period, between the middle of 2017 and the end of 2019. Also analyzed were arrest data on defendants for up to 24 months later.
Results showed that AI alone performed worse than the judge in predicting reoffenders—in this case, by imposing the tighter restriction of cash bail. At the same time, little to no difference was found between the accuracy of human-alone and AI-assisted decision-making. The judge went against AI recommendations in slightly more than 30% of cases.
“I was surprised by this,” Greiner said. “Given the evidence that we’ve cited that algorithms can sometimes outperform human decisions, it looked as though what happened is that this algorithm had been set to be too harsh. It was over-predicting that the arrestees would misbehave, predicting that they would do so too often, and, therefore, recommending measures that were too harsh.”
This issue could be fixed by recalibrating the algorithm, the professors argued.
“It’s a lot easier to understand and then fix the algorithm or AI than the human,” Imai said. “It’s a lot harder to change the human or understand why humans make their decisions.”
The AI studied here did not specifically account for race, instead focusing on age and nine factors related to past criminal experience. Imai, an expert on deploying statistical modeling to call out racial gerrymandering, attributed inequities concerning cash bail to a variety of societal factors, particularly relating to criminal history.
He acknowledged that the study’s findings may be cause for concern, but he noted that people are biased as well. “The advantage of AI or an algorithm is that it can be made transparent,” he said. The key is to have open-source AI that is readily available for empirical evaluation and analysis.
The way the criminal justice system is currently using AI as well as unguided human decisions should be studied with an eye to making improvements, Greiner said. “I don’t know whether this is comforting,” he offered, “but my reaction for folks who are afraid or skeptical of AI is to be afraid and skeptical of AI, but to be potentially more afraid or skeptical of unguided human decisions.”
He added that the way the criminal justice system is currently using AI as well as unguided human decisions should be studied to make improvements.
More information:
Eli Ben-Michael et al, Does AI help humans make better decisions? A methodological framework for experimental evaluation, arXiv (2024). DOI: 10.48550/arxiv.2403.12108
Harvard Gazette
This story is published courtesy of the Harvard Gazette, Harvard University’s official newspaper. For additional university news, visit Harvard.edu.
Does AI help humans make better decisions? One judge’s track record—with and without algorithm—surprises researchers (2024, June 17)
retrieved 17 June 2024
from https://techxplore.com/news/2024-06-ai-humans-decisions-track-algorithm.html
part may be reproduced without the written permission. The content is provided for information purposes only.
Should artificial intelligence be used to improve decision-making in the court of law? According to a new working paper posted to the arXiv preprint server, not only does one example of an AI algorithm fail to improve the accuracy of judicial calls, on its own the technology fares worse than humans.
“A lot of researchers have focused on whether the algorithm has a bias or AI has a bias,” noted co-author Kosuke Imai, professor of government and statistics. “What they haven’t really looked at is how the use of AI affects the human decision.”
While several sectors, including criminal justice, medicine, and even business, use AI recommendations, humans are typically the final decision-makers. The researchers took this into account by comparing criminal bail decisions made by a single judge with recommendations generated by an AI system. The researchers specifically analyzed AI’s influence on whether cash bail should be imposed.
The randomized controlled trial was conducted in Dane County, Wisconsin, focusing on whether arrestees were released on their own recognizance or subjected to cash bail. The researchers—led by Imai and Jim Greiner, the Honorable S. William Green Professor of Public Law at Harvard Law School—set their sights on hearings held by a single judge over a 30-month period, between the middle of 2017 and the end of 2019. Also analyzed were arrest data on defendants for up to 24 months later.
Results showed that AI alone performed worse than the judge in predicting reoffenders—in this case, by imposing the tighter restriction of cash bail. At the same time, little to no difference was found between the accuracy of human-alone and AI-assisted decision-making. The judge went against AI recommendations in slightly more than 30% of cases.
“I was surprised by this,” Greiner said. “Given the evidence that we’ve cited that algorithms can sometimes outperform human decisions, it looked as though what happened is that this algorithm had been set to be too harsh. It was over-predicting that the arrestees would misbehave, predicting that they would do so too often, and, therefore, recommending measures that were too harsh.”
This issue could be fixed by recalibrating the algorithm, the professors argued.
“It’s a lot easier to understand and then fix the algorithm or AI than the human,” Imai said. “It’s a lot harder to change the human or understand why humans make their decisions.”
The AI studied here did not specifically account for race, instead focusing on age and nine factors related to past criminal experience. Imai, an expert on deploying statistical modeling to call out racial gerrymandering, attributed inequities concerning cash bail to a variety of societal factors, particularly relating to criminal history.
He acknowledged that the study’s findings may be cause for concern, but he noted that people are biased as well. “The advantage of AI or an algorithm is that it can be made transparent,” he said. The key is to have open-source AI that is readily available for empirical evaluation and analysis.
The way the criminal justice system is currently using AI as well as unguided human decisions should be studied with an eye to making improvements, Greiner said. “I don’t know whether this is comforting,” he offered, “but my reaction for folks who are afraid or skeptical of AI is to be afraid and skeptical of AI, but to be potentially more afraid or skeptical of unguided human decisions.”
He added that the way the criminal justice system is currently using AI as well as unguided human decisions should be studied to make improvements.
More information:
Eli Ben-Michael et al, Does AI help humans make better decisions? A methodological framework for experimental evaluation, arXiv (2024). DOI: 10.48550/arxiv.2403.12108
Harvard Gazette
This story is published courtesy of the Harvard Gazette, Harvard University’s official newspaper. For additional university news, visit Harvard.edu.
Does AI help humans make better decisions? One judge’s track record—with and without algorithm—surprises researchers (2024, June 17)
retrieved 17 June 2024
from https://techxplore.com/news/2024-06-ai-humans-decisions-track-algorithm.html
part may be reproduced without the written permission. The content is provided for information purposes only.