699 results for search: %EB%A7%88%EC%BC%80%ED%8C%85%EC%A0%84%EB%AC%B8%E2%96%B6%EC%B9%B4%ED%86%A1adgogo%E2%96%B6%EA%B5%90%ED%95%98%EB%8F%99%ED%83%80%ED%88%AC%E3%84%85%EB%A7%88%EC%BC%80%ED%8C%85%E2%94%BE%EC%A0%84%EB%AC%B8%E2%99%A1%EA%B5%90%ED%95%98%EB%8F%99%E5%A5%A9%ED%83%80%ED%88%AC%E6%AF%8Econductor/feed/content/india/privacy
Data ‘hashing’ improves estimate of the number of victims in databases
But while HRDAG’s estimate relied on the painstaking efforts of human workers to carefully weed out potential duplicate records, hashing with statistical estimation proved to be faster, easier and less expensive. The researchers said hashing also had the important advantage of a sharp confidence interval: The range of error is plus or minus 1,772, or less than 1 percent of the total number of victims.
“The big win from this method is that we can quickly calculate the probable number of unique elements in a dataset with many duplicates,” said Patrick Ball, HRDAG’s director of research. “We can do a lot with this estimate.”
Selection Bias and the Statistical Patterns of Mortality in Conflict.
Megan Price and Patrick Ball. 2015. Statistical Journal of the IAOS 31: 263–272. doi: 10.3233/SJI-150899. © IOS Press and the authors. All rights reserved. Creative Commons BY-NC-SA.
Colombia (eng)
Big Data, Selection Bias, and the Statistical Patterns of Mortality in Conflict
Megan Price and Patrick Ball (2014). SAIS Review of International Affairs © 2014 The Johns Hopkins University Press. This article first appeared in SAIS Review, Volume 34, Issue 1, Winter-Spring 2014, pages 9-20. All rights reserved.
Assessing Claims of Declining Lethal Violence in Colombia
Patrick Ball, Tamy Guberek, Daniel Guzmán, Amelia Hoover, and Meghan Lynch (2007). “Assessing Claims of Declining Lethal Violence in Colombia.” Benetech. Also available in Spanish – “Para Evaluar Afirmaciones Sobre la Reducción de la Violencia Letal en Colombia.”
The Case Against a Golden Key
Patrick Ball (2016). The case against a golden key. Foreign Affairs. September 14, 2016. ©2016 Council on Foreign Relations, Inc. All Rights Reserved.
Recognising Uncertainty in Statistics
In Responsible Data Reflection Story #7—from the Responsible Data Forum—work by HRDAG affiliates Anita Gohdes and Brian Root is cited extensively to make the point about how quantitative data are the result of numerous subjective human decisions. An excerpt: “The Human Rights Data Analysis Group are pioneering the way in collecting and analysing figures of killings in conflict in a responsible way, using multiple systems estimation.”
Weighting for the Guatemalan National Police Archive Sample: Unusual Challenges and Problems.”
Gary M. Shapiro, Daniel R. Guzmán, Paul Zador, Tamy Guberek, Megan E. Price, Kristian Lum (2009).“Weighting for the Guatemalan National Police Archive Sample: Unusual Challenges and Problems.”In JSM Proceedings, Survey Research Methods Section. Alexandria, VA: American Statistical Association.
Kosovo Data – Killings, Migrations and More
Pretrial Risk Assessment Tools
Sarah L. Desmarais and Evan M. Lowder (2019). Pretrial Risk Assessment Tools: A Primer for Judges, Prosecutors, and Defense Attorneys. Safety and Justice Challenge, February 2019. © 2019 Safety and Justice Challenge. <<HRDAG’s Kristian Lum and Tarak Shah served as Project Members and made significant contributions to the primer.>>
Announcing New HRDAG Advisory Board Member
Using Math and Science to Count Killings in Syria
To predict and serve?
Kristian Lum and William Isaac (2016). To predict and serve? Significance. October 10, 2016. © 2016 The Royal Statistical Society.
Setting the Record Straight on Predictive Policing and Race
William Isaac and Kristian Lum (2018). Setting the Record Straight on Predictive Policing and Race. In Justice Today. 3 January 2018. © 2018 In Justice Today / Medium.
verdata: An R package for analyzing data from the Truth Commission in Colombia
Maria Gargiulo, María Julia Durán, Paula Andrea Amado, and Patrick Ball (2024). verdata: An R package for analyzing data from the Truth Commission in Colombia. The Journal of Open Source Software. 6 January, 2024. 9(93), 5844, https://doi.org/10.21105/joss.05844. Creative Commons Attribution 4.0 International License.
Palantir Has Secretly Been Using New Orleans to Test Its Predictive Policing Technology
One of the researchers, a Michigan State PhD candidate named William Isaac, had not previously heard of New Orleans’ partnership with Palantir, but he recognized the data-mapping model at the heart of the program. “I think the data they’re using, there are serious questions about its predictive power. We’ve seen very little about its ability to forecast violent crime,” Isaac said.
Celebrating Women in Statistics
In her work on statistical issues in criminal justice, Lum has studied uses of predictive policing—machine learning models to predict who will commit future crime or where it will occur. In her work, she has demonstrated that if the training data encodes historical patterns of racially disparate enforcement, predictions from software trained with this data will reinforce and—in some cases—amplify this bias. She also currently works on statistical issues related to criminal “risk assessment” models used to inform judicial decision-making. As part of this thread, she has developed statistical methods for removing sensitive information from training data, guaranteeing “fair” predictions with respect to sensitive variables such as race and gender. Lum is active in the fairness, accountability, and transparency (FAT) community and serves on the steering committee of FAT, a conference that brings together researchers and practitioners interested in fairness, accountability, and transparency in socio-technical systems.
Quantifying Injustice
“In 2016, two researchers, the statistician Kristian Lum and the political scientist William Isaac, set out to measure the bias in predictive policing algorithms. They chose as their example a program called PredPol. … Lum and Isaac faced a conundrum: if official data on crimes is biased, how can you test a crime prediction model? To solve this technique, they turned to a technique used in statistics and machine learning called the synthetic population.”
The True Dangers of AI are Closer Than We Think
William Isaac is quoted.
Here’s how an AI tool may flag parents with disabilities
HRDAG contributed to work by the ACLU showing that a predictive tool used to guide responses to alleged child neglect may forever flag parents with disabilities. “These predictors have the effect of casting permanent suspicion and offer no means of recourse for families marked by these indicators,” according to the analysis from researchers at the ACLU and the nonprofit Human Rights Data Analysis Group. “They are forever seen as riskier to their children.”