711 results for search: %E3%80%94%EC%A4%91%EB%85%84%ED%8F%B0%ED%8C%85%E3%80%95%20WWW%E0%BC%9DPAYO%E0%BC%9DPW%20%20%EB%B2%95%EC%A0%84%EB%A7%8C%EB%82%A8%ED%86%A1%20%EB%B2%95%EC%A0%84%EB%AA%A8%EC%9E%84%EC%96%B4%ED%94%8C%E2%88%83%EB%B2%95%EC%A0%84%EB%AF%B8%ED%8C%85%EC%96%B4%ED%94%8C%E2%97%86%EB%B2%95%EC%A0%84%EB%B2%88%EA%B0%9C%ED%8C%85%E2%92%AA%E3%81%B4%E9%B9%80lewdness/feed/content/colombia/privacy


Carnegie Mellon Partners With Human Rights Data Analysis Group To Improve Syrian Casualty Reporting


Former Leader of Guatemala Is Guilty of Genocide Against Mayan Group


The ghost in the machine

“Every kind of classification system – human or machine – has several kinds of errors it might make,” [Patrick Ball] says. “To frame that in a machine learning context, what kind of error do we want the machine to make?” HRDAG’s work on predictive policing shows that “predictive policing” finds patterns in police records, not patterns in occurrence of crime.


Counting the Dead in Syria


Guatemalan Ex-Cops Get 40 Years for Labor Leader’s Slaying


Chad: Habré Knew of Deaths in His Jails


Inside the Difficult, Dangerous Work of Tallying the ISIS Death Toll

HRDAG executive director Megan Price is interviewed by Mother Jones. An excerpt: “Violence can be hidden,” says Price. “ISIS has its own agenda. Sometimes that agenda is served by making public things they’ve done, and I have to assume, sometimes it’s served by hiding things they’ve done.”


Benetech Celebrates Milestone; Human Rights Data Analysis Group Transitioning into Independent Organization


60,000 Dead in Syria? Why the Death Toll is Likely Even Higher


5 Questions for Kristian Lum

Kristian Lum discusses the challenges of getting accurate data from conflict zones, as well as her concerns about predictive policing if law enforcement gets it wrong.


Syrian Death Toll Reaches 60,000, Says UN Rights Agency


Data Dive Reveals 15,000 New Victims of Syria War


Patrick Ball on the Perils of Misusing Human Rights Data


The Panic Button: High-Tech Protection for Human Rights Investigators


Palantir Has Secretly Been Using New Orleans to Test Its Predictive Policing Technology

One of the researchers, a Michigan State PhD candidate named William Isaac, had not previously heard of New Orleans’ partnership with Palantir, but he recognized the data-mapping model at the heart of the program. “I think the data they’re using, there are serious questions about its predictive power. We’ve seen very little about its ability to forecast violent crime,” Isaac said.


Celebrating Women in Statistics

kristian lum headshot 2018In her work on statistical issues in criminal justice, Lum has studied uses of predictive policing—machine learning models to predict who will commit future crime or where it will occur. In her work, she has demonstrated that if the training data encodes historical patterns of racially disparate enforcement, predictions from software trained with this data will reinforce and—in some cases—amplify this bias. She also currently works on statistical issues related to criminal “risk assessment” models used to inform judicial decision-making. As part of this thread, she has developed statistical methods for removing sensitive information from training data, guaranteeing “fair” predictions with respect to sensitive variables such as race and gender. Lum is active in the fairness, accountability, and transparency (FAT) community and serves on the steering committee of FAT, a conference that brings together researchers and practitioners interested in fairness, accountability, and transparency in socio-technical systems.


Data Analysis By Benetech Scientists Aid in Arrest of Former Guatemalan Police Chief


Benetech Scientists Publish Analysis of Indirect Sampling Methods in the Journal of the American Medical Association


The Body Counter


Trump’s “extreme-vetting” software will discriminate against immigrants “Under a veneer of objectivity,” say experts

Kristian Lum, lead statistician at the Human Rights Data Analysis Group (and letter signatory), fears that “in order to flag even a small proportion of future terrorists, this tool will likely flag a huge number of people who would never go on to be terrorists,” and that “these ‘false positives’ will be real people who would never have gone on to commit criminal acts but will suffer the consequences of being flagged just the same.”


Our work has been used by truth commissions, international criminal tribunals, and non-governmental human rights organizations. We have worked with partners on projects on five continents.

Donate