609 results for search: %ED%99%8D%EB%B3%B4%ED%8C%80%E2%98%8E%EC%B9%B4%ED%86%A1adgogo%E2%98%8E%EC%99%95%EC%8B%AD%EB%A6%AC%EC%97%AD%EC%A3%BC%EC%A0%90%E3%83%A5%ED%99%8D%EB%B3%B4%E2%94%82%ED%8C%80%E2%97%95%EC%99%95%EC%8B%AD%EB%A6%AC%EC%97%AD%E8%B1%B7%EC%A3%BC%EC%A0%90%E8%B1%88endocardium


Mining data on mutilations, beatings, murders


A Human Rights Statistician Finds Truth In Numbers

The tension started in the witness room. “You could feel the stress rolling off the walls in there,” Patrick Ball remembers. “I can remember realizing that this is why lawyers wear sport coats – you can’t see all the sweat on their arms and back.” He was, you could say, a little nervous to be cross-examined by Slobodan Milosevic.


Guatemala Struggles to Find War Crimes Justice


Doing Well By Doing Good


Guatemala Police Archive Yields Clues to ‘Dirty War’


The Quiet Revolution


How statistics caught Indonesia’s war-criminals


Coders Bare Invasion Death Count


Benetech: Using technology to improve human rights


El científico que usa estadísticas para encontrar desaparecidos en El Salvador, Guatemala y México

Patrick Ball es un sabueso de la verdad. Ese deseo de descubrir lo que otros quieren ocultar lo ha llevado a desarrollar fórmulas matemáticas para detectar desaparecidos.

Su trabajo consiste en aplicar métodos de medición científica para comprobar violaciones masivas de derechos humanos.


Data Analysis By Benetech Scientists Aid in Arrest of Former Guatemalan Police Chief


Patrick Ball on the Perils of Misusing Human Rights Data


One Better

The University of Michigan College of Literature, Science and the Arts profiled Patrick Ball in its fall 2016 issue of the alumni magazine. Here’s an excerpt:

Ball believes doing this laborious, difficult work makes the world a more just place because it leads to accountability.

“My part is a specific, narrow piece, which just happens to fit with the skills I have,” he says. “I don’t think that what we do is in any way the best or most important part of human rights activism. Sometimes, we are just a footnote—but we are a really good footnote.”


Rise of the racist robots – how AI is learning all our worst impulses

“If you’re not careful, you risk automating the exact same biases these programs are supposed to eliminate,” says Kristian Lum, the lead statistician at the San Francisco-based, non-profit Human Rights Data Analysis Group (HRDAG). Last year, Lum and a co-author showed that PredPol, a program for police departments that predicts hotspots where future crime might occur, could potentially get stuck in a feedback loop of over-policing majority black and brown neighbourhoods. The program was “learning” from previous crime reports. For Samuel Sinyangwe, a justice activist and policy researcher, this kind of approach is “especially nefarious” because police can say: “We’re not being biased, we’re just doing what the math tells us.” And the public perception might be that the algorithms are impartial.


Crean sistema para predecir fosas clandestinas en México

Por ello, Human Rights Data Analysis Group (HRDAG), el Programa de Derechos Humanos de la Universidad Iberoamericana (UIA) y Data Cívica, realizan un análisis estadístico construido a partir de una variable en la que se identifican fosas clandestinas a partir de búsquedas automatizadas en medios locales y nacionales, y usando datos geográficos y sociodemográficos.


Mapping Mexico’s hidden graves

When Patrick Ball was introduced to Ibero’s database, the director of research at the Human Rights Data Analysis Group in San Francisco, California, saw an opportunity to turn the data into a predictive model. Ball, who has used similar models to document human rights violations from Syria to Guatemala, soon invited Data Cívica, a Mexico City–based nonprofit that creates tools for analyzing data, to join the project.


Sobre fosas clandestinas, tenemos más información que el gobierno: Ibero

El modelo “puede distinguir entre los municipios en que vamos a encontrar fosas clandestinas, y en los que es improbable que vayamos a encontrar estas fosas”, explicó Patrick Ball, estadístico estadounidense que colabora con el Programa de Derechos Humanos de la Universidad Iberoamericana de la Ciudad de México.


The ghost in the machine

“Every kind of classification system – human or machine – has several kinds of errors it might make,” [Patrick Ball] says. “To frame that in a machine learning context, what kind of error do we want the machine to make?” HRDAG’s work on predictive policing shows that “predictive policing” finds patterns in police records, not patterns in occurrence of crime.


5 Questions for Kristian Lum

Kristian Lum discusses the challenges of getting accurate data from conflict zones, as well as her concerns about predictive policing if law enforcement gets it wrong.


What happens when you look at crime by the numbers

Kristian Lum’s work on the HRDAG Policing Project is referred to here: “In fact, Lum argues, it’s not clear how well this model worked at depicting the situation in Oakland. Those data on drug crimes were biased, she now reports. The problem was not deliberate, she says. Rather, data collectors just missed some criminals and crime sites. So data on them never made it into her model.”


Our work has been used by truth commissions, international criminal tribunals, and non-governmental human rights organizations. We have worked with partners on projects on five continents.

Donate