Law Enforcement

Jump to: navigation, search

Youtube search... ...Google search

Most Advanced Police And Security Robots Patrolling Everywhere
Watch these most advanced police and security robots in action. They are already in action to maintain peace, safety and security for all people. But they are also assigned with task to capture criminals and law-breakers. SUBSCRIBE To Richard Aguilar Channel: RECOMMENDED BOOKS ABOUT AI & ROBOTICS: Introduction to Artificial Intelligence: Artificial Intelligence: A Modern Approach: Hands-On Machine Learning: Building Automation Integration: Quick Video Summary: 1. Crime-Fighting Robots 2. Silicon Valley’s Police Robot 3. Huntington Park 400-Pound Robot Cop 4. New Robots Security Guards In The US 5. Northeast Ohio Police Robots 6. Robot Security Guards Now A Reality 7. Robots Add Another Layer Of Security 8. Security Robot Patrols At Gas Station 9. Twin Arrows Casino's First Security Robot

International Policing, Ethics, & the Use of AI in Law Enforcement, with Interpol's Jürgen Stock
In this episode of the Artificial Intelligence & Equality podcast, Senior Fellow Anja Kaspersen speaks with Dr. Jürgen Stock, secretary general of the International Criminal Police Organization (Interpol). In an engaging conversation, they discuss his professional journey towards leading the world police body, what keeps him up at night, and the critical role of global police work in keeping societies safe, especially as those seeking to evade justice increasingly hide behind screens, and operate via bits and bytes, as well as on the dark net.

Using Artificial Intelligence to Fight Human Trafficking | Emily Kennedy | TEDxPittsburgh
The horrors of human trafficking are widespread across the world. Big problems require top technology and ingenuity, that's where the idea of creating Artificial Intelligence powered software for law enforcement comes in. Technologist and founder Emily Kennedy's idea for fighting the spread of human trafficking centers on the heart of the mission and the technology created to find and rescue victims while bringing criminals to justice. Emily Kennedy is a startup founder, human trafficking subject matter expert, Forbes 30 Under 30, a Mother of Invention, keynote speaker, and activist. She has been developing technology solutions to human trafficking since 2011 at the Carnegie Mellon University Robotics Institute. Her company Marinus Analytics uses the latest advancements in AI to turn big data into actionable intelligence for sex trafficking investigations. As President and Co-Founder of Marinus Analytics, she leads development and deployment of these tools to law enforcement across the globe for use on criminal cases, with an emphasis on sex trafficking investigations. She routinely works alongside, advises, and teaches stakeholders—such as attorneys general, prosecutors, law enforcement agents, and non-profit victim services organizations—on micro and macro approaches to combating and measuring human trafficking in the United States and abroad. This talk was given at a TEDx event using the TED conference format but independently organized by a local community. Learn more at

Can Google Predict Who Will Commit Crimes? | Seth Stephens-Davidowitz
Are you a future criminal? You might not think so, says data scientist Seth Stephens-Davidowitz, but what do you look like on paper? Have you ever searched something suspicious online? Ever been curious about a dark topic? Just like the film Minority Report, where "future murderers" are arrested before they commit their crimes, we have a similar predictive tool ready-made: Google's search data. People really do search for things like 'how to kill your girlfriend' or 'how to dispose of a body', but as Stephens-Davidowitz points out, it’s not supposed to be illegal to have bad thoughts. Beyond privacy and ethics, data science also backs the idea that you can't predict with any accuracy who will commit a crime, as he says: "a lot of people have horrific thoughts or make horrific searches without ever going through with a horrific action." Data also provides intriguing correlations about who or won't will pay their loans based on a single word used in their loan application, and reveals the questions people in the Bible Belt are too afraid to ask aloud. This kind of data in the wrong hands can leave people vulnerable to discrimination or worse, if society lets its ethics slide. Stephens-Davidowitz is the author of Everybody Lies: Big Data, New Data, and What the Internet Can Tell Us About Who We Really Are. Seth Stephens-Davidowitz has used data from the internet, particularly Google searches, to get new insights into the human psyche. A book summarizing his research, Everybody Lies, was published in May 2017 by HarperCollins. Seth has used Google searches to measure racism, self-induced abortion, depression, child abuse, hateful mobs, the science of humor, sexual preference, anxiety, son preference, and sexual insecurity, among many other topics. He worked for one-and-a-half years as a data scientist at Google and is currently a contributing op-ed writer for the New York Times. He is designing and teaching a course about his research at The Wharton School at the University of Pennsylvania, where he will be a visiting lecturer.

How Cops Are Using Algorithms to Predict Crimes | WIRED
The LAPD is one of a growing number of police departments using algorithms to try to predict crimes before they happen. Proponents of these tools say they provide cops with added tools to keep their cities safe -- but critics argue it's just another form of profiling.

Police Unlock AI's Potential to Monitor, Surveil and Solve Crimes | WSJ
Law enforcement agencies like the New Orleans Police Department are adopting artificial-intelligence based systems to analyze surveillance footage. WSJ's Jason Bellini gets a demonstration of the tracking technology and hears why some think it’s a game changer, while for others it’s raising concerns around privacy and potential bias. Photo: Drew Evans/The Wall Street Journal

The danger of predictive algorithms in criminal justice | Hany Farid | TEDxAmoskeagMillyard
Predictive algorithms may help us shop, discover new music or literature, but do they belong in the courthouse? Dartmouth professor Dr. Hany Farid reverse engineers the inherent dangers and potential biases of recommendations engines built to mete out justice in today's criminal justice system. The co-founder and CTO of Fourandsix Technologies, an image authentication and forensics company, Hany Farid works to advance the field of digital forensics. Hany said, “For the past decade I have been working on technology and policy that will find a balance between an open and free Internet while reining in online abuses. With approximately a billion Facebook uploads per day and 400 hours of video uploaded to YouTube every minute, this task is technically and logistically complicated but also, I believe, critical to the long-term health of our online communities.” Hany is the Albert Bradley 1915 Third Century Professor and Chair of Computer Science at Dartmouth. He is also a Senior Adviser to the Counter Extremism Project. This talk was given at a TEDx event using the TED conference format but independently organized by a local community. Learn more at

The Future of Policing – Challenge and Opportunity | Simon O'Rourke | TEDxFulbrightPerth
The future of policing presents both challenge and opportunity as advances in technology continue to influence how we interact as a society. We are becoming dual citizens of the physical and digital worlds with accompanying expectations of police to provide similar services in both domains. Keeping pace with the expectations of an increasingly networked and digitised society will require police to become highly innovative. Simon O’Rourke is a career police officer with over 20 years operational experience and a keen interest in technology. He is the recipient of the 2017 Fulbright Western Australia Postdoctoral Scholarship, which saw him appointed as a Fellow at the Program on Crisis Leadership at the Harvard Kennedy School for the 2017-18 Academic Year. His research focused on Police Command at Critical Incidents, including Terrorism. His current role is to develop and prepare Police Commanders for the challenges they will face during a major incident, where they will be required to make critical decisions in a highly complex environment. This talk was given at a TEDx event using the TED conference format but independently organized by a local community. Learn more at

Reshaping the Future of Crime, Terrorism and Security - Artificial Intelligence and Robotics
Recent technological advancements in artificial intelligence (AI) and robotics have moved these technologies away from the realm of science fiction and into our daily lives. The massive growth in computational power and increasing abundance of data have vastly improved the capabilities of AI and robotics, giving them more real-world applications. In light of this, stakeholders in both the public and private sector have begun to pursue these technologies with a view to revolutionizing fields such as healthcare, transportation, agriculture and the financial and legal systems, by enhancing efficiency, optimizing resource allocation, reducing costs and creating new revenue opportunities. The technological advances taking place in the fields of AI and robotics can also have many positive effects for law enforcement and security agencies, for instance in terms of identifying persons of interest, stolen vehicles or suspicious sounds and behavior; predicting trends in criminality or terrorist action; tracking illicit money flows; flagging and responding to terrorist use of the internet, and even contributing to international cooperation by supporting the research, analysis and response to international mutual assistance requests from the International Criminal Police Organization (INTERPOL). At the same time however, these technologies are only as good as the user that employs them. In the hands of criminals or terrorist organizations such dual-use technologies could equally enable new digital, physical or even political threats. The event will seek to build upon the success of the UNICRI-INTERPOL meeting in Singapore by further raising awareness of the risks and benefits of AI and robotics for a crime, terrorism and security perspective and contributing to fostering a coordinated international movement on the issue. Key challenges, findings and recommendations identified during the UNICRI-INTERPOL meeting will also be spotlight and copies of the forthcoming meeting report will be distributed. The event organized by UNICRI and INTERPOL, with the support of the Permanent Missions of Georgia, the Kingdom of the Netherlands and the United Arab Emirates will have two substantive panels: Panel I – “The Future, Today” Panel II – “Facing the Challenges Together”

How AI Could Reinforce Biases In The Criminal Justice System
Increasingly, algorithms and machine learning are being implemented at various touch points throughout the criminal justice system, from deciding where to deploy police officers to aiding in bail and sentencing decisions. The question is, will this tech make the system more fair for minorities and low-income residents, or will it simply amplify our human biases? We all know humans are imperfect. We're subject to biases and stereotypes, and when these come into play in the criminal justice system, the most disadvantaged communities end up suffering. It's easy to imagine that there's a better way, that one day we'll find a tool that can make neutral, dispassionate decisions about policing and punishment. Some think that day has already arrived. Around the country, police departments and courtrooms are turning to artificial intelligence algorithms to help them decide everything from where to deploy police officers to whether to release defendants on bail. Supporters believe that the technology will lead to increased objectivity, ultimately creating safer communities. Others however, say that the data fed into these algorithms is encoded with human bias, meaning the tech will simply reinforce historical disparities. Learn more about the ways in which communities, policemen and judges across the U.S. are using these algorithms to make decisions about public safety and people's lives.

CPDP 2020: Regulating Artificial Intelligence in Criminal Justice?
MODERATOR: Juraj Sajfert SPEAKERS: Katalin Ligeti, University of Luxembourg (LU); Anna Moscibroda, DG JUST (EU); Lani Cossette, Microsoft (BE); Frank Schuermans, Supervisory Body for Police Information (BE) Panel Description AI can make predictions about where, when, and by whom crimes are likely to be committed. AI can also estimate how likely it is that a suspect, defendant or convict flees or commits further crimes. Against the backdrop that AI helps predictive policing and predictive justice, what should the EU’s legal and policy responses be, in particular after the adoption of the Artificial Intelligence Ethics Guidelines? One approach is to count on the vitality of recently adopted data protection laws -in particular, Law Enforcement Directive (EU) 2016/680. Another approach would be to launch a regulatory reform process, either in or out of the classical data protection realm. This panel will look at the usefulness and reliability of AI for criminal justice and will critically asses the different regulatory avenues the new European Commission might consider. - How does the idea of “trustworthy AI” translate into the area of criminal law? - Should we not ban the use of predictive policing systems or the use of AI in criminal law cases, on the basis of ethics? - Does the new European Commission plan to propose legislation in this area? If yes, what would be the objectives of such new laws? Should the actors leading such a reform be different from the ones that were leading the EU data protection reform? - Is it possible to develop predictive justice and predictive policing, and still respect the requirements of the GDPR and Directive (EU) 2016/680?

Artificial Intelligence: The World According to AI |Targeted by Algorithm (Ep1)| The Big Picture
Artificial intelligence is already here. There's a lot of debate and hype about AI, and it's tended to focus on the extreme possibilities of a technology still in its infancy. From self-aware computers and killer robots taking over the world, to a fully-automated world where humans are made redundant by machines, the brave new world of Artificial Intelligence is prophesied by some to be a doomed, scary place, no place for people. For others, AI is ushering in great technological advances for humanity, helping the world communicate, manufacture, trade and innovate faster, longer, better. But in between these competing utopian and dystopian visions, AI is allowing new ways of maintaining an old order. It is being used across public and private spheres to make decisions about the lives of millions of people around the world - and sometimes those decisions can mean life or death. "Communities, particularly vulnerable communities, children, people of colour, women are often characterised by these systems, in quite misrepresentative ways," says Safiya Umoja Noble, author of the book, Algorithms of Oppression. In episode one of The Big Picture: The World According to AI, we chart the evolution of artificial intelligence from its post-World War II origins and, dissect the mechanisms by which existing prejudices are built into the very systems that are supposed to be free of human bias. We shed a harsh light on computerised targeting everywhere from foreign drone warfare to civilian policing. In the UK, we witness the trialling of revolutionary new facial recognition technology by the London Metropolitan Police Service. We examine how these technologies, that are far from proven, are being sold as new policing solutions to maintain in some of the world's biggest cities. The Big Picture: The World According to AI explores how artificial intelligence is being used today, and what it means to those on its receiving end. Watch Episode 2 here:

The Future of Crime Detection and Prevention
Could an artificial intelligence predict a crime before it happens? Will we ever truly trust a machine? What new technology might be used against us in the future? Subscribe for regular science videos: Our expert panel will open our eyes and try to allay our fears regarding the future of crime. Gloria Laycock is a Professor of Crime Science in the Engineering Sciences Faculty at UCL. She was a researcher in the Home Office for many years, leaving as Head of the Home Office Police Research Group for a fellowship in the USA, before coming to UCL as Director of the Jill Dando Institute of Security and Crime Science. Mark Girolami holds the Chair of Statistics within the Department of Mathematics at Imperial College London where he is also Professor of Computing Science in the Department of Computing. He is an adjunct Professor of Statistics at the University of Warwick and is Director of the Lloyd’s Register Foundation Programme on Data Centric Engineering at the Alan Turing Institute where he served as one of the original founding Executive Directors. He is an elected member of the Royal Society of Edinburgh and previously was awarded a Royal Society - Wolfson Research Merit Award. Professor Girolami has been an EPSRC Research Fellow continuously since 2007 and in 2018 he was awarded the Royal Academy of Engineering Research Chair in Data Centric Engineering. His research focuses on applications of mathematical and computational statistics such as Machine Learning. Adrian Weller is a Senior Research Fellow in Machine Learning at the University of Cambridge, and Programme Director for Artificial Intelligence at The Alan Turing Institute. He has broad interests across machine learning and artificial intelligence, their applications, and their implications for society. This talk and Q&A was filmed in the Ri on 8 May 2018.

The Use of Artificial Intelligence in the Administration of Justice
UNODC - United Nations Office on Drugs and Crime Global Judicial Integrity Network Webinar Series The use of artificial intelligence in the administration of justice / Artificial Intelligence (AI) can improve the efficiency of judiciaries, especially in regards to court administration tasks (such as automatically assigning cases to judges, or marking cases as “urgent”). However, these benefits come with risks — judiciaries need to ensure that any court and case management software using AI is operating without error, and also without bias. Guest speaker: Professor Karen Yeung Interdisciplinary Professorial Fellow in Law, Ethics and Informatics at the University of Birmingham in the School of Law and the School of Computer Science.