Difference between revisions of "Boolean Satisfiability (SAT) Problem/Satisfiability Modulo Theories (SMT) Solvers"

From
Jump to: navigation, search
Line 1: Line 1:
 +
{{#seo:
 +
|title=PRIMO.ai
 +
|titlemode=append
 +
|keywords=artificial, intelligence, machine, learning, models, algorithms, data, singularity, moonshot, Tensorflow, Google, Nvidia, Microsoft, Azure, Amazon, AWS
 +
|description=Helpful resources for your journey with artificial intelligence; videos, articles, techniques, courses, profiles, and tools
 +
}}
 
[http://www.youtube.com/results?search_query=~SAT+SMT+Satisfiability+Modulo+Theories+Z3+Reluplex+Deep+Learning+Artificial+Intelligence Youtube search...]
 
[http://www.youtube.com/results?search_query=~SAT+SMT+Satisfiability+Modulo+Theories+Z3+Reluplex+Deep+Learning+Artificial+Intelligence Youtube search...]
 +
[http://www.google.com/search?q=SAT+SMT+Satisfiability+Modulo+Theories+Z3+Reluplex+deep+machine+learning+ML+artificial+intelligence ...Google search]
  
 
* [[Offense - Adversarial Threats/Attacks]]
 
* [[Offense - Adversarial Threats/Attacks]]

Revision as of 18:19, 2 February 2019

Youtube search... ...Google search


In what seems to be an endless back-and-forth between new adversarial attacks and new defenses against those attacks, we would like a means of formally verifying the robustness of machine learning algorithms to adversarial attacks. In the privacy domain, there is the idea of a differential privacy budget, which quantifies privacy over all possible attacks. In the following three papers, we see attempts at deriving an equivalent benchmark for security, one that will allow the evaluation of defenses against all possible attacks instead of just a specific one. Class 6: Measuring Robustness of ML Models