|
|
| Line 172: |
Line 172: |
| | </b><br>5 years ago, Joe Doliner and his co-founder Joey Zwicker decided to focus on the hard problems in data science, rather than building just another dashboard on top of the existing mess. It's been a long road, but it's really payed off. Last year, after an adventurous journey, they closed a $10m Series A led by Benchmark. In this episode, Erasmus Elsner is joined by Joe Doliner to explore what Pachyderm does and how it scaled from just an idea into a fast growing tech company. Listen to the podcast version | | </b><br>5 years ago, Joe Doliner and his co-founder Joey Zwicker decided to focus on the hard problems in data science, rather than building just another dashboard on top of the existing mess. It's been a long road, but it's really payed off. Last year, after an adventurous journey, they closed a $10m Series A led by Benchmark. In this episode, Erasmus Elsner is joined by Joe Doliner to explore what Pachyderm does and how it scaled from just an idea into a fast growing tech company. Listen to the podcast version |
| | http://apple.co/2W2g0nV | | http://apple.co/2W2g0nV |
| − | |}
| |
| − | |}<!-- B -->
| |
| − |
| |
| − | = <span id="Data Quality"></span>Data Quality =
| |
| − | * [http://greatexpectations.io/ Great Expectations] ...helps data teams eliminate pipeline debt, through data testing, documentation, and profiling.
| |
| − | {|<!-- T -->
| |
| − | | valign="top" |
| |
| − | {| class="wikitable" style="width: 550px;"
| |
| − | ||
| |
| − | <youtube>aUGCxTgvFf0</youtube>
| |
| − | <b>Testing and Documenting Your Data Doesn't Have to Suck | Superconductive
| |
| − | </b><br>Data teams everywhere struggle with pipeline debt: untested, undocumented assumptions that drain productivity, erode trust in data and kill team morale. Unfortunately, rolling your own data validation tooling usually takes weeks or months. In addition, most teams suffer from “documentation rot,” where data documentation is hard to maintain, and therefore chronically outdated, incomplete, and only semi-trusted. Great Expectations - http://bit.ly/2OtmY1W, the leading open source project for fighting pipeline debt, can solve these problems for you. We're excited to share new features and under-the-hood architecture with the data community. ABOUT THE SPEAKER
| |
| − | Abe Gong is a core contributor to the Great Expectations open source library, and CEO and Co-founder at Superconductive. Prior to Superconductive, Abe was Chief Data Officer at Aspire Health, the founding member of the Jawbone data science team, and lead data scientist at Massive Health. Abe has been leading teams using data and technology to solve problems in health care, consumer wellness, and public policy for over a decade. Abe earned his PhD at the University of Michigan in Public Policy, Political Science, and Complex Systems. He speaks and writes regularly on data, healthcare, and data ethics.
| |
| − | |}
| |
| − | |<!-- M -->
| |
| − | | valign="top" |
| |
| − | {| class="wikitable" style="width: 550px;"
| |
| − | ||
| |
| − | <youtube>DRGajth6OO4</youtube>
| |
| − | <b>"Data Quality Check In Machine Learning"
| |
| − | </b><br>The world of data quality check in Machine Learning is expanding at an unimaginable pace. Researchers estimate that by 2020, every human would create 1.7MB of information each second. The true power of data can be unlocked when it is refined and transformed into a high quality state where we can realize its true potential. Many businesses and researchers believe that data quality is one of the primary concerns for data-driven enterprises and associated processes considering the pace of data growth. Most of the operational processes and analytics rely on good quality data for being efficient and consistent in output.Data quality process has evolved in its capacity but the demand for pace and efficiency has been proliferating extensively. Data management experts believe that data quality remains a bottleneck that creeps repeatedly to bother the data management and business fraternity due to proliferating data volumes and the complexity involved to derive quality insights. Innovative technologies such as Big Data, AI, ML etc.ML algorithms can learn from human decision labels in the training datasets and replicate the scenarios in real-time. However, ML algorithms are also prone to biases that may reflect in these data sets and are learnt through fresh data sets. These biases could lead to erosion of data quality. External validity testing and audits on a regular basis will help in avoiding such situations.
| |
| − | |}
| |
| − | |}<!-- B -->
| |
| − | {|<!-- T -->
| |
| − | | valign="top" |
| |
| − | {| class="wikitable" style="width: 550px;"
| |
| − | ||
| |
| − | <youtube>t7vHpA39TXM</youtube>
| |
| − | <b>An Approach to Data Quality for Netflix Personalization Systems
| |
| − | </b><br>Personalization is one of the key pillars of Netflix as it enables each member to experience the vast collection of content tailored to their interests. Our personalization system is powered by several machine learning models. These models are only as good as the data that is fed to them. They are trained using hundreds of terabytes of data everyday, that make it a non-trivial challenge to track and maintain data quality. To ensure high data quality, we require three things: automated monitoring of data; visualization to observe changes in the metrics over time; and mechanisms to control data related regressions, wherein a data regression is defined as data loss or distributional shifts over a given period of time. In this talk, we will describe infrastructure and methods that we used to achieve the above: – ‘Swimlanes’ that help us define data boundaries for different environments that are used to develop, evaluate and deploy ML models, – Pipelines that aggregate data metrics from various sources within each swimlane – Time series and dashboard visualization tools across an atypically larger period of time – Automated audits that periodically monitor these metrics to detect data regressions. We will explain how we run aggregation jobs to optimize metric computations, SQL queries to quickly define/test individual metrics and other ETL jobs to power the visualization/audits tools using Spark.’ About: Databricks provides a unified data analytics platform, powered by Apache Spark™, that accelerates innovation by unifying data science, engineering and business. Connect with us: Website: http://databricks.com [[Facebook]]: http://www.facebook.com/databricksinc
| |
| − | |}
| |
| − | |<!-- M -->
| |
| − | | valign="top" |
| |
| − | {| class="wikitable" style="width: 550px;"
| |
| − | ||
| |
| − | <youtube>ID2</youtube>
| |
| − | <b>HH2
| |
| − | </b><br>BB2
| |
| | |} | | |} |
| | |}<!-- B --> | | |}<!-- B --> |