Difference between revisions of "Algorithm Administration"
m (BPeat moved page Data Lineage / Data Catalog to Feature Store / Data Lineage / Data Catalog without leaving a redirect) |
m |
||
(219 intermediate revisions by the same user not shown) | |||
Line 1: | Line 1: | ||
− | + | {{#seo: | |
+ | |title=PRIMO.ai | ||
+ | |titlemode=append | ||
+ | |keywords=ChatGPT, artificial, intelligence, machine, learning, GPT-4, GPT-5, NLP, NLG, NLC, NLU, models, data, singularity, moonshot, Sentience, AGI, Emergence, Moonshot, Explainable, TensorFlow, Google, Nvidia, Microsoft, Azure, Amazon, AWS, Hugging Face, OpenAI, Tensorflow, OpenAI, Google, Nvidia, Microsoft, Azure, Amazon, AWS, Meta, LLM, metaverse, assistants, agents, digital twin, IoT, Transhumanism, Immersive Reality, Generative AI, Conversational AI, Perplexity, Bing, You, Bard, Ernie, prompt Engineering LangChain, Video/Image, Vision, End-to-End Speech, Synthesize Speech, Speech Recognition, Stanford, MIT |description=Helpful resources for your journey with artificial intelligence; videos, articles, techniques, courses, profiles, and tools | ||
− | + | <!-- Google tag (gtag.js) --> | |
− | + | <script async src="https://www.googletagmanager.com/gtag/js?id=G-4GCWLBVJ7T"></script> | |
+ | <script> | ||
+ | window.dataLayer = window.dataLayer || []; | ||
+ | function gtag(){dataLayer.push(arguments);} | ||
+ | gtag('js', new Date()); | ||
− | <youtube> | + | gtag('config', 'G-4GCWLBVJ7T'); |
+ | </script> | ||
+ | }} | ||
+ | [https://www.youtube.com/results?search_query=ai+Data+Algorithm+Administration YouTube] | ||
+ | [https://www.quora.com/search?q=ai%20Data%20Algorithm%20Administration ... Quora] | ||
+ | [https://www.google.com/search?q=ai+Data+Algorithm+Administration ...Google search] | ||
+ | [https://news.google.com/search?q=ai+Data+Algorithm+Administration ...Google News] | ||
+ | [https://www.bing.com/news/search?q=ai+Data+Algorithm+Administration&qft=interval%3d%228%22 ...Bing News] | ||
+ | |||
+ | * [[AI Solver]] ... [[Algorithms]] ... [[Algorithm Administration|Administration]] ... [[Model Search]] ... [[Discriminative vs. Generative]] ... [[Train, Validate, and Test]] | ||
+ | * [[Development]] ... [[Notebooks]] ... [[Development#AI Pair Programming Tools|AI Pair Programming]] ... [[Codeless Options, Code Generators, Drag n' Drop|Codeless]] ... [[Hugging Face]] ... [[Algorithm Administration#AIOps/MLOps|AIOps/MLOps]] ... [[Platforms: AI/Machine Learning as a Service (AIaaS/MLaaS)|AIaaS/MLaaS]] | ||
+ | * [[Natural Language Processing (NLP)#Managed Vocabularies |Managed Vocabularies]] | ||
+ | * [[Analytics]] ... [[Visualization]] ... [[Graphical Tools for Modeling AI Components|Graphical Tools]] ... [[Diagrams for Business Analysis|Diagrams]] & [[Generative AI for Business Analysis|Business Analysis]] ... [[Requirements Management|Requirements]] ... [[Loop]] ... [[Bayes]] ... [[Network Pattern]] | ||
+ | * [[Backpropagation]] ... [[Feed Forward Neural Network (FF or FFNN)|FFNN]] ... [[Forward-Forward]] ... [[Activation Functions]] ...[[Softmax]] ... [[Loss]] ... [[Boosting]] ... [[Gradient Descent Optimization & Challenges|Gradient Descent]] ... [[Algorithm Administration#Hyperparameter|Hyperparameter]] ... [[Manifold Hypothesis]] ... [[Principal Component Analysis (PCA)|PCA]] | ||
+ | * [[Strategy & Tactics]] ... [[Project Management]] ... [[Best Practices]] ... [[Checklists]] ... [[Project Check-in]] ... [[Evaluation]] ... [[Evaluation - Measures|Measures]] | ||
+ | ** [[Evaluation - Measures#Accuracy|Accuracy]] | ||
+ | ** [[Evaluation - Measures#Precision & Recall (Sensitivity)|Precision & Recall (Sensitivity)]] | ||
+ | ** [[Evaluation - Measures#Specificity|Specificity]] | ||
+ | * [[Artificial General Intelligence (AGI) to Singularity]] ... [[Inside Out - Curious Optimistic Reasoning| Curious Reasoning]] ... [[Emergence]] ... [[Moonshots]] ... [[Explainable / Interpretable AI|Explainable AI]] ... [[Algorithm Administration#Automated Learning|Automated Learning]] | ||
+ | * NLP [[Natural Language Processing (NLP)#Workbench / Pipeline | Workbench / Pipeline]] | ||
+ | * [[Data Science]] ... [[Data Governance|Governance]] ... [[Data Preprocessing|Preprocessing]] ... [[Feature Exploration/Learning|Exploration]] ... [[Data Interoperability|Interoperability]] ... [[Algorithm Administration#Master Data Management (MDM)|Master Data Management (MDM)]] ... [[Bias and Variances]] ... [[Benchmarks]] ... [[Datasets]] | ||
+ | * [[Data Quality]] ...[[AI Verification and Validation|validity]], [[Evaluation - Measures#Accuracy|accuracy]], [[Data Quality#Data Cleaning|cleaning]], [[Data Quality#Data Completeness|completeness]], [[Data Quality#Data Consistency|consistency]], [[Data Quality#Data Encoding|encoding]], [[Data Quality#Zero Padding|padding]], [[Data Quality#Data Augmentation, Data Labeling, and Auto-Tagging|augmentation, labeling, auto-tagging]], [[Data Quality#Batch Norm(alization) & Standardization| normalization, standardization]], and [[Data Quality#Imbalanced Data|imbalanced data]] | ||
+ | * [[Building Your Environment]] | ||
+ | * [[Service Capabilities]] | ||
+ | * [[AI Marketplace & Toolkit/Model Interoperability]] | ||
+ | * [[Directed Acyclic Graph (DAG)]] - programming pipelines | ||
+ | * [[Containers; Docker, Kubernetes & Microservices]] | ||
+ | * [https://getmanta.com/?gclid=CjwKCAjwsfreBRB9EiwAikSUHSSOxld0nZNyLNXmiPM43x7jEAgeTxkXRH_s5XPJlfTekPdO8N1Y1xoCKwwQAvD_BwE Automate your data lineage] | ||
+ | * [https://www.information-age.com/benefiting-ai-data-management-123471564/ Benefiting from AI: A different approach to data management is needed] | ||
+ | * [[Libraries & Frameworks Overview]] ... [[Libraries & Frameworks]] ... [[Git - GitHub and GitLab]] ... [[Other Coding options]] | ||
+ | ** [[Writing/Publishing#Model Publishing|publishing your model]] | ||
+ | * [https://github.com/JonTupitza/Data-Science-Process/blob/master/10-Modeling-Pipeline.ipynb Use a Pipeline to Chain PCA with a RandomForest Classifier Jupyter Notebook |] [https://github.com/jontupitza Jon Tupitza] | ||
+ | * [https://devblogs.microsoft.com/cesardelatorre/ml-net-model-lifecycle-with-azure-devops-ci-cd-pipelines/ ML.NET Model Lifecycle with Azure DevOps CI/CD pipelines | Cesar de la Torre - Microsoft] | ||
+ | * [https://medium.com/data-ops/a-great-model-is-not-enough-deploying-ai-without-technical-debt-70e3d5fecfd3 A Great Model is Not Enough: Deploying AI Without Technical Debt | DataKitchen - Medium] | ||
+ | * [https://towardsdatascience.com/ml-infrastructure-tools-for-production-part-2-model-deployment-and-serving-fcfc75c4a362ML Infrastructure Tools for Production | Aparna Dhinakaran - Towards Data Science] ...Model Deployment and Serving | ||
+ | * [https://www.camelot-mc.com/en/client-services/information-data-management/global-community-for-artificial-intelligence-in-mdm/ Global Community for Artificial Intelligence (AI) in Master Data Management (MDM) | Camelot Management Consultants] | ||
+ | * [https://ce.aut.ac.ir/~meybodi/paper/Dynamic%20environment/PSO%20in%20dynamic%20environment==/Particle%20Swarms%20for%20Dynamic%20Optimization%20Problems.2008.pdf Particle Swarms for Dynamic Optimization Problems | T. Blackwell, J. Branke, and X. Li] | ||
+ | * [[Telecommunications#5G_Security|5G_Security]] | ||
+ | |||
+ | = Tools = | ||
+ | * [[Google AutoML]] automatically build and deploy state-of-the-art machine learning models | ||
+ | ** [[TensorBoard]] | [[Google ]] | ||
+ | ** [[Kubeflow Pipelines]] - a platform for building and deploying portable, scalable machine learning (ML) workflows based on Docker containers. [https://cloud.google.com/blog/products/ai-machine-learning/introducing-ai-hub-and-kubeflow-pipelines-making-ai-simpler-faster-and-more-useful-for-businesses Introducing AI Hub and Kubeflow Pipelines: Making AI simpler, faster, and more useful for businesses] | [[Google ]] | ||
+ | * [[SageMaker]] | [[Amazon]] | ||
+ | * [https://docs.microsoft.com/en-us/azure/machine-learning/concept-model-management-and-deployment MLOps] | [[Microsoft]] ...model management, deployment, and monitoring with Azure | ||
+ | ** [https://feedback.azure.com/forums/906052-data-catalog How can we improve Azure Data Catalog?] | ||
+ | ** [https://docs.microsoft.com/en-us/azure/machine-learning/concept-automated-ml AutoML] | ||
+ | * [[Ludwig]] - a [[Python]] toolbox from Uber that allows to train and test deep learning models | ||
+ | * TPOT a [[Python]] library that automatically creates and optimizes full machine learning pipelines using genetic programming. Not for NLP, strings need to be coded to numerics. | ||
+ | * [[H2O]] [https://www.h2o.ai/products/h2o-driverless-ai/ Driverless AI] for automated [[Visualization]], feature engineering, model training, [[Algorithm Administration#Hyperparameter|hyperparameter]] optimization, and explainability. | ||
+ | * [https://www.alteryx.com/ alteryx:] [https://www.featurelabs.com/ Feature Labs], [https://www.alteryx.com/innovation-labs Featuretools] | ||
+ | * [https://mlbox.readthedocs.io/en/latest/ MLBox] Fast reading and distributed data preprocessing/cleaning/formatting. Highly robust feature selection and leak detection. Accurate hyper-parameter optimization in high-dimensional space. State-of-the art predictive models for classification and regression (Deep Learning, Stacking, [[LightGBM]],…). Prediction with models interpretation. Primarily Linux. | ||
+ | * [https://automl.github.io/auto-sklearn/master/ auto-sklearn] algorithm selection and [[Algorithm Administration#Hyperparameter|hyperparameter]] tuning. It leverages recent advantages in Bayesian optimization, meta-learning and ensemble construction.is a Bayesian [[Algorithm Administration#Hyperparameter|hyperparameter]] optimization layer on top of [[Python#scikit-learn|scikit-learn]]. Not for large datasets. | ||
+ | * [[Auto Keras]] is an open-source [[Python]] package for neural architecture search. | ||
+ | * [https://github.com/HDI-Project/ATM ATM] -auto tune models - a multi-tenant, multi-data system for automated machine learning (model selection and tuning). ATM is an open source software library under the [https://github.com/HDI-Project Human Data Interaction project (HDI)] at MIT. | ||
+ | * [https://www.cs.ubc.ca/labs/beta/Projects/autoweka/ Auto-WEKA] is a Bayesian [[Algorithm Administration#Hyperparameter|hyperparameter]] optimization layer on top of [https://www.cs.waikato.ac.nz/ml/weka/ Weka]. [https://www.cs.waikato.ac.nz/ml/weka/ Weka] is a collection of machine learning algorithms for data mining tasks. The algorithms can either be applied directly to a dataset or called from your own Java code. [https://www.cs.waikato.ac.nz/ml/weka/ Weka] contains tools for data pre-processing, classification, regression, clustering, association rules, and [[visualization]]. | ||
+ | * [https://github.com/salesforce/TransmogrifAI TransmogrifAI] - an AutoML library for building modular, reusable, strongly typed machine learning workflows. A Scala/SparkML library created by Salesforce for automated data cleansing, feature engineering, model selection, and [[Algorithm Administration#Hyperparameter|hyperparameter]] optimization | ||
+ | * [https://github.com/laic-ufmg/Recipe RECIPE] - a framework based on grammar-based genetic programming that builds customized [[Python#scikit-learn|scikit-learn]] classification pipelines. | ||
+ | * [https://github.com/laic-ufmg/automlc AutoMLC] Automated Multi-Label Classification. GA-Auto-MLC and Auto-MEKAGGP are freely-available methods that perform automated multi-label classification on the MEKA software. | ||
+ | * [https://databricks.com/mlflow Databricks MLflow] an open source framework to manage the complete Machine Learning lifecycle using Managed MLflow as an integrated service with the [[Databricks]] Unified Analytics Platform... ...manage the ML lifecycle, including experimentation, reproducibility and deployment | ||
+ | * [https://www.sas.com/en_us/software/viya/new-features.html SAS Viya] automates the process of data cleansing, data transformations, feature engineering, algorithm matching, model training and ongoing governance. | ||
+ | * [https://www.comet.ml/site/ Comet ML] ...self-hosted and cloud-based meta machine learning platform allowing data scientists and teams to track, compare, explain and optimize experiments and models | ||
+ | * [https://www.dominodatalab.com/product/domino-model-monitor/ Domino Model Monitor (DMM) | Domino] ...monitor the performance of all models across your entire organization | ||
+ | * [https://www.wandb.com/ Weights and Biases] ...experiment tracking, model optimization, and dataset versioning | ||
+ | * [https://sigopt.com/ SigOpt] ...optimization platform and API designed to unlock the potential of modeling pipelines. This fully agnostic software solution accelerates, amplifies, and scales the model development process | ||
+ | * [https://dvc.org/ DVC] ...Open-source Version Control System for Machine Learning Projects | ||
+ | * [https://www.modelop.com/modelops-and-mlops/ ModelOp Center | ModelOp] | ||
+ | * [https://www.moogsoft.com/aiops-platform/ Moogsoft] and [https://www.ansible.com/ Red Hat Ansible] Tower | ||
+ | * [https://www.dataiku.com/product/ DSS | Dataiku] | ||
+ | * [https://www.sas.com/en_us/software/model-manager.html Model Manager | SAS] | ||
+ | * [https://www.datarobot.com/platform/mlops/ Machine Learning Operations (MLOps) | DataRobot] ...build highly accurate predictive models with full transparency | ||
+ | * [https://metaflow.org/ Metaflow], Netflix and AWS open source [[Python]] library | ||
+ | |||
+ | = Master Data Management (MDM) = | ||
+ | [https://www.youtube.com/results?search_query=Master+Data+Management+MDM+data+lineage+catalog+management+deep+machine+learning+ai YouTube search...] | ||
+ | [https://www.google.com/search?q=Master+Data+Management+MDM+data+lineage+catalog+management+deep+machine+learning+ai ...Google search] | ||
+ | Feature Store / Data Lineage / Data Catalog | ||
+ | |||
+ | * [[Data Science]] ... [[Data Governance|Governance]] ... [[Data Preprocessing|Preprocessing]] ... [[Feature Exploration/Learning|Exploration]] ... [[Data Interoperability|Interoperability]] ... [[Algorithm Administration#Master Data Management (MDM)|Master Data Management (MDM)]] ... [[Bias and Variances]] ... [[Benchmarks]] ... [[Datasets]] | ||
+ | |||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>x_ZlsC4h_4A</youtube> | ||
+ | <b>How is AI changing the game for Master Data Management? | ||
+ | </b><br>Tony Brownlee talks about the ability to inspect and find [[Data Quality|data quality]] issues as one of several ways cognitive computing technology is influencing master data management. | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>8OcGcScgzOo</youtube> | ||
+ | <b>Introducing Roxie. Data Management Meets Artificial Intelligence. | ||
+ | </b><br>Introducing Roxie, Rubrik's Intelligent Personal [[Assistants|Assistant]]. A hackathon project by Manjunath Chinni. Created in 10 hours with the power of Rubrik APIs. | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>RWX2mj_yh5o</youtube> | ||
+ | <b>DAS Webinar: Master Data Management – Aligning Data, Process, and Governance | ||
+ | </b><br>Getting MDM “right” requires a strategic mix of Data Architecture, business process, and Data Governance. | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>7q3-CdmCMoE</youtube> | ||
+ | <b>How to manage Artificial Intelligence Data Collection [Enterprise AI Governance Data Management ] | ||
+ | </b><br>Mind Data AI AI researcher Brian Ka Chan's AI ML DL introduction series. Collecting Data is an important step to the success of Artificial intelligence Program in the 4th industrial Revolution. In the current advancement of Artificial Intelligence technologies, machine learning has always been associated with AI, and in many cases, Machine Learning is considered equivalent of Artifical Intelligence. Machine learning is actually a subset of Artificial Intelligence, this discipline of machine learning relies on data to perform AI training, supervised or unsupervised. On average, 80% of the time that my team spent in AI or Data Sciences projects is about preparing data. Preparing data includes, but not limited to: | ||
+ | Identify Data required, Identify the availability of data, and location of them, Profiling the data, Source the data, Integrating the data, Cleanse the data, and prepare the data for learning | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>sHPY8zIhy60</youtube> | ||
+ | <b>What is Data Governance? | ||
+ | </b><br>Understand what problems a Data Governance program is intended to solve and why the Business Users must own it. Also learn some sample roles that each group might need to play. | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>5Pl671FH6MQ</youtube> | ||
+ | <b>Top 10 Mistakes in Data Management | ||
+ | </b><br>Come learn about the mistakes we most often see organizations make in managing their data. Also learn more about Intricity's Data Management Health Check which you can download here: | ||
+ | https://www.intricity.com/intricity101/ To Talk with a Specialist go to: https://www.intricity.com/intricity101/ www.intricity.com | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | |||
+ | = <span id="Versioning"></span>Versioning = | ||
+ | [https://www.youtube.com/results?search_query=~version+versioning+ai YouTube search...] | ||
+ | [https://www.google.com/search?q=~version+versioning+ai ...Google search] | ||
+ | |||
+ | * [https://dvc.org/ DVC | DVC.org] | ||
+ | * [https://www.pachyderm.com/ Pachyderm] …[https://medium.com/bigdatarepublic/pachyderm-for-data-scientists-d1d1dff3a2fa Pachyderm for data scientists | Gerben Oostra - bigdata - Medium] | ||
+ | * [https://www.dataiku.com/ Dataiku] | ||
+ | * [[Algorithm Administration#Continuous Machine Learning (CML)|Continuous Machine Learning (CML)]] | ||
+ | |||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>LQF6vHm_QIY</youtube> | ||
+ | <b>How to manage model and data versions | ||
+ | </b><br>[[Creatives#Raj Ramesh|Raj Ramesh]] Managing data versions and model versions is critical in deploying machine learning models. This is because if you want to re-create the models or go back to fix them, you will need both the data that went into training the model and as well as the model [[Algorithm Administration#Hyperparameter|hyperparameter]]s itself. In this video I explained that concept. | ||
+ | Here's what I can do to help you. I speak on the topics of architecture and AI, help you integrate AI into your organization, educate your team on what AI can or cannot do, and make things simple enough that you can take action from your new knowledge. I work with your organization to understand the nuances and challenges that you face, and together we can understand, frame, analyze, and address challenges in a systematic way so you see improvement in your overall business, is aligned with your strategy, and most importantly, you and your organization can incrementally change to transform and thrive in the future. If any of this sounds like something you might need, please reach out to me at dr.raj.ramesh@topsigma.com, and we'll get back in touch within a day. Thanks for watching my videos and for subscribing. www.topsigma.com www.linkedin.com/in/rajramesh | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>UbL7VUpv1Bs</youtube> | ||
+ | <b>Version Control for Data Science Explained in 5 Minutes (No Code!) | ||
+ | </b><br>In this code-free, five-minute explainer for complete beginners, we'll teach you about Data Version Control (DVC), a tool for adapting Git version control to machine learning projects. | ||
+ | |||
+ | - Why data science and machine learning badly need tools for versioning | ||
+ | - Why Git version control alone will fall short | ||
+ | - How DVC helps you use Git with big datasets and models | ||
+ | - Cool features in DVC, like metrics, pipelines, and plots | ||
+ | |||
+ | Check out the DVC open source project on GitHub: https://github.com/iterative/dvc | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>rUTlqpcmiQw</youtube> | ||
+ | <b>How to easily set up and version your Machine Learning pipelines, using Data Version Control (DVC) and Machine Learning Versioning (MLV)-tools | PyData Amsterdam 2019 | ||
+ | </b><br>Stephanie Bracaloni, Sarah Diot-Girard Have you ever heard about Machine Learning versioning solutions? Have you ever tried one of them? And what about automation? Come with us and learn how to easily build versionable pipelines! This tutorial explains through small exercises how to setup a project using DVC and MLV-tools. www.pydata.org | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>IH2gEtxIbqM</youtube> | ||
+ | <b>Alessia Marcolini: Version Control for Data Science | PyData Berlin 2019 | ||
+ | </b><br>Track:PyData Are you versioning your Machine Learning project as you would do in a traditional software project? How are you keeping track of changes in your datasets? Recorded at the PyConDE & PyData Berlin 2019 conference. https://pycon.de | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>Pno7P3fVM7o</youtube> | ||
+ | <b>Introduction to Pachyderm | ||
+ | </b><br>Joey Zwicker A high-level introduction to the core concepts and features of Pachyderm as well as a quick demo. Learn more at: pachyderm.io github.com/pachyderm/pachyderm docs.pachyderm.io | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>YG8VFOZBb2A</youtube> | ||
+ | <b>E05 Pioneering version control for data science with Pachyderm co-founder and CEO Joe Doliner | ||
+ | </b><br>5 years ago, Joe Doliner and his co-founder Joey Zwicker decided to focus on the hard problems in data science, rather than building just another dashboard on top of the existing mess. It's been a long road, but it's really payed off. Last year, after an adventurous journey, they closed a $10m Series A led by Benchmark. In this episode, Erasmus Elsner is joined by Joe Doliner to explore what Pachyderm does and how it scaled from just an idea into a fast growing tech company. Listen to the podcast version | ||
+ | https://apple.co/2W2g0nV | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | |||
+ | == Model Versioning - ModelDB == | ||
+ | * ModelDB: An open-source system for Machine Learning model versioning, metadata, and experiment management | ||
+ | |||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>U0lyF_lHngo</youtube> | ||
+ | <b>Model Versioning Done Right: A ModelDB 2.0 Walkthrough | ||
+ | </b><br>In a field that is rapidly evolving but lacks infrastructure to operationalize and govern models, ModelDB 2.0 provides the ability to version the full modeling process including the underlying data and training configurations, ensuring that teams can always go back and re-create a model, whether to remedy a production incident or to answer a regulatory query. | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>nj7PUD7IazM</youtube> | ||
+ | <b>Model Versioning Why, When, and How | ||
+ | </b><br>Models are the new code. While machine learning models are increasingly being used to make critical product and business decisions, the process of developing and deploying ML models remain ad-hoc. In the “wild-west” of data science and ML tools, versioning, management, and deployment of models are massive hurdles in making ML efforts successful. As creators of ModelDB, an open-source model management solution developed at MIT CSAIL, we have helped manage and deploy a host of models ranging from cutting-edge deep learning models to traditional ML models in finance. In each of these applications, we have found that the key to enabling production ML is an often-overlooked but critical step: model versioning. Without a means to uniquely identify, reproduce, or rollback a model, production ML pipelines remain brittle and unreliable. In this webinar, we draw upon our experience with ModelDB and Verta to present best practices and tools for model versioning and how having a robust versioning solution (akin to Git for code) can streamlining DS/ML, enable rapid deployment, and ensure high quality of deployed ML models. | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | |||
+ | = <span id="Hyperparameter"></span>Hyperparameter = | ||
+ | [https://www.youtube.com/results?search_query=hyperparameter+deep+learning+tuning+optimization+ai YouTube search...] | ||
+ | [https://www.google.com/search?q=hyperparameter+optimization+deep+machine+learning+ML+ai ...Google search] | ||
+ | |||
+ | * [[Backpropagation]] ... [[Feed Forward Neural Network (FF or FFNN)|FFNN]] ... [[Forward-Forward]] ... [[Activation Functions]] ...[[Softmax]] ... [[Loss]] ... [[Boosting]] ... [[Gradient Descent Optimization & Challenges|Gradient Descent]] ... [[Algorithm Administration#Hyperparameter|Hyperparameter]] ... [[Manifold Hypothesis]] ... [[Principal Component Analysis (PCA)|PCA]] | ||
+ | * [[Hypernetworks]] | ||
+ | * [https://cloud.google.com/ml-engine/docs/tensorflow/using-hyperparameter-tuning Using TensorFlow Tuning] | ||
+ | * [https://towardsdatascience.com/understanding-hyperparameters-and-its-optimisation-techniques-f0debba07568 Understanding Hyperparameters and its Optimisation techniques | Prabhu - Towards Data Science] | ||
+ | * [https://nanonets.com/blog/hyperparameter-optimization/ How To Make Deep Learning Models That Don’t Suck | Ajay Uppili Arasanipalai] | ||
+ | * [[Optimization Methods]] | ||
+ | |||
+ | In machine learning, a hyperparameter is a parameter whose value is set before the learning process begins. By contrast, the values of other parameters are derived via training. Different model training algorithms require different hyperparameters, some simple algorithms (such as ordinary least squares regression) require none. Given these hyperparameters, the training algorithm learns the parameters from the data. [https://en.wikipedia.org/wiki/Hyperparameter_(machine_learning) Hyperparameter (machine learning) | Wikipedia] | ||
+ | |||
+ | Machine learning algorithms train on data to find the best set of [[Activation Functions#Weights|weights]] for each independent variable that affects the predicted value or class. The algorithms themselves have variables, called hyperparameters. They’re called hyperparameters, as opposed to parameters, because they control the operation of the algorithm rather than the [[Activation Functions#Weights|weights]] being determined. The most important hyperparameter is often the learning rate, which determines the step size used when finding the next set of [[Activation Functions#Weights|weights]] to try when optimizing. If the learning rate is too high, the gradient descent may quickly converge on a plateau or suboptimal point. If the learning rate is too low, the gradient descent may stall and never completely converge. Many other common hyperparameters depend on the algorithms used. Most algorithms have stopping parameters, such as the maximum number of epochs, or the maximum time to run, or the minimum improvement from epoch to epoch. Specific algorithms have hyperparameters that control the shape of their search. For example, a [[Random Forest (or) Random Decision Forest]] Classifier has hyperparameters for minimum samples per leaf, max depth, minimum samples at a split, minimum [[Activation Functions#Weights|weight]] fraction for a leaf, and about 8 more. [https://www.infoworld.com/article/3394399/machine-learning-algorithms-explained.html Machine learning algorithms explained | Martin Heller - InfoWorld] | ||
+ | |||
+ | https://nanonets.com/blog/content/images/2019/03/HPO1.png | ||
+ | |||
+ | |||
+ | == Hyperparameter Tuning == | ||
+ | Hyperparameters are the variables that govern the training process. Your model parameters are optimized (you could say "tuned") by the training process: you run data through the operations of the model, compare the resulting prediction with the actual value for each data instance, evaluate the accuracy, and adjust until you find the best combination to handle the problem. These algorithms automatically adjust (learn) their internal parameters based on data. However, there is a subset of parameters that is not learned and that have to be configured by an expert. Such parameters are often referred to as “hyperparameters” — and they have a big impact ...For example, the tree depth in a decision tree model and the number of layers in an artificial [[Neural Network]] are typical hyperparameters. The performance of a model can drastically depend on the choice of its hyperparameters. [https://thenextweb.com/podium/2019/11/11/machine-learning-algorithms-and-the-art-of-hyperparameter-selection/ Machine learning algorithms and the art of hyperparameter selection - A review of four optimization strategies | Mischa Lisovyi and Rosaria Silipo - TNW] | ||
+ | |||
+ | There are four commonly used optimization strategies for hyperparameters: | ||
+ | # Bayesian optimization | ||
+ | # Grid search | ||
+ | # Random search | ||
+ | # Hill climbing | ||
+ | |||
+ | Bayesian optimization tends to be the most efficient. You would think that tuning as many hyperparameters as possible would give you the best answer. However, unless you are running on your own personal hardware, that could be very expensive. There are diminishing returns, in any case. With experience, you’ll discover which hyperparameters matter the most for your data and choice of algorithms. [https://www.infoworld.com/article/3394399/machine-learning-algorithms-explained.html Machine learning algorithms explained | Martin Heller - InfoWorld] | ||
+ | |||
+ | Hyperparameter Optimization libraries: | ||
+ | * [https://github.com/maxim5/hyper-engine hyper-engine - Gaussian Process Bayesian optimization and some other techniques, like learning curve prediction] | ||
+ | * [https://ray.readthedocs.io/en/latest/tune.html Ray Tune: Hyperparameter Optimization Framework] | ||
+ | * [https://sigopt.com/ SigOpt’s API tunes your model’s parameters through state-of-the-art Bayesian optimization] | ||
+ | * [https://github.com/hyperopt/hyperopt hyperopt; Distributed Asynchronous Hyperparameter Optimization in Python - random search and tree of parzen estimators optimization.] | ||
+ | * [https://scikit-optimize.github.io/#skopt.Optimizer Scikit-Optimize, or skopt - Gaussian process Bayesian optimization] | ||
+ | * [https://github.com/polyaxon/polyaxon polyaxon] | ||
+ | * [https://github.com/SheffieldML/GPyOpt GPyOpt; Gaussian Process Optimization] | ||
+ | |||
+ | Tuning: | ||
+ | * Optimizer type | ||
+ | * Learning rate (fixed or not) | ||
+ | * Epochs | ||
+ | * Regularization rate (or not) | ||
+ | * Type of Regularization - L1, L2, ElasticNet | ||
+ | * Search type for local minima | ||
+ | ** Gradient descent | ||
+ | ** Simulated | ||
+ | ** Annealing | ||
+ | ** Evolutionary | ||
+ | * Decay rate (or not) | ||
+ | * Momentum (fixed or not) | ||
+ | * Nesterov Accelerated Gradient momentum (or not) | ||
+ | * Batch size | ||
+ | * Fitness measurement type | ||
+ | ** MSE, accuracy, MAE, [[Cross-Entropy Loss]] | ||
+ | ** Precision, recall | ||
+ | |||
+ | * Stop criteria | ||
+ | |||
+ | <youtube>oaxf3rk0KGM</youtube> | ||
+ | <youtube>wKkcBPp3F1Y</youtube> | ||
+ | <youtube>giBAxWeuysM</youtube> | ||
+ | <youtube>WYLoNEcVeZo</youtube> | ||
+ | <youtube>ttE0F7fghfk</youtube> | ||
+ | |||
+ | = Automated Learning = | ||
+ | [https://www.youtube.com/results?search_query=~Automated+~Learning+ai YouTube search...] | ||
+ | [https://www.google.com/search?q=~Automated+~Learning+ai ...Google search] | ||
+ | |||
+ | * [[Artificial General Intelligence (AGI) to Singularity]] ... [[Inside Out - Curious Optimistic Reasoning| Curious Reasoning]] ... [[Emergence]] ... [[Moonshots]] ... [[Explainable / Interpretable AI|Explainable AI]] ... [[Algorithm Administration#Automated Learning|Automated Learning]] | ||
+ | * [[Immersive Reality]] ... [[Metaverse]] ... [[Omniverse]] ... [[Transhumanism]] ... [[Religion]] | ||
+ | * [[Telecommunications]] ... [[Computer Networks]] ... [[Telecommunications#5G|5G]] ... [[Satellite#Satellite Communications|Satellite Communications]] ... [[Quantum Communications]] ... [[Agents#Communication | Communication Agents]] ... [[Smart Cities]] ... [[Digital Twin]] ... [[Internet of Things (IoT)]] | ||
+ | * [[Large Language Model (LLM)]] ... [[Large Language Model (LLM)#Multimodal|Multimodal]] ... [[Foundation Models (FM)]] ... [[Generative Pre-trained Transformer (GPT)|Generative Pre-trained]] ... [[Transformer]] ... [[GPT-4]] ... [[GPT-5]] ... [[Attention]] ... [[Generative Adversarial Network (GAN)|GAN]] ... [[Bidirectional Encoder Representations from Transformers (BERT)|BERT]] | ||
+ | * [[Development]] ... [[Notebooks]] ... [[Development#AI Pair Programming Tools|AI Pair Programming]] ... [[Codeless Options, Code Generators, Drag n' Drop|Codeless, Generators, Drag n' Drop]] ... [[Algorithm Administration#AIOps/MLOps|AIOps/MLOps]] ... [[Platforms: AI/Machine Learning as a Service (AIaaS/MLaaS)|AIaaS/MLaaS]] | ||
+ | * [[AdaNet]] | ||
+ | * [[Creatives]] ... [[History of Artificial Intelligence (AI)]] ... [[Neural Network#Neural Network History|Neural Network History]] ... [[Rewriting Past, Shape our Future]] ... [[Archaeology]] ... [[Paleontology]] | ||
+ | * [[Generative Pre-trained Transformer (GPT)#Generative Pre-trained Transformer 5 (GPT-5) | Generative Pre-trained Transformer 5 (GPT-5)]] | ||
+ | * [https://www.technologyreview.com/s/603381/ai-software-learns-to-make-ai-software/ AI Software Learns to Make AI Software] | ||
+ | * [https://www.nextgov.com/emerging-tech/2018/08/pentagon-wants-ai-take-over-scientific-process/150807/ The Pentagon Wants AI to Take Over the Scientific Process | Automating Scientific Knowledge Extraction (ASKE) | DARPA] | ||
+ | ** [https://www.nextgov.com/emerging-tech/2018/09/inside-pentagons-plan-make-computers-collaborative-partners/151014/ Inside the Pentagon's Plan to Make Computers ‘Collaborative Partners’ | DARPA - Nextgov] | ||
+ | ** [https://www.newscientist.com/article/dn28434-ai-tool-scours-all-the-science-on-the-web-to-find-new-knowledge/ AI tool scours all the science on the web to find new knowledge | Allen Institute for Artificial Intelligence (AI2)] | ||
+ | ** [https://www.fbo.gov/index.php?s=opportunity&mode=form&id=f6149249b0f3c04c5b8994be1a492726&tab=core&tabmode=list&= Program Announcement for Artificial Intelligence Exploration (AIE) | DARPA - FedBizOpps.gov] | ||
+ | * [https://medium.com/applied-data-science/how-to-build-your-own-world-model-using-python-and-keras-64fb388ba459 Hallucinogenic Deep Reinforcement Learning Using Python and Keras | David Foster] | ||
+ | * [https://towardsdatascience.com/automated-feature-engineering-in-python-99baf11cc219 Automated Feature Engineering in Python - How to automatically create machine learning features | Will Koehrsen - Towards Data Science] | ||
+ | * [https://chatbotslife.com/why-meta-learning-is-crucial-for-further-advances-of-artificial-intelligence-c2df55959adf Why Meta-learning is Crucial for Further Advances of Artificial Intelligence? | Pavel Kordik] | ||
+ | * [https://www.darpa.mil/attachments/AssuredAutonomyProposersDay_Program%20Brief.pdf Assured Autonomy | Dr. Sandeep Neema, DARPA] | ||
+ | * [https://www.kdnuggets.com/2019/02/automatic-machine-learning-broken.html Automatic Machine Learning is Broken | Piotr Plonski - KDnuggets] | ||
+ | * [https://www.gigabitmagazine.com/ai/why-2020-will-be-year-automated-machine-learning Why 2020 will be the Year of Automated Machine Learning | Senthil Ravindran - Gigabit] | ||
+ | * [https://en.wikipedia.org/wiki/Meta_learning_(computer_science) Meta Learning | Wikipedia] | ||
+ | |||
+ | Several production machine-learning platforms now offer automatic hyperparameter tuning. Essentially, you tell the system what hyperparameters you want to vary, and possibly what metric you want to optimize, and the system sweeps those hyperparameters across as many runs as you allow. ([[Google Cloud]] hyperparameter tuning extracts the appropriate metric from the TensorFlow model, so you don’t have to specify it.) | ||
+ | |||
+ | An emerging class of data science toolkit that is finally making machine learning accessible to business subject matter experts. We anticipate that these innovations will mark a new era in data-driven decision support, where business analysts will be able to access and deploy machine learning on their own to analyze hundreds and thousands of dimensions simultaneously. Business analysts at highly competitive organizations will shift from using [[visualization]] tools as their only means of analysis, to using them in concert with AML. Data [[visualization]] tools will also be used more frequently to communicate model results, and to build task-oriented user interfaces that enable stakeholders to make both operational and strategic decisions based on output of scoring engines. They will also continue to be a more effective means for analysts to perform inverse analysis when one is seeking to identify where relationships in the data do not exist. [https://www.ironsidegroup.com/2018/06/06/five-essential-capabilities-automated-machine-learning/ 'Five Essential Capabilities: Automated Machine Learning' | Gregory Bonnette] | ||
+ | |||
+ | [[H2O]] Driverless AI automatically performs feature engineering and [[Algorithm Administration#Hyperparameter|hyperparameter]] tuning, and claims to perform as well as Kaggle masters. [[AmazonML]] [[SageMaker]] supports [[Algorithm Administration#Hyperparameter|hyperparameter]] optimization. [[Microsoft]] Azure Machine Learning AutoML automatically sweeps through features, algorithms, and [[Algorithm Administration#Hyperparameter|hyperparameter]]s for basic machine learning algorithms; a separate [https://docs.microsoft.com/en-us/azure/machine-learning/service/concept-automated-ml Azure Machine Learning [[Algorithm Administration#Hyperparameter|hyperparameter]] tuning facility] allows you to sweep specific [[Algorithm Administration#Hyperparameter|hyperparameter]]s for an existing experiment. [https://cloud.google.com/automl/ Google Cloud AutoML] implements automatic deep [[transfer learning]] (meaning that it starts from an existing [[Neural Network#Deep Neural Network (DNN)|Deep Neural Network (DNN)]] trained on other data) and neural architecture search (meaning that it finds the right combination of extra network layers) for language pair translation, natural language classification, and image classification. [https://www.infoworld.com/article/3344596/review-google-cloud-automl-is-truly-automated-machine-learning.html Review: Google Cloud AutoML is truly automated machine learning | Martin Heller] | ||
+ | |||
+ | <img src="https://miro.medium.com/max/588/1*pgTLoLGw0PVaP7ViSyQabA.png" width="600"> | ||
+ | |||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>ynYnZywayC4</youtube> | ||
+ | <b>Hyperparameter Tuning with [[Amazon]] SageMaker's Automatic Model Tuning - AWS Online Tech Talks | ||
+ | </b><br>Learn how to use Automatic Model Tuning with [[Amazon]] SageMaker to get the best machine learning model for your dataset. Training machine models requires choosing seemingly arbitrary hyperparameters like learning rate and regularization to control the learning algorithm. Traditionally, finding the best values for the hyperparameters requires manual trial-and-error experimentation. [[Amazon]] SageMaker makes it easy to get the best possible outcomes for your machine learning models by providing an option to create hyperparameter tuning jobs. These jobs automatically search over ranges of hyperparameters to find the best values. Using sophisticated Bayesian optimization, a meta-model is built to accurately predict the quality of your trained model from the hyperparameters. Learning Objectives: | ||
+ | - Understand what hyperparameters are and what they do for training machine learning models | ||
+ | - Learn how to use Automatic Model Tuning with [[Amazon]] SageMaker for creating hyperparameter tuning of your training jobs | ||
+ | - Strategies for choosing and iterating on tuning ranges of a hyperparameter tuning job with [[Amazon]] SageMaker | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>mSvw0TfxqDo</youtube> | ||
+ | <b>Automatic Hyperparameter Optimization in [[Keras]] for the MediaEval 2018 Medico Multimedia Task | ||
+ | </b><br>Rune Johan Borgli, Pål Halvorsen, Michael Riegler, Håkon Kvale Stensland, Automatic Hyperparameter Optimization in Keras for the MediaEval 2018 Medico Multimedia Task. Proc. of MediaEval 2018, 29-31 October 2018, Sophia Antipolis, France. Abstract: This paper details the approach to the MediaEval 2018 Medico Multimedia Task made by the Rune team. The decided upon approach uses a work-in-progress hyperparameter optimization system called Saga. Saga is a system for creating the best hyperparameter finding in Keras, a popular machine learning framework, using [[Bayes|Bayesian]] optimization and transfer learning. In addition to optimizing the Keras classifier configuration, we try manipulating the dataset by adding extra images in a class lacking in images and splitting a commonly misclassified class into two classes. Presented by Rune Johan Borgli | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | |||
+ | |||
+ | == <span id="AutoML"></span>AutoML == | ||
+ | [https://www.youtube.com/results?search_query=AutoML+ai YouTube search...] | ||
+ | [https://www.google.com/search?q=AutoML+ai ...Google search] | ||
+ | |||
+ | * [https://en.wikipedia.org/wiki/Automated_machine_learning Automated Machine Learning (AutoML) | Wikipedia] | ||
+ | * [https://www.automl.org/ AutoML.org] ...[https://ml.informatik.uni-freiburg.de/ ML Freiburg] ... [https://github.com/automl GitHub] and [https://www.tnt.uni-hannover.de/project/automl/ ML Hannover] | ||
+ | |||
+ | |||
+ | New cloud software suite of machine learning tools. It’s based on [[Google]]’s state-of-the-art research in image recognition called [[Neural Architecture]] Search (NAS). NAS is basically an algorithm that, given your specific dataset, searches for the most optimal neural network to perform a certain task on that dataset. AutoML is then a suite of machine learning tools that will allow one to easily train high-performance deep networks, without requiring the user to have any knowledge of deep learning or AI; all you need is labelled data! [[Google]] will use NAS to then find the best network for your specific dataset and task. [https://www.kdnuggets.com/2018/08/autokeras-killer-google-automl.html AutoKeras: The Killer of Google’s AutoML | George Seif - KDnuggets] | ||
+ | |||
+ | * [https://www.androidauthority.com/google-cloud-automl-vision-guide-894671/ Cloud AutoML Vision: Train your own machine learning model | Jessica Thornsby] | ||
+ | |||
+ | |||
+ | <img src="https://cloud.google.com/images/products/natural-language/automl-nl-works.png" width="800"> | ||
+ | |||
+ | |||
+ | <youtube>PVo3Wu8ZUFk</youtube> | ||
+ | <youtube>MqO_L9nIOWM</youtube> | ||
+ | <youtube>6JZNEb5uDu4</youtube> | ||
+ | <youtube>kgxfdTh9lz0</youtube> | ||
+ | <youtube>aUfIFoMEIgg</youtube> | ||
+ | <youtube>OHIEZ-Scek8</youtube> | ||
+ | <youtube>-zteIdpQ5UE</youtube> | ||
+ | <youtube>GbLQE2C181U</youtube> | ||
+ | <youtube>iBuW8D3hGaI</youtube> | ||
+ | <youtube>fuNt_4DCkd4</youtube> | ||
+ | <youtube>42Oo8TOl85I</youtube> | ||
+ | <youtube>_a9VyXhPnrM</youtube> | ||
+ | <youtube>jn-22XyKsgo</youtube> | ||
+ | |||
+ | == Automatic Machine Learning (AML) == | ||
+ | * [https://www.forbes.com/sites/tomdavenport/2019/09/03/dotdata-and-the-explosion-of-automated-machine-learning/#4549ede92c3a dotData And The Explosion Of Automated Machine Learning | Tom Davenport - Forbes] | ||
+ | |||
+ | <img src="https://thumbor.forbes.com/thumbor/960x0/https%3A%2F%2Fblogs-images.forbes.com%2Ftomdavenport%2Ffiles%2F2019%2F09%2F0-1200x534.jpg" width="800"> | ||
+ | |||
+ | <youtube>OR-IKyP4ZpI</youtube> | ||
+ | <youtube>qpU2RqimexM</youtube> | ||
+ | |||
+ | == Self-Learning == | ||
+ | <youtube>hVv68aHYSs4</youtube> | ||
+ | <youtube>QrJlj0VCHys</youtube> | ||
+ | <youtube>X2tr0lEmslw</youtube> | ||
+ | <youtube>GdTBqBnqhaQ</youtube> | ||
+ | <youtube>VX0vFZF3c2w</youtube> | ||
+ | <youtube>YNLC0wJSHxI</youtube> | ||
+ | <youtube>TnUYcTuZJpM</youtube> | ||
+ | <youtube>0g9SlVdv1PY</youtube> | ||
+ | <youtube>9EN_HoEk3KY</youtube> | ||
+ | |||
+ | == DARTS: Differentiable Architecture Search == | ||
+ | [https://www.youtube.com/results?search_query=DARTS+Differentiable+Architecture+Search YouTube search...] | ||
+ | [https://www.google.com/search?q=Differentiable+Architecture+Search ...Google search] | ||
+ | |||
+ | * [https://arxiv.org/pdf/1806.09055.pdf DARTS: Differentiable Architecture Search | H. Liu, K. Simonyan, and Y. Yang] addresses the scalability challenge of architecture search by formulating the task in a differentiable manner. Unlike conventional approaches of applying evolution or reinforcement learning over a discrete and non-differentiable search space, the method is based on the continuous relaxation of the architecture representation, allowing efficient search of the architecture using gradient descent. | ||
+ | * [https://www.microsoft.com/en-us/research/uploads/prod/2018/12/Neural-Architecture-Search-SLIDES.pdf Neural Architecture Search | Debadeepta Dey - Microsoft Research] | ||
+ | |||
+ | <img src="https://ai2-s2-public.s3.amazonaws.com/figures/2017-08-08/c1f457e31b611da727f9aef76c283a18157dfa83/3-Figure1-1.png" width="600"> | ||
+ | |||
+ | <youtube>wL-p5cjDG64</youtube> | ||
+ | |||
+ | <img src="https://www.ironsidegroup.com/wp-content/uploads/2018/05/aml.png" width="600"> | ||
+ | |||
+ | = <span id="AIOps/MLOps"></span>AIOps/MLOps = | ||
+ | [https://www.youtube.com/results?search_query=~AIOps+~MLOps+~devops+~secdevops+~devsecops+pipeline+toolchain+CI+CD+machine+learning+ai Youtube search...] | ||
+ | [https://www.google.com/search?q=~AIOps+~MLOps+~devops+~secdevops+~devsecops+pipeline+toolchain+CI+CD+machine+learning+ai ...Google search] | ||
+ | |||
+ | * [[Development]] ... [[Notebooks]] ... [[Development#AI Pair Programming Tools|AI Pair Programming]] ... [[Codeless Options, Code Generators, Drag n' Drop|Codeless]] ... [[Hugging Face]] ... [[Algorithm Administration#AIOps/MLOps|AIOps/MLOps]] ... [[Platforms: AI/Machine Learning as a Service (AIaaS/MLaaS)|AIaaS/MLaaS]] | ||
+ | * [[MLflow]] | ||
+ | * [[ChatGPT#DevSecOps| ChatGPT integration with DevSecOps]] | ||
+ | * [https://devops.com/?s=ai DevOps.com] | ||
+ | * [https://www.forbes.com/sites/servicenow/2020/02/26/a-silver-bullet-for-cios/#53a1381e6870 A Silver Bullet For CIOs; Three ways AIOps can help IT leaders get strategic - Lisa Wolfe - Forbes] | ||
+ | * [https://www.forbes.com/sites/tomtaulli/2020/08/01/mlops-what-you-need-to-know/#37b536da1214 MLOps: What You Need To Know | Tom Taulli - Forbes] | ||
+ | * [https://devops.com/what-is-so-special-about-aiops-for-mission-critical-workloads/ What is so Special About AIOps for Mission Critical Workloads? | Rebecca James - DevOps] | ||
+ | * [https://www.bmc.com/blogs/what-is-aiops/ What is AIOps? Artificial Intelligence for IT Operations Explained | BMC] | ||
+ | * [https://www.splunk.com/en_us/it-operations/artificial-intelligence-aiops.html AIOps: Artificial Intelligence for IT Operations, Modernize and transform IT Operations with solutions built on the only Data-to-Everything platform | splunk>] | ||
+ | * [https://www.gartner.com/smarterwithgartner/how-to-get-started-with-aiops/ How to Get Started With AIOps | Susan Moore - Gartner] | ||
+ | * [https://hackernoon.com/why-ai-ml-will-shake-software-testing-up-in-2019-b3f86a30bcfa Why AI & ML Will Shake Software Testing up in 2019 | Oleksii Kharkovyna - Medium] | ||
+ | * [https://martinfowler.com/articles/cd4ml.html Continuous Delivery for Machine Learning D. Sato, A. Wider and C. Windheuser - MartinFowler] | ||
+ | * [[Defense]]: [[Joint Capabilities Integration and Development System (JCIDS)#Adaptive Acquisition Framework (AAF)|Adaptive Acquisition Framework (AAF)]] | ||
+ | |||
+ | |||
+ | Machine learning capabilities give IT operations teams [[context]]ual, actionable insights to make better decisions on the job. More importantly, AIOps is an approach that transforms how systems are automated, detecting important signals from vast amounts of data and relieving the operator from the headaches of managing according to tired, outdated runbooks or policies. In the AIOps future, the environment is continually improving. The administrator can get out of the impossible business of refactoring rules and policies that are immediately outdated in today’s modern IT environment. Now that we have AI and machine learning technologies embedded into IT operations systems, the game changes drastically. AI and machine learning-enhanced automation will bridge the gap between DevOps and IT Ops teams: helping the latter solve issues faster and more accurately to keep pace with business goals and user needs. [https://it.toolbox.com/guest-article/how-aiops-helps-it-operators-on-the-job How AIOps Helps IT Operators on the Job | Ciaran Byrne - Toolbox] | ||
+ | |||
+ | <img src="https://martinfowler.com/articles/cd4ml/cd4ml-end-to-end.png" width="1000" height="500"> | ||
+ | |||
+ | |||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>joTF9BRwWp4</youtube> | ||
+ | <b>MLOps #28 ML Observability // Aparna Dhinakaran - Chief Product Officer at Arize AI | ||
+ | </b><br>MLOps.community As more and more machine learning models are deployed into production, it is imperative we have better observability tools to monitor, troubleshoot, and explain their decisions. In this talk, Aparna Dhinakaran, Co-Founder, CPO of Arize AI (Berkeley-based startup focused on ML Observability), will discuss the state of the commonly seen ML Production Workflow and its challenges. She will focus on the lack of model observability, its impacts, and how Arize AI can help. This talk highlights common challenges seen in models deployed in production, including model drift, [[Data Quality|data quality]]data quality issues, distribution changes, outliers, and bias. The talk will also cover best practices to address these challenges and where observability and explainability can help identify model issues before they impact the business. Aparna will be sharing a demo of how the Arize AI platform can help companies validate their models performance, provide real-time performance monitoring and alerts, and automate troubleshooting of slices of model performance with explainability. The talk will cover best practices in ML Observability and how companies can build more transparency and trust around their models. Aparna Dhinakaran is Chief Product Officer at Arize AI, a startup focused on ML Observability. She was previously an ML engineer at Uber, Apple, and Tubemogul (acquired by Adobe). During her time at Uber, she built a number of core ML Infrastructure platforms including Michaelangelo. She has a bachelors from Berkeley's Electrical Engineering and Computer Science program where she published research with Berkeley's AI Research group. She is on a leave of absence from the Computer Vision PhD program at Cornell University. | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>HwZlGQuCTj4</youtube> | ||
+ | <b>DevOps for AI - [[Microsoft]] | ||
+ | </b><br>DOES18 Las Vegas — Because the AI field is young compared to traditional software [[development]], best practices and solutions around life cycle management for these AI systems have yet to solidify. This talk will discuss how we did this at [[Microsoft]] in different departments (one of them being Bing). DevOps for AI - [[Microsoft]] | ||
+ | Gabrielle Davelaar, Data Platform Solution Architect/A.I., [[Microsoft]] Jordan Edwards, Senior Program Manager, [[Microsoft]] Gabrielle Davelaar is a Data Platform Solution Architect specialized in Artificial Intelligence solutions at [[Microsoft]]. She was originally trained as a computational neuroscientist. Currently she helps [[Microsoft]]’s top 15 Fortune 500 customers build trustworthy and scalable platforms able to create the next generation of A.I. applications. While helping customers with their digital A.I. transformation, she started working with engineering to tackle one key issue: A.I. maturity. The demand for this work is high, and Gabrielle is now working on bringing together the right people to create a full offering. Her aspirations are to be a technical leader in the healthcare digital transformation. Empowering people to find new treatments using A.I. while insuring privacy and taking data governance in consideration. Jordan Edwards is a Senior Program Manager on the Azure AI Platform team. He has worked on a number of highly performant, globally distributed systems across Bing, Cortana and [[Microsoft]] Advertising and is currently working on CI/CD experiences for the next generation of Azure Machine Learning. Jordan has been a key driver of dev-ops modernization in AI+R, including but not limited to: moving to Git, moving the organization at large to CI/CD, packaging and build language modernization, movement from monolithic services to microservice platforms and driving for a culture of friction free devOps and flexible engineering culture. His passion is to continue driving [[Microsoft]] towards a culture which enables our engineering talent to do and achieve more. DOES18 Las Vegas DOES 2018 US DevOps Enterprise Summit 2018 https://events.itrevolution.com/us/ | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>NKN6HawFuts</youtube> | ||
+ | <b>Productionizing Machine Learning with AI Platform Pipelines (Cloud AI Huddle) | ||
+ | </b><br>In this virtual edition of AI Huddle, Yufeng introduces Cloud AI Platform Pipelines, a newly launched product from GCP. We'll first cover the basics of what it is and how to get it set up, and then dive deeper into how to operationalize your ML pipeline in this environment, using the Kubeflow Pipelines SDK. All the code we'll be using is already in the open, so boot up your GCP console and try it out yourself! | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>M8ZG2Oz_NWU</youtube> | ||
+ | <b>How AIOps is transforming customer experience by breaking down DevOps silos at KPN | ||
+ | </b><br>DevOps means dev fails fast and ops fails never. The challenge is that the volume of data exceeds human ability to analyze it and take action before user experience degrades. In this session, Dutch telecom KPN shares how AIOps from Broadcom is helping connect operational systems and increase automation across their toolchain to breakdown DevOps silos and provide faster feedback loops by preventing problems from getting into production. | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>_uDyPoh58Gk</youtube> | ||
+ | <b>Getting to Sub-Second Incident Response with AIOps (FutureStack19) | ||
+ | </b><br>Machine learning and artificial intelligence are revolutionizing the way operations teams work. In this talk, we’ll explore the incident management practices that can get you toward zero downtime, and talk about how New Relic can help you along that path. Speaker: Dor Sasson, Senior Product Manager AIOps, New Relic | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>kmKuEQrO-l4</youtube> | ||
+ | <b>AWS re:Invent 2018: [REPEAT 1] AIOps: Steps Towards Autonomous Operations (DEV301-R1) | ||
+ | </b><br>In this session, learn how to architect a predictive and preventative remediation solution for your applications and infrastructure resources. We show you how to collect performance and operational intelligence, understand and predict patterns using AI & ML, and fix issues. We show you how to do all this by using AWS native solutions: [[Amazon]] SageMaker and [[Amazon]] CloudWatch. | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>UYNygjTY4xw</youtube> | ||
+ | <b>Machine Learning & AIOps: Why IT Operations & Monitoring Teams Should Care | ||
+ | </b><br>In this webinar, we break down what machine learning is, what it can do for your organization, and questions to ask when evaluating ML tools. For more information - please visit out our product pages: https://www.bigpanda.io/our-product/ | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>xUhn0GuK_4g</youtube> | ||
+ | <b>Anthony Bulk @ NTT DATA, "AI/ML Democratization" | ||
+ | </b><br>North Technology People would like to welcome you to the CAMDEA Digital Forum for Tuesday 15th September Presenter: Anthony Bulk Job title: Director AI & Data | ||
+ | Presentation: AI/ML Democratization | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>_KbKjux7TCM</youtube> | ||
+ | <b>AIOps: The next 5 years | ||
+ | </b><br>A Conversation with Will Cappelli, CTO of EMEA, Moogsoft & Rich Lane, Senior Analyst, Forrester In this webinar, the speakers discuss the role that AIOps can and will play in the enterprise of the future, how the scope of AIOps platforms will expand, and what new functionality may be deployed. Watch this webinar to learn how AIOps will enable and support the digitalization of key business processes, and what new AI technologies and algorithms are likely to have the most impact on the continuing evolution of AIOps. For more information, visit our website www.moogsoft.com | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>sC1FTcuu3sc</youtube> | ||
+ | <b>Building an MLOps Toolchain The Fundamentals | ||
+ | </b><br>Artificial intelligence and machine learning are the latest “must-have” technologies in helping organizations realize better business outcomes. However, most organizations don’t have a structured process for rolling out AI-infused applications. Data scientists create AI models in isolation from IT, which then needs to insert those models into applications—and ensure their security—to deliver any business value. In this ebook/webinar, we examine the best way to set up an MLOps process to ensure successful delivery of AI-infused applications. | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>VCUDo9umKEQ</youtube> | ||
+ | <b>Webinar: MLOps automation with Git Based CI/CD for ML | ||
+ | </b><br>Deploying AI/ML based applications is far from trivial. On top of the traditional DevOps challenges, you need to foster collaboration between multidisciplinary teams (data-scientists, data/ML engineers, software developers and DevOps), handle model and experiment versioning, data versioning, etc. Most ML/AI deployments involve significant manual work, but this is changing with the introduction of new frameworks that leverage cloud-native paradigms, Git and Kubernetes to automate the process of ML/AI-based application deployment. In this session we will explain how ML Pipelines work, the main challenges and the different steps involved in producing models and data products (data gathering, preparation, training/AutoML, validation, model deployment, drift monitoring and so on). We will demonstrate how the [[development]] and deployment process can be greatly simplified and automated. We’ll show how you can: a. maximize the efficiency and collaboration between the various teams, b. harness Git review processes to evaluate models, and c. abstract away the complexity of Kubernetes and DevOps. We will demo how to enable continuous delivery of machine learning to production using Git, CI frameworks (e.g. GitHub Actions) with hosted Kubernetes, Kubeflow, MLOps orchestration tools (MLRun), and [[Serverless]] functions (Nuclio) using real-world application examples. Presenter: Yaron Haviv, Co-Founder and CTO @Iguazio | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>MSTYOOCg4bg</youtube> | ||
+ | <b>How RealPage Leveraged Full Stack Visibility and Integrated AIOps for SaaS Innovation and Customer S | ||
+ | </b><br>[[Development]] and operation teams continue to struggle with having a unified view of their applications and infrastructure. | ||
+ | In this webinar, you'll learn how RealPage, the industry leader in SaaS-based Property Management Solutions, leverages an integrated AppDynamics and Virtana AIOps solution to deliver a superior customer experience by managing the performance of its applications as well as its infrastructure. RealPage must ensure its applications and infrastructure are always available and continuously performing, while constantly innovating to deliver new capabilities to sustain their market leadership. The combination of their Agile [[development]] process and highly virtualized infrastructure environment only adds to the complexity of managing both. To meet this challenge, RealPage is leveraging the visibility and AIOps capabilities delivered by the integrated solutions from AppDynamics and Virtana. | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>P5wcE4IwKgQ</youtube> | ||
+ | <b>Machine Learning on Kubernetes with Kubeflow | ||
+ | </b><br>[[Google[[ Cloud Platform Join Fei and Ivan as they talk to us about the benefits of running your [[TensorFlow]] models in Kubernetes using Kubeflow. Working on a cool project and want to get in contact with us? Fill out Don't forget to subscribe to the channel! → https://goo.gl/UzeAiN this form → https://take5.page.link/csf1 Watch more Take5 episodes here → https://bit.ly/2MgTllk | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>gWgy3EdDObQ</youtube> | ||
+ | <b>PipelineAI End-to-End [[TensorFlow]] Model Training + Deploying + Monitoring + Predicting (Demo) | ||
+ | </b><br>100% open source and reproduce-able on your laptop through Docker - or in production through Kubernetes! Details at https://pipeline.io and https://github.com/fluxcapacitor/pipeline End-to-end pipeline demo: Train and deploy a [[TensorFlow]] model from research to live production. Includes full metrics and insight into the offline training and online predicting phases. | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>PHVtFQpAbsY</youtube> | ||
+ | <b>Deep Learning Pipelines: Enabling AI in Production | ||
+ | </b><br>Deep learning has shown tremendous successes, yet it often requires a lot of effort to leverage its power. Existing deep learning frameworks require [[Writing/Publishing|writing]] a lot of code to run a model, let alone in a distributed manner. Deep Learning Pipelines is an Apache Spark Package library that makes practical deep learning simple based on the Spark MLlib Pipelines API. Leveraging Spark, Deep Learning Pipelines scales out many compute-intensive deep learning tasks. In this talk, we discuss the philosophy behind Deep Learning Pipelines, as well as the main tools it provides, how they fit into the deep learning ecosystem, and how they demonstrate Spark's role in deep learning. About: [[Databricks]] provides a unified data analytics platform, powered by Apache Spark™, that accelerates innovation by unifying data science, engineering and business. Website: https://databricks.com | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>qL8t1gkzs8M</youtube> | ||
+ | <b>PipelineAI: High Performance Distributed [[TensorFlow]] AI + GPU + Model Optimizing Predictions | ||
+ | </b><br>We will each build an end-to-end, continuous [[TensorFlow]] AI model training and deployment pipeline on our own GPU-based cloud instance. At the end, we will combine our cloud instances to create the LARGEST Distributed [[TensorFlow]] AI Training and Serving Cluster in the WORLD! Pre-requisites Just a modern browser and an internet connection. We'll provide the rest! Agenda Spark ML [[TensorFlow]] AI Storing and Serving Models with HDFS Trade-offs of CPU vs. *GPU, Scale Up vs. Scale Out | ||
+ | CUDA + cuDNN GPU [[Development]] Overview [[TensorFlow]] Model Checkpointing, Saving, Exporting, and Importing Distributed [[TensorFlow]] AI Model Training (Distributed [[TensorFlow]]) [[TensorFlow]]'s Accelerated Linear Algebra Framework (XLA) [[TensorFlow]]'s Just-in-Time (JIT) Compiler, Ahead of Time (AOT) Compiler Centralized Logging and Visualizing of Distributed [[TensorFlow]] Training (Tensorboard) Distributed [[TensorFlow]] AI Model Serving/Predicting ([[TensorFlow]] Serving) Centralized Logging and Metrics Collection (Prometheus, Grafana) Continuous [[TensorFlow]] AI Model Deployment ([[TensorFlow]], Airflow) Hybrid Cross-Cloud and On-Premise Deployments (Kubernetes) High-Performance and Fault-Tolerant Micro-services (NetflixOSS) https://pipeline.ai | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>f_-3rQoudnc</youtube> | ||
+ | <b>Bringing Your Data Pipeline into The Machine Learning Era - Chris Gaun & Jörg Schad, Mesosphere | ||
+ | </b><br>Want to view more sessions and keep the conversations going? Join us for KubeCon + CloudNativeCon North America in Seattle, December 11 - 13, 2018 (https://bit.ly/KCCNCNA18) or in Shanghai, November 14-15 (https://bit.ly/kccncchina18). Bringing Your Data Pipeline into The Machine Learning Era - Chris Gaun, Mesosphere (Intermediate Skill Level) Kubeflow is a new tool that makes it easy to run distributed machine learning solutions (e.g. Tensorflow) on Kubernetes. However, much of the data that can feed machine learning algorithms is already in existing distributed data stores. This presentation shows how to connect existing distributed data services running on Apache Mesos to Tensorflow on Kubernetes using the Kubeflow tool. Chris Gaun will show you how this existing data can now leverage machine learning, such as Tensorflow, on Kubernetes using the Kubeflow tool. These lessons can be extrapolated to any local distributed data. Chris Gaun is a CNCF ambassador and product marketing manager at Mesosphere. He has presented at Kubecon in 2016 and has put on over 40 free Kubernetes workshops across US and EU in 2017. About Jörg He is a technical lead at Mesosphere https://kubecon.io. The conference features presentations from developers and end users of Kubernetes, Prometheus, Envoy and all of the other CNCF-hosted projects. Learn more at https://bit.ly/2XTN3ho. The conference features presentations from developers and end users of Kubernetes, Prometheus, Envoy and all of the other CNCF-hosted projects. | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>RY7p9fNmi2E</youtube> | ||
+ | <b>Deep Dive into Deep Learning Pipelines - Sue Ann Hong & Tim Hunter | ||
+ | </b><br>"Deep learning has shown tremendous successes, yet it often requires a lot of effort to leverage its power. Existing deep learning frameworks require writing a lot of code to run a model, let alone in a distributed manner. Deep Learning Pipelines is a Spark Package library that makes practical deep learning simple based on the Spark MLlib Pipelines API. Leveraging Spark, Deep Learning Pipelines scales out many compute-intensive deep learning tasks. In this talk we dive into - the various use cases of Deep Learning Pipelines such as prediction at massive scale, transfer learning, and hyperparameter tuning, many of which can be done in just a few lines of code. - how to work with complex data such as images in Spark and Deep Learning Pipelines. - how to deploy deep learning models through familiar Spark APIs such as MLlib and Spark SQL to empower everyone from machine learning practitioners to business analysts. Finally, we discuss integration with popular deep learning frameworks. About: [[Databricks]] provides a unified data analytics platform, powered by Apache Spark™, that accelerates innovation by unifying data science, engineering and business. Connect with us: | ||
+ | Website: https://databricks.com | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>8ZE5x5VAnQo</youtube> | ||
+ | <b>How to Build Machine Learning Pipelines in a Breeze with Docker | ||
+ | </b><br>Docker Let's review and see in action a few projects based on docker containers that could help you prototype ML based projects (detect faces, nudity, evaluate sentiment, building a smart bot) within a few hours. Let´s see in practice where containers could help in this type of projects. | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>UmCB9ycz55Q</youtube> | ||
+ | <b>End to End Streaming ML Recommendation Pipeline Spark 2 0, Kafka, [[TensorFlow]] Workshop | ||
+ | </b><br>End to End Streaming ML Recommendation Pipeline Spark 2 0, Kafka, TensorFlow Workshop Presented at Bangalore Apache Spark Meetup by Chris Fregly on 10/12/2016. Connect with Chris Fregly at https://www.linkedin.com/in/cfregly https://twitter.com/cfregly https://www.slideshare.net/cfregly | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>8-LriJaYuQw</youtube> | ||
+ | <b>Machine Learning with caret: Building a pipeline | ||
+ | </b><br>Building a Machine Learning model with RandomForest and caret in R. | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>84gqSbLcBFE</youtube> | ||
+ | <b>Let’s Write a Pipeline - Machine Learning Recipes #4 | ||
+ | </b><br>[[Google]] Developer In this episode, we’ll write a basic pipeline for supervised learning with just 12 lines of code. Along the way, we'll talk about training and testing data. Then, we’ll work on our intuition for what it means to “learn” from data. Check out [[TensorFlow]] Playground: https://goo.gl/cv7Dq5 | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>URdnFlZnlaE</youtube> | ||
+ | <b>Kevin Goetsch | Deploying Machine Learning using sklearn pipelines | ||
+ | </b><br>PyData Chicago 2016 Sklearn pipeline objects provide an framework that simplifies the lifecycle of data science models. This talk will cover the how and why of encoding feature engineering, estimators, and model ensembles in a single deployable object. | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>ZwwneZ6iU3Y</youtube> | ||
+ | <b>Building online churn prediction ML model using XGBoost, Spark, Featuretools, [[Python]] and GCP | ||
+ | </b><br>Mariusz Jacyno This video shows step by step how to build, evaluate and deploy churn prediction model using [[Python]], Spark, automated feature engineering (Featuretools), extreme gradient boosting algorithm (XGBoost) and [[Google]] ML service. | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>ml4vlXzVFeE</youtube> | ||
+ | <b>MLOps #34 Owned By Statistics: How Kubeflow & MLOps Can Help Secure ML Workloads // David Aronchick | ||
+ | </b><br>While machine learning is spreading like wildfire, very little attention has been paid to the ways that it can go wrong when moving from [[development]] to production. Even when models work perfectly, they can be attacked and/or degrade quickly if the data changes. Having a well understood MLOps process is necessary for ML security! Using Kubeflow, we demonstrated how to the common ways machine learning workflows go wrong, and how to mitigate them using MLOps pipelines to provide reproducibility, validation, versioning/tracking, and safe/compliant deployment. We also talked about the direction for MLOps as an industry, and how we can use it to move faster, with less risk, than ever before. David leads Open Source Machine Learning Strategy at Azure. This means he spends most of his time helping humans to convince machines to be smarter. He is only moderately successful at this. Previously, he led product management for Kubernetes on behalf of [[Google]], launched [[Google]] Kubernetes Engine, and co-founded the Kubeflow project. He has also worked at [[Microsoft]], [[Amazon]] and Chef and co-founded three startups. When not spending too much time in service of electrons, he can be found on a mountain (on skis), traveling the world (via restaurants) or participating in kid activities, of which there are a lot more than he remembers than when he was that age. | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>6gdrwFMaEZ0</youtube> | ||
+ | <b>An introduction to MLOps on Google Cloud | ||
+ | </b><br>The enterprise machine learning life cycle is expanding as firms increasingly look to automate their production ML systems. MLOps is an ML engineering culture and practice that aims at unifying ML system [[development]] and ML system operation enabling shorter [[development]] cycles, increased deployment velocity, and more dependable releases in close alignment with business objectives. Learn how to construct your systems to standardize and manage the life cycle of machine learning in production with MLOps on Google Cloud. Speaker: Nate Keating Watch more: Google Cloud Next ’20: OnAir → https://goo.gle/next2020 | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>w0n-yapq_7w</youtube> | ||
+ | <b>MLOps: Planning for the new normal | ||
+ | </b><br>Join us for a conversation with Diego Oppenheimer, CEO of Algorithmia, and Kahini Shah, Investor at Gradient Ventures, as they discuss the impacts that the global pandemic is having on AI/ML. | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>d1Rcl0x8zTU</youtube> | ||
+ | <b>Artificial Intelligence (AI) Operations in ServiceNow | Share The Wealth | ||
+ | </b><br>Mike Desmond of GlideFast Consulting explains AI Operations in ServiceNow in this Share the Wealth session. | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>yLfWpaYalVI</youtube> | ||
+ | <b>GitOps for MLOps - David Aronchick | ||
+ | </b><br>GitOps Days 2020 Day 1, Session 8 (May 20) While DevOps and GitOps have made huge traction in recent years, many customers struggle to apply these practices to ML workloads. This talk will focus on some best practices to bring GitOps to MLOps for production-grade applications and possibilities for the future. David Aronchick, Head of Open Source ML Strategy, [[Microsoft]] David Aronchick leads open source machine learning strategy at Azure. He spends most of his time helping humans convince machines to be smarter. (He’s only moderately successful at this.) Previously, he led product management for Kubernetes on behalf of [[Google]], launched [[Google]] Kubernetes Engine, and cofounded the Kubeflow project. He’s also worked at [[Microsoft]], [[Amazon]], and Chef and cofounded three startups. When not spending too much time in service of electrons, he can be found on a mountain (on skis), traveling the world (via restaurants), or participating in kid activities, of which there are a lot more than he remembers when he was that age. | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>5L7jD8uqmD0</youtube> | ||
+ | <b>MLOps Demo — A Day in the Life | ||
+ | </b><br>DataRobot MLOps is the way to manage all your production models. In this demo, we will walk through a typical day-in-a-life of a financial services company running a Center of Excellence for delivering ML in Production. We'll see the types of administrative activities our MLOps engineer George performs, and understand how he works with Mary, our lead Data Scientist. George and Mary rely on each other to effectively build, deploy, and manage models for their company over their full lifecycle. | ||
+ | Visit the MLOps product page to find out more: datarobot.com/platform/mlops/ | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>vSgneL06oOk</youtube> | ||
+ | <b>MLOPs topics in the cloud | ||
+ | </b><br>Pragmatic AI Labs A lecture on emerging topics on MLOps in the cloud including Azure, GCP and Sagemaker. | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>KidlhiqSNmM</youtube> | ||
+ | <b>Commit Virtual 2020: MLOps DevOps for Machine Learning | ||
+ | </b><br>GitLab Speaker: Monmayuri Ray The practice of Devops - developing software and operationalizing the [[development]] cycle has been evolving for over a decade. Now, a new addition has joined this holistic [[development]] cycle: Machine Predictions. The emerging art and science of machine learning algorithms integrated into current operational systems is opening new possibilities for engineers, scientists, and architects in the tech world. This presentation will take the audience on a journey in understanding the fundamentals of orchestrating machine predictions using MLOps in this ever-changing, agile world of software [[development]]. You’ll learn how to excel at the craft of DevOps for Machine Learning (ML). Monmayuri will unpack the theoretical constructs and show they apply to real-world scenarios. Get in touch with Sales: https://bit.ly/2IygR7z | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>_Ni6JWdeCew</youtube> | ||
+ | <b>AIDevFest20: Machine Learning Design Patterns for MLOps | ||
+ | </b><br>Speaker: Valliappa Lakshmanan (LAK), [[Google]] Design patterns are formalized best practices to solve common problems when designing a software system. As machine learning moves from being a research discipline to a software one, it is useful to catalog tried-and-proven methods to help engineers tackle frequently occurring problems that crop up during the ML process. In this talk, I will cover five patterns (Workflow Pipelines, Transform, Multimodal Input, Feature Store, Cascade) that are useful in the [[context]] of adding flexibility, resilience and reproducibility to ML in production. For data scientists and ML engineers, these patterns provide a way to apply hard-won knowledge from hundreds of ML experts to your own projects. Anyone designing infrastructure for machine learning will have to be able to provide easy ways for the data engineers, data scientists, and ML engineers to implement these, and other, design patterns. Website: https://gdgdevfest20.xnextcon.com | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>uClvvlfJxqo</youtube> | ||
+ | <b>[[Python]] Lunch Break presents MLOps | ||
+ | </b><br>Deploy your [[Assistants#Chatbot | Chatbot]] using CI/CD by William Arias Leverage DevOps practices in your next [[Assistants#Chatbot | Chatbot]] deployments, create pipelines for training and testing your [[Assistants#Chatbot | Chatbot]] before automating the deployment to staging or production. NLP developers can focus on the [[Assistants#Chatbot | Chatbot]] [[development]] and worry less about infrastructure and environments configuration Bio: Born and raised in Colombia, studied Electronic Engineering and specialized in Digital Electronics, former exchange student in Czech Technical University in the faculty of Data Science. I started my career working in Intel as Field Engineer, then moved to Oracle in Bogota. After 5 years in Oracle as Solution Architect I decided to get deeper in my knowledge of Machine Learning and Databases, relocated to Prague in Czech Republic, where I studied Data Science for a while from the academia [[perspective]], later jumped to Solution Architect in CA technologies, later moved on to Developer of Automations and NLP in Adecco. Currently I work at Gitlab as Technical Marketing Engineer, teaching and evangelizing developers about DevOps | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | |||
+ | == <span id="Large Language Model Operations (LLMOps)"></span>Large Language Model Operations (LLMOps) == | ||
+ | * [https://www.databricks.com/glossary/llmops What Is LLMOps? | Databricks] | ||
+ | |||
+ | |||
+ | Large Language Model Operations (LLMOps) is a set of practices and tools for the operational management of large language models (LLMs) in production environments. LLMOps encompasses the entire lifecycle of LLMs, from development and deployment to monitoring and maintenance. LLMOps is important because LLMs are complex and resource-intensive models that require careful management to ensure that they are performing as expected and meeting the needs of their users. LLMOps tools and practices can help to streamline and automate the LLM lifecycle, making it easier to deploy and manage LLMs in production. | ||
+ | |||
+ | Here are some of the key components of LLMOps: | ||
+ | |||
+ | * <b>Model development</b>: LLMOps tools and practices can help to streamline the process of developing new LLMs, from data collection and preparation to model training and evaluation. | ||
+ | * <b>Model deployment</b>: LLMOps tools and practices can help to deploy LLMs to production environments in a safe and efficient manner. | ||
+ | * <b>Model monitoring</b>: LLMOps tools and practices can help to monitor the performance of LLMs in production and identify any problems early on. | ||
+ | * <b>Model maintenance</b>: LLMOps tools and practices can help to maintain LLMs in production, including fine-tuning models to improve their performance and updating models with new data. | ||
+ | |||
+ | |||
+ | LLMOps is still a relatively new field, but it is rapidly growing as more and more organizations adopt LLMs. There are a number of LLMOps tools and platforms available, both open source and proprietary. Here are some of the benefits of using LLMOps: | ||
+ | |||
+ | * <b>Improved efficiency and productivity</b>: LLMOps tools and practices can help to automate and streamline the LLM lifecycle, freeing up data scientists and engineers to focus on more creative and strategic tasks. | ||
+ | * <b>Reduced costs</b>: LLMOps can help to reduce the costs of developing and managing LLMs by improving efficiency and reducing the risk of errors. | ||
+ | * <b>Increased reliability and performance</b>: LLMOps can help to ensure that LLMs are performing reliably and meeting the needs of their users by providing tools and practices for monitoring and maintaining LLMs in production. | ||
+ | |||
+ | === How is LLMOps different from AIOps/MLOps? === | ||
+ | * [https://wandb.ai/mostafaibrahim17/ml-articles/reports/AIOps-vs-MLOps-vs-LLMOps--Vmlldzo1MTQzODMz AIOps vs. MLOps vs. LLMOps | Mostafa Ibrahim - Weights & Biases] | ||
+ | * [https://www.analyticsvidhya.com/blog/2023/08/llmops-vs-mlops/ LLMOPS vs MLOPS: Choosing the Best Path for AI Development | Analytics Vidhya] | ||
+ | * [https://www.xenonstack.com/blog/aiops-vs-mlops AIOps vs MLOps | Know Everything in Detail | Jagreet Kaur - Xenonstack] | ||
+ | |||
+ | |||
+ | LLMOps is a subset of MLOps that focuses specifically on the operational management of large language models (LLMs). MLOps is a set of practices and tools for managing the entire lifecycle of machine learning models, from development to deployment to production. AIOps is a broader field that focuses on using AI and machine learning to improve IT operations and management. Here is a table that summarizes the key differences between LLMOps, AIOps, and MLOps: | ||
+ | |||
+ | <table> | ||
+ | |||
+ | <tr> | ||
+ | <th>Characteristic</th> | ||
+ | <th>LLMOps</th> | ||
+ | <th>AIOps</th> | ||
+ | <th>MLOps</th> | ||
+ | </tr> | ||
+ | |||
+ | |||
+ | <tr> | ||
+ | <td>Focus</td> | ||
+ | <td>Operational management of LLMs</td> | ||
+ | <td>Use of AI and ML to improve IT operations and management</td> | ||
+ | <td>Operational management of machine learning models</td> | ||
+ | </tr> | ||
+ | <tr> | ||
+ | <td>Use cases</td> | ||
+ | <td>Language-related tasks, such as chatbots, language translation, sentiment analysis, and content generation</td> | ||
+ | <td>IT operations tasks, such as incident management, capacity planning, and performance monitoring</td> | ||
+ | <td>Machine learning tasks, such as fraud detection, product recommendation, and predictive maintenance</td> | ||
+ | </tr> | ||
+ | <tr> | ||
+ | <td>Team composition</td> | ||
+ | <td>NLP specialists, linguists, data scientists with language modeling expertise, and software engineers experienced in building language-related applications</td> | ||
+ | <td>Data scientists, analytics professionals, data engineers, domain experts, and IT professionals</td> | ||
+ | <td>Data scientists, machine learning engineers, software engineers, and DevOps engineers</td> | ||
+ | </tr> | ||
+ | |||
+ | </table> | ||
+ | |||
+ | |||
+ | Here are some examples of how LLMOps is different from AIOps/MLOps: | ||
+ | |||
+ | * <b>Model development</b>: LLMOps tools and practices are specifically designed for developing and training LLMs. For example, LLMOps tools may help to automate the process of collecting and cleaning text data, or they may provide specialized libraries for training LLMs. | ||
+ | * <b>Model deployment</b>: LLMOps tools and practices can help to deploy LLMs to production environments in a way that is optimized for performance and reliability. For example, LLMOps tools may help to scale LLMs to handle large volumes of traffic, or they may provide specialized infrastructure for hosting LLMs. | ||
+ | * <b>Model monitoring and maintenance</b>: LLMOps tools and practices can help to monitor the performance of LLMs in production and identify any problems early on. For example, LLMOps tools may track the accuracy of LLMs on real-world data, or they may detect potential biases in LLMs. | ||
+ | |||
+ | == <span id="Continuous Machine Learning (CML)"></span>Continuous Machine Learning (CML) == | ||
+ | * [https://cml.dev/ Continuous Machine Learning (CML)] ...is Continuous Integration/Continuous Deployment (CI/CD) for Machine Learning Projects | ||
+ | * [https://dvc.org/ DVC | DVC.org] | ||
+ | |||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>9BgIDqAzfuA</youtube> | ||
+ | <b>MLOps Tutorial #1: Intro to Continuous Integration for ML | ||
+ | </b><br>DVCorg Learn how to use one of the most powerful ideas from the DevOps revolution, continuous integration, in your data science and machine learning projects. This hands-on tutorial shows you how to create an automatic model training & testing setup using GitHub Actions and Continuous Machine Learning (CML), two free and open-source tools in the Git ecosystem. Designed for total beginners! We'll be using: GitHub Actions: https://github.com/features/actions CML: https://github.com/iterative/cml | ||
+ | Resources: Code: https://github.com/andronovhopf/wine GitLab support: https://github.com/iterative/cml/wiki | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>xPncjKH6SPk</youtube> | ||
+ | <b>MLOps Tutorial #3: Track ML models with Git & GitHub Actions | ||
+ | </b><br>DVCorg In this tutorial, we'll compare ML models across two different Git branches of a project- and we'll do it in a continuous integration system (GitHub Actions) for automation superpowers! We'll cover: | ||
+ | |||
+ | - Why comparing model metrics takes more than a git diff | ||
+ | - How pipelines, a method for making model training more reproducible, help you standardize model comparisons across Git branches | ||
+ | - How to display a table comparing model performance to the main branch in a GitHub Pull Request | ||
+ | |||
+ | ** Need an intro to GitHub Actions and continuous integration? Check out the first video in this series! https://youtu.be/9BgIDqAzfuA ** | ||
+ | |||
+ | Helpful links: | ||
+ | Dataset: Data on farmers’ adoption of climate change mitigation measures, individual characteristics, risk attitudes and social influences in a region of Switzerland https://www.sciencedirect.com/science/article/pii/S2352340920303048 | ||
+ | Code: https://github.com/elleobrien/farmer | ||
+ | DVC pipelines & metrics documentation: https://dvc.org/doc/start/data-pipelines#data-pipelines | ||
+ | CML project repo: https://github.com/iterative/cml | ||
+ | DVC Discord channel: https://discord.gg/bzA6uY7 | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | |||
+ | == <span id="DevSecOps"></span>DevSecOps == | ||
+ | [https://www.youtube.com/results?search_query=~AIOps+~MLOps+~devops+~secdevops+~devsecops+pipeline+toolchain+CI+CD+machine+learning+ai Youtube search...] | ||
+ | [https://www.google.com/search?q=~AIOps+~MLOps+~devops+~secdevops+~devsecops+pipeline+toolchain+CI+CD+machine+learning+ai ...Google search] | ||
+ | |||
+ | * [[Cybersecurity]] | ||
+ | * [[Containers; Docker, Kubernetes & Microservices]] | ||
+ | * [https://safecode.org/ SafeCode] ...nonprofit organization that brings business leaders and technical experts together to exchange insights and ideas on creating, improving and promoting scalable and effective software security programs. | ||
+ | * [https://techbeacon.com/devops/3-ways-ai-will-advance-devsecops 3 ways AI will advance DevSecOps | Joseph Feiman - TechBeacon] | ||
+ | * [https://dzone.com/articles/leveraging-ai-and-automation-for-successful-devsec Leveraging AI and Automation for Successful DevSecOps | Vishnu Nallani - DZone] | ||
+ | |||
+ | DevSecOps (also known as SecDevOps and DevOpsSec) is the process of integrating secure [[development]] best practices and methodologies into continuous design, development, [[deployment]] and integration processes | ||
+ | |||
+ | <hr> | ||
+ | |||
+ | <b><i><center>Dull, Dangerous, and/or Dirty</center></i></b> | ||
+ | |||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>DzX9Vi_UQ8o</youtube> | ||
+ | <b>DevSecOps : What, Why and How | ||
+ | </b><br>Black Hat In this talk, we shall focus on how a DevOps pipeline can easily be metamorphosed into a DevSecOps and the benefits which can be achieved with this transformation. The talk (assisted with various demos) will focus on developing a DevSecOps pipeline using free/open-source tools in various deployment platforms, i.e. on-premise, cloud native and hybrid scenarios. By Anant Shrivastava [https://www.blackhat.com/us-19/briefings/schedule/#devsecops--what-why-and-how-17058 Full Abstract & Presentation Material] | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>19FfDsadajk</youtube> | ||
+ | <b>DevSecOps State of the Union | ||
+ | </b><br>Clint Gibler, Research Director, NCC Group It’s tough to keep up with the DevSecOps resources out there, or even know where to start. This talk will summarize and distill the unique tips and tricks, lessons learned, and tools discussed in dozens of blog posts and more than 50 conference talks over the past few years, and combine it with knowledge gained from in-person discussions with security leaders at companies with mature security programs. Pre-Requisites: General understanding of the fundamental areas of modern application security programs, including threat modeling, secure code reviews, security training, building security culture/developing security champions, security scanning (static and dynamic analysis tools), monitoring and logging in production, etc. Understanding of how software generally moves from [[development]] to production in agile environments that embrace CI/CD practices. Basic understanding of the principles of network/infrastructure and cloud security. | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>eteN4J4610I</youtube> | ||
+ | <b>Webinar: Critical DevSecOps considerations for multi-cloud Kubernetes | ||
+ | </b><br>Cloud Native Computing Foundation (CNCF) The distributed nature of Kubernetes has turned both legacy infrastructure and traditional cybersecurity approaches on their heads. Organizations building cloud-native environments in their own data centers grapple with operationalizing and scaling Kubernetes clusters, and then ensuring system-wide security from the infrastructure layer all the way up to each container. In this webinar, you’ll hear from two cloud-native experts in infrastructure and security who will offer up valuable insights on: - How containerized applications use compute, storage, and network resources differently than do legacy applications - Why hyper-converged infrastructure is suited for Kubernetes - How a Kubernetes stack should be instrumented for observability - Best practices for implementing system-wide security for multi-cloud Kubernetes Presented by Nutanix and Sysdig Presenters: Sylvain Huguet, Sr. Product Manager - Karbon/Kubernetes @Nutanix Loris Digioanni, CTO & Founder @Sysdig. | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>MTurwl9QDu8</youtube> | ||
+ | <b>DevOps for Beginners Where to Start | ||
+ | </b><br>DevOpsTV Tim Beattie Global Head of Product, Open Innovation Labs, RedHat If you're beginning to learn more about DevOps, you may be confused about where to start or where to focus. The term “DevOps” refers to this idea that we no longer have pure and separated [[development]] streams and operations streams – and to a set of DevOps practices that optimizes and improves both functions. Join this webinar to explore how to get going with this cross-functional way of working that breaks down walls, removes bottlenecks, improves speed of delivery, and increases experimentation. | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>1lvPxEQpwVg</youtube> | ||
+ | <b>DevSecOps Days DC 2020 | ||
+ | </b><br>DevSecOps Days is a global series of one and two day conferences helping to educate, evolve, and explore concepts around developing security as code. Join us online to meet fellow practitioners integrating security into their DevOps practices, learn about their journeys, share ideas on integrating security into your teams, and trade insights on automating security tools in your DevOps pipeline. | ||
+ | |||
+ | Welcome (Hasan Yasar) | ||
+ | 00:00--10:47 | ||
+ | |||
+ | The Need for Threat Modeling in a DevSecOps World (Simone Curzi) | ||
+ | 10:48--44:50 | ||
+ | |||
+ | Close the Gap: Bringing Engineering & Security/Compliance Together (Sare Ehmann) | ||
+ | 49:15--1:20:29 | ||
+ | |||
+ | Delivering DevSecOps Value Faster? Use Metrics and Standards (Steven Woodward) | ||
+ | 1:24:15--1:54:13 | ||
+ | |||
+ | Navigating DevOps Requirements with Risk Management (Victoria (Vicky) Hailey) | ||
+ | 2:02:03--2:25:05 | ||
+ | |||
+ | Make it Personal to Make it Happen (Ruth Lennon) | ||
+ | 02:34:11--03:03:00 | ||
+ | |||
+ | Keynote: Successful DevSecOps Implementation for Production Environments is a Journey Not a Milestone (Christopher Brazier) | ||
+ | 03:44:11--04:29:24 | ||
+ | |||
+ | Evolve Your DevSecOps to Manage Speed and Risk (Altaz Valani) | ||
+ | 04:34:09--05:03:12 | ||
+ | |||
+ | How Software Acquisition and DevSecOps Can Improve the Lethality of the [[Defense|DOD]] (Sean Brady) | ||
+ | 05:14:05--05:42:48 | ||
+ | |||
+ | Why Organizations Need DevSecOps Now More Than Ever (Krishna Guru) | ||
+ | 05:47:41--06:14:05 | ||
+ | |||
+ | Wrap-up (HasanYasar) | ||
+ | 06:14:17--06:26:29 | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>ud9QtQFQjVk</youtube> | ||
+ | <b>DevSecOps Days Pittsburgh 2020 | ||
+ | </b><br>Software Engineering Institute | Carnegie Mellon University Speaker Presentations (with timestamps) | ||
+ | |||
+ | Welcome and Introductions | ||
+ | Hasan Yasar 00:00:32 – 00:17:13 | ||
+ | |||
+ | Keynote: Core Values and Customer-Focused Delivery: DevOps in Special Operations | ||
+ | Jason Turse 00:27:57 – 01:11:05 | ||
+ | |||
+ | Are We Building Just Containers or Secure Containers? | ||
+ | Arun Chaudhary 01:15:28 – 01:43:48 | ||
+ | https://resources.sei.cmu.edu/library... | ||
+ | |||
+ | Are your Pipelines Secure? Lock-down Unsecured CI/CD Pipelines | ||
+ | Angel Rivera 01:47:25 – 02:20:05 | ||
+ | |||
+ | Risk: A Bias Toward Action | ||
+ | Jason Lamb 02:30:25 – 02:55:33 | ||
+ | |||
+ | The Innovation Ninja | ||
+ | Amber Vanderburg 03:00:30 – 03:30:34 | ||
+ | |||
+ | Security is an Awesome Product Feature | ||
+ | Mark P. Hahn 04:16:57 – 04:43:08 | ||
+ | https://resources.sei.cmu.edu/library... | ||
+ | |||
+ | Exhilarating Journey of Transformation into Digital Edge | ||
+ | Himanshu Patel 04:47:57 – 05:16:12 | ||
+ | https://resources.sei.cmu.edu/library... | ||
+ | |||
+ | Invited Speaker: Intentional Culture Architecture | ||
+ | Jeff Shupack 05:20:11 – 05:40:17 | ||
+ | |||
+ | Service Mess to Service Mesh | ||
+ | Rob Richardson 05:59:56 – 06:26:35 | ||
+ | https://resources.sei.cmu.edu/library... | ||
+ | |||
+ | Code to Kubernetes: Deployment Shouldn’t be an Afterthought | ||
+ | Lakmal Warusawithana 06:32:25 – 06:53:30 | ||
+ | |||
+ | Design Thinking, Security, and Your Life | ||
+ | Alex Cowan 06:59:56 – 07:23:28 | ||
+ | https://resources.sei.cmu.edu/library... | ||
+ | |||
+ | Keynote: How to Use Hacker Personas to Successfully Build DevSecOps Pipeline | ||
+ | Robin Yeman 07:29:54 – 08:00:45 | ||
+ | https://resources.sei.cmu.edu/library/... | ||
+ | |||
+ | Outbrief and Closing Remarks | ||
+ | Suzanne Miller • Hasan Yasar 08:12:18 – 08:26:42 | ||
+ | |||
+ | Meet fellow practitioners integrating security into their DevOps practices. Learn about their journeys, share ideas on integrating security into your teams, and trade insights on automating security tools in your DevOps pipeline. | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | |||
+ | === <span id="DevSecOps in Government"></span>DevSecOps in Government === | ||
+ | * [[Evaluation]]: [https://tech.gsa.gov/guides/dev_sec_ops_guide/ DevSecOps Guide | General Services Administration (GSA)] | ||
+ | * [[Joint_Capabilities_Integration_and_Development_System_(JCIDS)#Cybersecurity & Acquisition Lifecycle Integration|Cybersecurity & Acquisition Lifecycle Integration]] | ||
+ | * [https://dodcio.defense.gov/Portals/0/Documents/DoD%20Enterprise%20DevSecOps%20Reference%20Design%20v1.0_Public%20Release.pdf?ver=2019-09-26-115824-583 Defense|DOD Enterprise DevSecOps Reference Design Version 1.0 12 August 2019 | Department of Defense (DOD) Chief Information Officer (CIO] | ||
+ | * [https://tech.gsa.gov/guides/understanding_differences_agile_devsecops/ Understanding the Differences Between Agile & DevSecOps - from a Business Perspective | General Services Administration (GSA)] | ||
+ | |||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>2STTK52eAbM</youtube> | ||
+ | <b>[[Defense|DOD]] Enterprise DevSecOps Initiative | ||
+ | </b><br>CSIAC The current [[Defense|Department of Defense (DOD)]] software acquisition process is not responsive to the needs of our warfighters. Therefore, it is difficult for the DOD to keep pace with our potential adversaries and avoid falling behind them. To address this situation, the [[Defense|DOD]] is pursuing a new software [[development]] activity called the [[Defense|DOD]] Enterprise DevSecOps Initiative. This webinar will present the vision for transforming [[Defense|DOD]] software acquisition into secure, responsive software factories. It will examine and explore the utilization of modern software [[development]] processes and tools to revolutionize the Department’s ability to provide responsive, timely, and secure software capabilities for our warfighters. The focus of the effort involves exploiting automated software tools, services, and standards so warfighters can rapidly create, deploy, and operate software applications in a secure, flexible, and interoperable manner. [https://www.csiac.org/podcast/dod-enterprise-devsecops-initiative/ Slides] | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>mE9HMNn7_Pg</youtube> | ||
+ | <b>DevSecOps Implementation in the [[Defense|DoD]]: Barriers and Enablers | ||
+ | </b><br>Software Engineering Institute | Carnegie Mellon University Today's [[Defense|DOD]] software [[development]] and deployment is not responsive to warfighter needs. As a result, the [[Defense|DOD]]'s ability to keep pace with potential adversaries is falling behind. In this webcast, panelists discuss potential enablers of and barriers to using modern software [[development]] techniques and processes in the [[Defense|DOD]] or similar segregated environments. These software [[development]] techniques and processes are as commonly known as DevSecOps. | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>rCqFMKq9ZwA</youtube> | ||
+ | <b>Andrew Weiss/Harshal Dharia: Government DevSecOps with Azure and GitHub | ||
+ | </b><br>Andrew Weiss, Federal Solutions Engineer, GitHub and Harshal Dharia, Senior Azure Specialist, [[Microsoft]] Federal presents Government DevSecOps with Azure and GitHub at the Azure Government February 2020 meetup. | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>J-Ew13L6F9c</youtube> | ||
+ | <b>Infrastructure: Foundations of the Future | ||
+ | </b><br>GovernmentCIO Media & Research As agencies continue to modernize and move away from their legacy systems, how are they reshaping their IT infrastructure and leveraging data to transform their enterprise? Hear from top government IT leaders at the VA, [[Defense|DOD]], and HHS about how they’re adopting cloud and data center infrastructure, strategies and capabilities to move their organizations toward the future. | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | |||
+ | === Strangler Fig / Strangler Pattern === | ||
+ | * [https://martinfowler.com/bliki/StranglerFigApplication.html Strangler Fig Application | Martin Fowler] | ||
+ | * [https://www.michielrook.nl/2016/11/strangler-pattern-practice/ The Strangler pattern in practice | Michiel Rook] | ||
+ | * [https://docs.microsoft.com/en-us/azure/architecture/patterns/strangler Strangler pattern] ...[https://docs.microsoft.com/en-us/azure/architecture/patterns/ Cloud Design Patterns | ][[Microsoft]] | ||
+ | * [https://www.castsoftware.com/blog/how-to-use-strangler-pattern-for-microservices-modernization How to use strangler pattern for microservices modernization | N. Natean - Software Intelligence Plus] | ||
+ | * [https://medium.com/@rmmeans/serverless-strangler-pattern-on-aws-31c88191268d Serverless Strangler Pattern on AWS | Ryan Means - Medium] ... [[Serverless]] | ||
+ | |||
+ | Strangulation of a legacy or undesirable solution is a safe way to phase one thing out for something better, cheaper, or more expandable. You make something new that obsoletes a small percentage of something old, and put them live together. You do some more work in the same style, and go live again (rinse, repeat). | ||
+ | [https://paulhammant.com/2013/07/14/legacy-application-strangulation-case-studies/ Strangler Applications | Paul Hammant] ...case studies | ||
+ | |||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>jm_ZBc5z6B0</youtube> | ||
+ | <b>"Choking the Monolith − The Strangler Pattern Applied" by Tobias Goeschel (@w3ltraumpirat) | ||
+ | </b><br>The so called "Strangler Fig" (aka "Strangler") pattern is a much cited strategy for replacing legacy systems with new, often microservice-based architectures. However, it's not actually a microservice pattern, and there are several - quite different - ways to implement it. Are you confused yet? Fear not: We will have a look at the theory, and then explore together, how Strangler Fig could be used to improve and replace a project worthy being called "The Most Horrible Piece Of Code in the World". | ||
+ | === As a principal consultant at codecentric, Tobias has seen, improved and survived enough of "the code that earns our money" to enjoy it. [https://www.slideshare.net/TobiasGoeschel/choking-the-monolith-the-strangler-fig-pattern-applied Slides] | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>E2dnSg-IHdo</youtube> | ||
+ | <b>AWS New York Summit 2019: Migrating Monolithic Applications with the Strangler Pattern (FSV303) | ||
+ | </b><br>Learn more about [[Amazon]] AWS Global Summits at – [https://amzn.to/2Obv2Hs “Lifting and shifting”] an enterprise-scale application will yield some of the benefits of the cloud, but elasticity and agility may still be limited. Conversely, rewriting that application to be cloud-native can be costly in terms of both time and money and could cause you to miss market opportunities. This session explores the challenges financial institutions face when migrating their existing portfolio of applications to the cloud. Then, we share practical tips to migrate applications to realize cloud-native architectural benefits incrementally using the strangler pattern. | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | |||
+ | = <span id="Model Monitoring"></span>Model Monitoring = | ||
+ | [https://www.youtube.com/results?search_query=model+monitoring+machine+learning+ai YouTube search...] | ||
+ | [https://www.google.com/search?q=model+monitoring+machine+learning+ML+artificial+intelligence+ai ...Google search] | ||
+ | |||
+ | * [https://www.quora.com/How-do-you-evaluate-the-performance-of-a-machine-learning-model-thats-deployed-into-production How do you evaluate the performance of a machine learning model that's deployed into production? | Quora] | ||
+ | * [https://towardsdatascience.com/why-your-models-need-maintenance-faff545b38a2 Why your Models need Maintenance | Martin Schmitz - Towards Data Science] ...Change of concept & drift of Concept | ||
+ | * [https://www.analyticsvidhya.com/blog/2019/10/deployed-machine-learning-model-post-production-monitoring/ Deployed your Machine Learning Model? Here’s What you Need to Know About Post-Production Monitoring | Om Deshmukh - Analytics Vidhya] ...proactive & reactive model monitoring | ||
+ | |||
+ | Monitoring production systems is essential to keeping them running well. For ML systems, monitoring becomes even more important, because their performance depends not just on factors that we have some control over, like infrastructure and our own software, but also on data, which we have much less control over. Therefore, in addition to monitoring standard metrics like latency, traffic, errors and saturation, we also need to monitor model prediction performance. An obvious challenge with monitoring model performance is that we usually don’t have a verified label to compare our model’s predictions to, since the model works on new data. In some cases we might have some indirect way of assessing the model’s effectiveness, for example by measuring click rate for a recommendation model. In other cases, we might have to rely on comparisons between time periods, for example by calculating a percentage of positive classifications hourly and alerting if it deviates by more than a few percent from the average for that time. Just like when validating the model, it’s also important to monitor metrics across slices, and not just globally, to be able to detect problems affecting specific segments. [https://towardsdatascience.com/ml-ops-machine-learning-as-an-engineering-discipline-b86ca4874a3f ML Ops: Machine Learning as an Engineering Discipline | Cristiano Breuel - Towards Data Science] | ||
+ | |||
+ | |||
+ | <img src="https://miro.medium.com/max/1000/1*U7Efc4rSPsXDTeKRzM86eg.png" width="800"> | ||
+ | |||
+ | |||
+ | |||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>zbS9jBB8fz8</youtube> | ||
+ | <b>Josh Wills: Visibility and Monitoring for Machine Learning Models | ||
+ | </b><br>Josh started our Meetup with a short talk on deploying machine learning models into production. He's worked as the Director of Data Science at Cloudera, he wrote the Java version of [[Google]]'s AB testing framework, and he recently held the position of Director of Data Engineering at Slack. In his opinion the most important question is: "How often do you want to deploy this?" You should never deploy a machine learning model once. If the problem is not important enough to keep working on it and deploy new models, then its not important enough to pay the cost of putting it into production in the first place. Watch his talk to get his thoughts on testing machine learning models in production. | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>woRmeGvOaz4</youtube> | ||
+ | <b>Concept Drift: Monitoring Model Quality in Streaming Machine Learning Applications | ||
+ | </b><br>Most machine learning algorithms are designed to work on stationary data. Yet, real-life streaming data is rarely stationary. Models lose prediction accuracy over time if they are not retrained. Without model quality monitoring, retraining decisions are suboptimal and costly. Here, we review the monitoring methods and evaluate them for applicability in modern fast data and streaming applications. | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>IqKunD0Bl5c</youtube> | ||
+ | <b>Monitoring models in production - Jannes Klaas | ||
+ | </b><br>PyData Amsterdam 2018 A Data Scientists work is not done once machine learning models are in production. In this talk, Jannes will explain ways of monitoring Keras neural network models in production, how to track model decay and set up alerting using Flask, Docker and a range of self-built tools. www.pydata.org | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>x4oIdooDzKo</youtube> | ||
+ | <b>Machine Learning Models in Production | ||
+ | </b><br>Data Scientists and Machine Learning practitioners, nowadays, seem to be churning out models by the dozen and they continuously experiment to find ways to improve their accuracies. They also use a variety of ML and DL frameworks & languages , and a typical organization may find that this results in a heterogenous, complicated bunch of assets that require different types of runtimes, resources and sometimes even specialized compute to operate efficiently. But what does it mean for an enterprise to actually take these models to "production" ? How does an organization scale inference engines out & make them available for real-time applications without significant latencies ? There needs to be different techniques for batch ,offline, inferences and instant, online scoring. Data needs to be accessed from various sources and cleansing, transformations of data needs to be enabled prior to any predictions. In many cases, there maybe no substitute for customized data handling with scripting either. Enterprises also require additional auditing and authorizations built in, approval processes and still support a "continuous delivery" paradigm whereby a data scientist can enable insights faster. Not all models are created equal, nor are consumers of a model - so enterprises require both metering and allocation of compute resources for SLAs. In this session, we will take a look at how machine learning is operationalized in IBM Data Science Experience, DSX, a Kubernetes based offering for the Private Cloud and optimized for the HortonWorks Hadoop Data Platform. DSX essentially brings in typical software engineering [[development]] practices to Data Science, organizing the dev-test-production for machine learning assets in much the same way as typical software deployments. We will also see what it means to deploy, monitor accuracies and even rollback models & custom scorers as well as how API based techniques enable consuming business processes and applications to remain relatively stable amidst all the chaos. Speaker: Piotr Mierzejewski, Program Director [[Development]] [[IBM]] DSX Local, [[IBM]] | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | |||
+ | == <span id="A/B Testing"></span>A/B Testing == | ||
+ | [https://www.youtube.com/results?search_query=AB+A/B+~test+~Scoring+~score+~ai YouTube search...] | ||
+ | [https://www.google.com/search?q=AB+A/B+~test+~Scoring+~score+~ai ...Google search] | ||
+ | |||
+ | * [https://www.optimizely.com/ Optimizely] | ||
+ | * [[Math for Intelligence#P-Value|P-Value]] | ||
+ | * [[Math for Intelligence#Confidence Interval|Confidence Interval]] | ||
+ | |||
+ | <b>A/B testing (also known as bucket testing or split-run testing)</b> is a user experience research methodology. A/B tests consist of a randomized experiment with two variants, A and B. It includes application of statistical hypothesis testing or "two-sample hypothesis testing" as used in the field of statistics. A/B testing is a way to compare two versions of a single variable, typically by testing a subject's response to variant A against variant B, and determining which of the two variants is more effective. [https://en.wikipedia.org/wiki/A/B_testing A/B testing | Wikipedia] | ||
+ | |||
+ | A <b>randomized controlled trial (or randomized control trial; RCT)</b> is a type of scientific (often medical) experiment that aims to reduce certain sources of bias when testing the effectiveness of new treatments; this is accomplished by randomly allocating subjects to two or more groups, treating them differently, and then comparing them with respect to a measured response. One group—the experimental group—receives the intervention being assessed, while the other—usually called the control group—receives an alternative treatment, such as a placebo or no intervention. The groups are monitored under conditions of the trial design to determine the effectiveness of the experimental intervention, and efficacy is assessed in comparison to the control. There may be more than one treatment group or more than one control group. [https://en.wikipedia.org/wiki/Randomized_controlled_trial Randomized controlled trial | Wikipedia] | ||
+ | |||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>AJX4W3MwKzU</youtube> | ||
+ | <b>Stanford Seminar: Peeking at A/B Tests - Why It Matters and What to Do About It | ||
+ | </b><br>Ramesh Johari Stanford University I'll describe a novel statistical methodology that has been deployed by the commercial A/B testing platform Optimizely to communicate experimental results to their customers. Our methodology addresses the issue that traditional [[Math for Intelligence#P-Value|p-values]] and [[Math for Intelligence#Confidence Interval|confidence intervals]] give unreliable inference. This is because users of A/B testing software are known to continuously monitor these measures as the experiment is running. We provide always valid [[Math for Intelligence#P-Value|p-values]] and [[Math for Intelligence#Confidence Interval|confidence intervals]] that are provably robust to this effect. Not only does this make it safe for a user to continuously monitor, but it empowers her to detect true effects more efficiently. I'll cover why continuous monitoring is a problem, and how our solution addresses it. The talk will be presented at a level that does not presume much statistical background. Ramesh Johari is an Associate Professor at Stanford University, with a full-time appointment in the Department of Management Science and Engineering (MS&E), and courtesy appointments in the Departments of Computer Science (CS) and Electrical Engineering (EE). He is a member of the Operations Research group and the Social Algorithms Lab (SOAL) in MS&E, the Information Systems Laboratory (ISL) in EE, and the Institute for Computational and Mathematical Engineering (ICME). Learn more about Stanford's Human-Computer Interaction Group: https://hci.stanford.edu | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>Iw40wdwkkLA</youtube> | ||
+ | <b>[[Meta|Facebook]]'s A B Platform Interactive Analysis in Realtime - @Scale 2014 - Data | ||
+ | </b><br>Itamar Rosenn, Engineering Manager at [[Meta|Facebook]] | ||
+ | This talk presents [[Meta|Facebook]]¹s platform for streamlined and automated A/B test analysis. We will discuss the system architecture, how it enables interactive analysis on near-realtime results, and the challenges we've faced in scaling the system to meet our customer needs. | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | |||
+ | == Scoring Deployed Models == | ||
+ | [https://www.youtube.com/results?search_query=~Scoring+~score+~rate+~Deploy+~production+~installed+~Model+~ai YouTube search...] | ||
+ | [https://www.google.com/search?q=Scoring+~score+~rate+~Deploy+~production+~installed+~Model+~ai ...Google search] | ||
+ | |||
+ | {|<!-- T --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>gB0bTH-L6DE</youtube> | ||
+ | <b>ML Model Deployment and Scoring on the Edge with Automatic ML & DF / Flink2Kafka | ||
+ | </b><br>recorded on June 18, 2020. Machine Learning Model Deployment and Scoring on the Edge with Automatic Machine Learning and Data Flow Deploying Machine Learning models to the edge can present significant ML/IoT challenges centered around the need for low latency and accurate scoring on minimal resource environments. H2O.ai's Driverless AI AutoML and Cloudera Data Flow work nicely together to solve this challenge. Driverless AI automates the building of accurate Machine Learning models, which are deployed as light footprint and low latency Java or C++ artifacts, also known as a MOJO (Model Optimized). And Cloudera Data Flow leverage Apache NiFi that offers an innovative data flow framework to host MOJOs to make predictions on data moving on the edge. Speakers: James Medel (H2O.ai - Technical Community Maker) Greg Keys (H2O.ai - Solution Engineer) Kafka 2 Flink - An Apache Love Story This project has heavily inspired by two existing efforts from Data In Motion's FLaNK Stack and Data Artisan's blog on stateful streaming applications. The goal of this project is to provide insight into connecting an Apache Flink applications to Apache Kafka. Speaker: Ian R Brooks, PhD (Cloudera - Senior Solutions Engineer & Data) | ||
+ | |} | ||
+ | |<!-- M --> | ||
+ | | valign="top" | | ||
+ | {| class="wikitable" style="width: 550px;" | ||
+ | || | ||
+ | <youtube>q-VPALG6ogY</youtube> | ||
+ | <b>Shawn Scully: Production and Beyond: Deploying and Managing Machine Learning Models | ||
+ | </b><br>PyData NYC 2015 Machine learning has become the key component in building intelligence-infused applications. However, as companies increase the number of such deployments, the number of machine learning models that need to be created, maintained, monitored, tracked, and improved grow at a tremendous pace. This growth has lead to a huge (and well-documented) accumulation of technical debt. Developing a machine learning application is an iterative process that involves building multiple models over a dataset. The dataset itself evolves over time as new features and new data points are collected. Furthermore, once deployed, the models require updates over time. Changes in models and datasets become difficult to track over time, and one can quickly lose track of which version of the model used which data and why it was subsequently replaced. In this talk, we outline some of the key challenges in large-scale deployments of many interacting machine learning models. We then describe a methodology for management, monitoring, and optimization of such models in production, which helps mitigate the technical debt. In particular, we demonstrate how to: Track models and versions, and visualize their quality over time Track the provenance of models and datasets, and quantify how changes in data impact the models being served Optimize model ensembles in real time, based on changing data, and provide alerts when such ensembles no longer provide the desired accuracy. | ||
+ | |} | ||
+ | |}<!-- B --> | ||
+ | |||
+ | <hr> | ||
+ | |||
+ | https://miro.medium.com/max/1000/1*ldWxdWDzEYnSbvchuL5k1w.png | ||
+ | |||
+ | <img src="https://images.contentful.com/pqts2v0qq7kz/4Mcjw0xAi4auqweOQQyWCu/80236c975b5026ec67e61f767a646b45/machine_learning_flow--4j88rajonr_s600x0_q80_noupscale.png" width="500" height="500"> | ||
+ | |||
+ | <img src="https://miro.medium.com/max/1327/1*9xjlXSJ9i2DBB-BJIrY26w.png" width="800" height="500"> |
Latest revision as of 15:54, 28 April 2024
YouTube ... Quora ...Google search ...Google News ...Bing News
- AI Solver ... Algorithms ... Administration ... Model Search ... Discriminative vs. Generative ... Train, Validate, and Test
- Development ... Notebooks ... AI Pair Programming ... Codeless ... Hugging Face ... AIOps/MLOps ... AIaaS/MLaaS
- Managed Vocabularies
- Analytics ... Visualization ... Graphical Tools ... Diagrams & Business Analysis ... Requirements ... Loop ... Bayes ... Network Pattern
- Backpropagation ... FFNN ... Forward-Forward ... Activation Functions ...Softmax ... Loss ... Boosting ... Gradient Descent ... Hyperparameter ... Manifold Hypothesis ... PCA
- Strategy & Tactics ... Project Management ... Best Practices ... Checklists ... Project Check-in ... Evaluation ... Measures
- Artificial General Intelligence (AGI) to Singularity ... Curious Reasoning ... Emergence ... Moonshots ... Explainable AI ... Automated Learning
- NLP Workbench / Pipeline
- Data Science ... Governance ... Preprocessing ... Exploration ... Interoperability ... Master Data Management (MDM) ... Bias and Variances ... Benchmarks ... Datasets
- Data Quality ...validity, accuracy, cleaning, completeness, consistency, encoding, padding, augmentation, labeling, auto-tagging, normalization, standardization, and imbalanced data
- Building Your Environment
- Service Capabilities
- AI Marketplace & Toolkit/Model Interoperability
- Directed Acyclic Graph (DAG) - programming pipelines
- Containers; Docker, Kubernetes & Microservices
- Automate your data lineage
- Benefiting from AI: A different approach to data management is needed
- Libraries & Frameworks Overview ... Libraries & Frameworks ... Git - GitHub and GitLab ... Other Coding options
- Use a Pipeline to Chain PCA with a RandomForest Classifier Jupyter Notebook | Jon Tupitza
- ML.NET Model Lifecycle with Azure DevOps CI/CD pipelines | Cesar de la Torre - Microsoft
- A Great Model is Not Enough: Deploying AI Without Technical Debt | DataKitchen - Medium
- Infrastructure Tools for Production | Aparna Dhinakaran - Towards Data Science ...Model Deployment and Serving
- Global Community for Artificial Intelligence (AI) in Master Data Management (MDM) | Camelot Management Consultants
- Particle Swarms for Dynamic Optimization Problems | T. Blackwell, J. Branke, and X. Li
- 5G_Security
Contents
Tools
- Google AutoML automatically build and deploy state-of-the-art machine learning models
- TensorBoard | Google
- Kubeflow Pipelines - a platform for building and deploying portable, scalable machine learning (ML) workflows based on Docker containers. Introducing AI Hub and Kubeflow Pipelines: Making AI simpler, faster, and more useful for businesses | Google
- SageMaker | Amazon
- MLOps | Microsoft ...model management, deployment, and monitoring with Azure
- Ludwig - a Python toolbox from Uber that allows to train and test deep learning models
- TPOT a Python library that automatically creates and optimizes full machine learning pipelines using genetic programming. Not for NLP, strings need to be coded to numerics.
- H2O Driverless AI for automated Visualization, feature engineering, model training, hyperparameter optimization, and explainability.
- alteryx: Feature Labs, Featuretools
- MLBox Fast reading and distributed data preprocessing/cleaning/formatting. Highly robust feature selection and leak detection. Accurate hyper-parameter optimization in high-dimensional space. State-of-the art predictive models for classification and regression (Deep Learning, Stacking, LightGBM,…). Prediction with models interpretation. Primarily Linux.
- auto-sklearn algorithm selection and hyperparameter tuning. It leverages recent advantages in Bayesian optimization, meta-learning and ensemble construction.is a Bayesian hyperparameter optimization layer on top of scikit-learn. Not for large datasets.
- Auto Keras is an open-source Python package for neural architecture search.
- ATM -auto tune models - a multi-tenant, multi-data system for automated machine learning (model selection and tuning). ATM is an open source software library under the Human Data Interaction project (HDI) at MIT.
- Auto-WEKA is a Bayesian hyperparameter optimization layer on top of Weka. Weka is a collection of machine learning algorithms for data mining tasks. The algorithms can either be applied directly to a dataset or called from your own Java code. Weka contains tools for data pre-processing, classification, regression, clustering, association rules, and visualization.
- TransmogrifAI - an AutoML library for building modular, reusable, strongly typed machine learning workflows. A Scala/SparkML library created by Salesforce for automated data cleansing, feature engineering, model selection, and hyperparameter optimization
- RECIPE - a framework based on grammar-based genetic programming that builds customized scikit-learn classification pipelines.
- AutoMLC Automated Multi-Label Classification. GA-Auto-MLC and Auto-MEKAGGP are freely-available methods that perform automated multi-label classification on the MEKA software.
- Databricks MLflow an open source framework to manage the complete Machine Learning lifecycle using Managed MLflow as an integrated service with the Databricks Unified Analytics Platform... ...manage the ML lifecycle, including experimentation, reproducibility and deployment
- SAS Viya automates the process of data cleansing, data transformations, feature engineering, algorithm matching, model training and ongoing governance.
- Comet ML ...self-hosted and cloud-based meta machine learning platform allowing data scientists and teams to track, compare, explain and optimize experiments and models
- Domino Model Monitor (DMM) | Domino ...monitor the performance of all models across your entire organization
- Weights and Biases ...experiment tracking, model optimization, and dataset versioning
- SigOpt ...optimization platform and API designed to unlock the potential of modeling pipelines. This fully agnostic software solution accelerates, amplifies, and scales the model development process
- DVC ...Open-source Version Control System for Machine Learning Projects
- ModelOp Center | ModelOp
- Moogsoft and Red Hat Ansible Tower
- DSS | Dataiku
- Model Manager | SAS
- Machine Learning Operations (MLOps) | DataRobot ...build highly accurate predictive models with full transparency
- Metaflow, Netflix and AWS open source Python library
Master Data Management (MDM)
YouTube search... ...Google search Feature Store / Data Lineage / Data Catalog
- Data Science ... Governance ... Preprocessing ... Exploration ... Interoperability ... Master Data Management (MDM) ... Bias and Variances ... Benchmarks ... Datasets
|
|
|
|
|
|
Versioning
YouTube search... ...Google search
- DVC | DVC.org
- Pachyderm …Pachyderm for data scientists | Gerben Oostra - bigdata - Medium
- Dataiku
- Continuous Machine Learning (CML)
|
|
|
|
|
|
Model Versioning - ModelDB
- ModelDB: An open-source system for Machine Learning model versioning, metadata, and experiment management
|
|
Hyperparameter
YouTube search... ...Google search
- Backpropagation ... FFNN ... Forward-Forward ... Activation Functions ...Softmax ... Loss ... Boosting ... Gradient Descent ... Hyperparameter ... Manifold Hypothesis ... PCA
- Hypernetworks
- Using TensorFlow Tuning
- Understanding Hyperparameters and its Optimisation techniques | Prabhu - Towards Data Science
- How To Make Deep Learning Models That Don’t Suck | Ajay Uppili Arasanipalai
- Optimization Methods
In machine learning, a hyperparameter is a parameter whose value is set before the learning process begins. By contrast, the values of other parameters are derived via training. Different model training algorithms require different hyperparameters, some simple algorithms (such as ordinary least squares regression) require none. Given these hyperparameters, the training algorithm learns the parameters from the data. Hyperparameter (machine learning) | Wikipedia
Machine learning algorithms train on data to find the best set of weights for each independent variable that affects the predicted value or class. The algorithms themselves have variables, called hyperparameters. They’re called hyperparameters, as opposed to parameters, because they control the operation of the algorithm rather than the weights being determined. The most important hyperparameter is often the learning rate, which determines the step size used when finding the next set of weights to try when optimizing. If the learning rate is too high, the gradient descent may quickly converge on a plateau or suboptimal point. If the learning rate is too low, the gradient descent may stall and never completely converge. Many other common hyperparameters depend on the algorithms used. Most algorithms have stopping parameters, such as the maximum number of epochs, or the maximum time to run, or the minimum improvement from epoch to epoch. Specific algorithms have hyperparameters that control the shape of their search. For example, a Random Forest (or) Random Decision Forest Classifier has hyperparameters for minimum samples per leaf, max depth, minimum samples at a split, minimum weight fraction for a leaf, and about 8 more. Machine learning algorithms explained | Martin Heller - InfoWorld
Hyperparameter Tuning
Hyperparameters are the variables that govern the training process. Your model parameters are optimized (you could say "tuned") by the training process: you run data through the operations of the model, compare the resulting prediction with the actual value for each data instance, evaluate the accuracy, and adjust until you find the best combination to handle the problem. These algorithms automatically adjust (learn) their internal parameters based on data. However, there is a subset of parameters that is not learned and that have to be configured by an expert. Such parameters are often referred to as “hyperparameters” — and they have a big impact ...For example, the tree depth in a decision tree model and the number of layers in an artificial Neural Network are typical hyperparameters. The performance of a model can drastically depend on the choice of its hyperparameters. Machine learning algorithms and the art of hyperparameter selection - A review of four optimization strategies | Mischa Lisovyi and Rosaria Silipo - TNW
There are four commonly used optimization strategies for hyperparameters:
- Bayesian optimization
- Grid search
- Random search
- Hill climbing
Bayesian optimization tends to be the most efficient. You would think that tuning as many hyperparameters as possible would give you the best answer. However, unless you are running on your own personal hardware, that could be very expensive. There are diminishing returns, in any case. With experience, you’ll discover which hyperparameters matter the most for your data and choice of algorithms. Machine learning algorithms explained | Martin Heller - InfoWorld
Hyperparameter Optimization libraries:
- hyper-engine - Gaussian Process Bayesian optimization and some other techniques, like learning curve prediction
- Ray Tune: Hyperparameter Optimization Framework
- SigOpt’s API tunes your model’s parameters through state-of-the-art Bayesian optimization
- hyperopt; Distributed Asynchronous Hyperparameter Optimization in Python - random search and tree of parzen estimators optimization.
- Scikit-Optimize, or skopt - Gaussian process Bayesian optimization
- polyaxon
- GPyOpt; Gaussian Process Optimization
Tuning:
- Optimizer type
- Learning rate (fixed or not)
- Epochs
- Regularization rate (or not)
- Type of Regularization - L1, L2, ElasticNet
- Search type for local minima
- Gradient descent
- Simulated
- Annealing
- Evolutionary
- Decay rate (or not)
- Momentum (fixed or not)
- Nesterov Accelerated Gradient momentum (or not)
- Batch size
- Fitness measurement type
- MSE, accuracy, MAE, Cross-Entropy Loss
- Precision, recall
- Stop criteria
Automated Learning
YouTube search... ...Google search
- Artificial General Intelligence (AGI) to Singularity ... Curious Reasoning ... Emergence ... Moonshots ... Explainable AI ... Automated Learning
- Immersive Reality ... Metaverse ... Omniverse ... Transhumanism ... Religion
- Telecommunications ... Computer Networks ... 5G ... Satellite Communications ... Quantum Communications ... Communication Agents ... Smart Cities ... Digital Twin ... Internet of Things (IoT)
- Large Language Model (LLM) ... Multimodal ... Foundation Models (FM) ... Generative Pre-trained ... Transformer ... GPT-4 ... GPT-5 ... Attention ... GAN ... BERT
- Development ... Notebooks ... AI Pair Programming ... Codeless, Generators, Drag n' Drop ... AIOps/MLOps ... AIaaS/MLaaS
- AdaNet
- Creatives ... History of Artificial Intelligence (AI) ... Neural Network History ... Rewriting Past, Shape our Future ... Archaeology ... Paleontology
- Generative Pre-trained Transformer 5 (GPT-5)
- AI Software Learns to Make AI Software
- The Pentagon Wants AI to Take Over the Scientific Process | Automating Scientific Knowledge Extraction (ASKE) | DARPA
- Hallucinogenic Deep Reinforcement Learning Using Python and Keras | David Foster
- Automated Feature Engineering in Python - How to automatically create machine learning features | Will Koehrsen - Towards Data Science
- Why Meta-learning is Crucial for Further Advances of Artificial Intelligence? | Pavel Kordik
- Assured Autonomy | Dr. Sandeep Neema, DARPA
- Automatic Machine Learning is Broken | Piotr Plonski - KDnuggets
- Why 2020 will be the Year of Automated Machine Learning | Senthil Ravindran - Gigabit
- Meta Learning | Wikipedia
Several production machine-learning platforms now offer automatic hyperparameter tuning. Essentially, you tell the system what hyperparameters you want to vary, and possibly what metric you want to optimize, and the system sweeps those hyperparameters across as many runs as you allow. (Google Cloud hyperparameter tuning extracts the appropriate metric from the TensorFlow model, so you don’t have to specify it.)
An emerging class of data science toolkit that is finally making machine learning accessible to business subject matter experts. We anticipate that these innovations will mark a new era in data-driven decision support, where business analysts will be able to access and deploy machine learning on their own to analyze hundreds and thousands of dimensions simultaneously. Business analysts at highly competitive organizations will shift from using visualization tools as their only means of analysis, to using them in concert with AML. Data visualization tools will also be used more frequently to communicate model results, and to build task-oriented user interfaces that enable stakeholders to make both operational and strategic decisions based on output of scoring engines. They will also continue to be a more effective means for analysts to perform inverse analysis when one is seeking to identify where relationships in the data do not exist. 'Five Essential Capabilities: Automated Machine Learning' | Gregory Bonnette
H2O Driverless AI automatically performs feature engineering and hyperparameter tuning, and claims to perform as well as Kaggle masters. AmazonML SageMaker supports hyperparameter optimization. Microsoft Azure Machine Learning AutoML automatically sweeps through features, algorithms, and hyperparameters for basic machine learning algorithms; a separate Azure Machine Learning hyperparameter tuning facility allows you to sweep specific hyperparameters for an existing experiment. Google Cloud AutoML implements automatic deep transfer learning (meaning that it starts from an existing Deep Neural Network (DNN) trained on other data) and neural architecture search (meaning that it finds the right combination of extra network layers) for language pair translation, natural language classification, and image classification. Review: Google Cloud AutoML is truly automated machine learning | Martin Heller
|
|
AutoML
YouTube search... ...Google search
- Automated Machine Learning (AutoML) | Wikipedia
- AutoML.org ...ML Freiburg ... GitHub and ML Hannover
New cloud software suite of machine learning tools. It’s based on Google’s state-of-the-art research in image recognition called Neural Architecture Search (NAS). NAS is basically an algorithm that, given your specific dataset, searches for the most optimal neural network to perform a certain task on that dataset. AutoML is then a suite of machine learning tools that will allow one to easily train high-performance deep networks, without requiring the user to have any knowledge of deep learning or AI; all you need is labelled data! Google will use NAS to then find the best network for your specific dataset and task. AutoKeras: The Killer of Google’s AutoML | George Seif - KDnuggets
Automatic Machine Learning (AML)
Self-Learning
DARTS: Differentiable Architecture Search
YouTube search... ...Google search
- DARTS: Differentiable Architecture Search | H. Liu, K. Simonyan, and Y. Yang addresses the scalability challenge of architecture search by formulating the task in a differentiable manner. Unlike conventional approaches of applying evolution or reinforcement learning over a discrete and non-differentiable search space, the method is based on the continuous relaxation of the architecture representation, allowing efficient search of the architecture using gradient descent.
- Neural Architecture Search | Debadeepta Dey - Microsoft Research
AIOps/MLOps
Youtube search... ...Google search
- Development ... Notebooks ... AI Pair Programming ... Codeless ... Hugging Face ... AIOps/MLOps ... AIaaS/MLaaS
- MLflow
- ChatGPT integration with DevSecOps
- DevOps.com
- A Silver Bullet For CIOs; Three ways AIOps can help IT leaders get strategic - Lisa Wolfe - Forbes
- MLOps: What You Need To Know | Tom Taulli - Forbes
- What is so Special About AIOps for Mission Critical Workloads? | Rebecca James - DevOps
- What is AIOps? Artificial Intelligence for IT Operations Explained | BMC
- AIOps: Artificial Intelligence for IT Operations, Modernize and transform IT Operations with solutions built on the only Data-to-Everything platform | splunk>
- How to Get Started With AIOps | Susan Moore - Gartner
- Why AI & ML Will Shake Software Testing up in 2019 | Oleksii Kharkovyna - Medium
- Continuous Delivery for Machine Learning D. Sato, A. Wider and C. Windheuser - MartinFowler
- Defense: Adaptive Acquisition Framework (AAF)
Machine learning capabilities give IT operations teams contextual, actionable insights to make better decisions on the job. More importantly, AIOps is an approach that transforms how systems are automated, detecting important signals from vast amounts of data and relieving the operator from the headaches of managing according to tired, outdated runbooks or policies. In the AIOps future, the environment is continually improving. The administrator can get out of the impossible business of refactoring rules and policies that are immediately outdated in today’s modern IT environment. Now that we have AI and machine learning technologies embedded into IT operations systems, the game changes drastically. AI and machine learning-enhanced automation will bridge the gap between DevOps and IT Ops teams: helping the latter solve issues faster and more accurately to keep pace with business goals and user needs. How AIOps Helps IT Operators on the Job | Ciaran Byrne - Toolbox
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Large Language Model Operations (LLMOps)
Large Language Model Operations (LLMOps) is a set of practices and tools for the operational management of large language models (LLMs) in production environments. LLMOps encompasses the entire lifecycle of LLMs, from development and deployment to monitoring and maintenance. LLMOps is important because LLMs are complex and resource-intensive models that require careful management to ensure that they are performing as expected and meeting the needs of their users. LLMOps tools and practices can help to streamline and automate the LLM lifecycle, making it easier to deploy and manage LLMs in production.
Here are some of the key components of LLMOps:
- Model development: LLMOps tools and practices can help to streamline the process of developing new LLMs, from data collection and preparation to model training and evaluation.
- Model deployment: LLMOps tools and practices can help to deploy LLMs to production environments in a safe and efficient manner.
- Model monitoring: LLMOps tools and practices can help to monitor the performance of LLMs in production and identify any problems early on.
- Model maintenance: LLMOps tools and practices can help to maintain LLMs in production, including fine-tuning models to improve their performance and updating models with new data.
LLMOps is still a relatively new field, but it is rapidly growing as more and more organizations adopt LLMs. There are a number of LLMOps tools and platforms available, both open source and proprietary. Here are some of the benefits of using LLMOps:
- Improved efficiency and productivity: LLMOps tools and practices can help to automate and streamline the LLM lifecycle, freeing up data scientists and engineers to focus on more creative and strategic tasks.
- Reduced costs: LLMOps can help to reduce the costs of developing and managing LLMs by improving efficiency and reducing the risk of errors.
- Increased reliability and performance: LLMOps can help to ensure that LLMs are performing reliably and meeting the needs of their users by providing tools and practices for monitoring and maintaining LLMs in production.
How is LLMOps different from AIOps/MLOps?
- AIOps vs. MLOps vs. LLMOps | Mostafa Ibrahim - Weights & Biases
- LLMOPS vs MLOPS: Choosing the Best Path for AI Development | Analytics Vidhya
- AIOps vs MLOps | Know Everything in Detail | Jagreet Kaur - Xenonstack
LLMOps is a subset of MLOps that focuses specifically on the operational management of large language models (LLMs). MLOps is a set of practices and tools for managing the entire lifecycle of machine learning models, from development to deployment to production. AIOps is a broader field that focuses on using AI and machine learning to improve IT operations and management. Here is a table that summarizes the key differences between LLMOps, AIOps, and MLOps:
Characteristic | LLMOps | AIOps | MLOps |
---|---|---|---|
Focus | Operational management of LLMs | Use of AI and ML to improve IT operations and management | Operational management of machine learning models |
Use cases | Language-related tasks, such as chatbots, language translation, sentiment analysis, and content generation | IT operations tasks, such as incident management, capacity planning, and performance monitoring | Machine learning tasks, such as fraud detection, product recommendation, and predictive maintenance |
Team composition | NLP specialists, linguists, data scientists with language modeling expertise, and software engineers experienced in building language-related applications | Data scientists, analytics professionals, data engineers, domain experts, and IT professionals | Data scientists, machine learning engineers, software engineers, and DevOps engineers |
Here are some examples of how LLMOps is different from AIOps/MLOps:
- Model development: LLMOps tools and practices are specifically designed for developing and training LLMs. For example, LLMOps tools may help to automate the process of collecting and cleaning text data, or they may provide specialized libraries for training LLMs.
- Model deployment: LLMOps tools and practices can help to deploy LLMs to production environments in a way that is optimized for performance and reliability. For example, LLMOps tools may help to scale LLMs to handle large volumes of traffic, or they may provide specialized infrastructure for hosting LLMs.
- Model monitoring and maintenance: LLMOps tools and practices can help to monitor the performance of LLMs in production and identify any problems early on. For example, LLMOps tools may track the accuracy of LLMs on real-world data, or they may detect potential biases in LLMs.
Continuous Machine Learning (CML)
- Continuous Machine Learning (CML) ...is Continuous Integration/Continuous Deployment (CI/CD) for Machine Learning Projects
- DVC | DVC.org
|
|
DevSecOps
Youtube search... ...Google search
- Cybersecurity
- Containers; Docker, Kubernetes & Microservices
- SafeCode ...nonprofit organization that brings business leaders and technical experts together to exchange insights and ideas on creating, improving and promoting scalable and effective software security programs.
- 3 ways AI will advance DevSecOps | Joseph Feiman - TechBeacon
- Leveraging AI and Automation for Successful DevSecOps | Vishnu Nallani - DZone
DevSecOps (also known as SecDevOps and DevOpsSec) is the process of integrating secure development best practices and methodologies into continuous design, development, deployment and integration processes
|
|
|
|
|
|
DevSecOps in Government
- Evaluation: DevSecOps Guide | General Services Administration (GSA)
- Cybersecurity & Acquisition Lifecycle Integration
- Defense|DOD Enterprise DevSecOps Reference Design Version 1.0 12 August 2019 | Department of Defense (DOD) Chief Information Officer (CIO
- Understanding the Differences Between Agile & DevSecOps - from a Business Perspective | General Services Administration (GSA)
|
|
|
|
Strangler Fig / Strangler Pattern
- Strangler Fig Application | Martin Fowler
- The Strangler pattern in practice | Michiel Rook
- Strangler pattern ...Cloud Design Patterns | Microsoft
- How to use strangler pattern for microservices modernization | N. Natean - Software Intelligence Plus
- Serverless Strangler Pattern on AWS | Ryan Means - Medium ... Serverless
Strangulation of a legacy or undesirable solution is a safe way to phase one thing out for something better, cheaper, or more expandable. You make something new that obsoletes a small percentage of something old, and put them live together. You do some more work in the same style, and go live again (rinse, repeat). Strangler Applications | Paul Hammant ...case studies
|
|
Model Monitoring
YouTube search... ...Google search
- How do you evaluate the performance of a machine learning model that's deployed into production? | Quora
- Why your Models need Maintenance | Martin Schmitz - Towards Data Science ...Change of concept & drift of Concept
- Deployed your Machine Learning Model? Here’s What you Need to Know About Post-Production Monitoring | Om Deshmukh - Analytics Vidhya ...proactive & reactive model monitoring
Monitoring production systems is essential to keeping them running well. For ML systems, monitoring becomes even more important, because their performance depends not just on factors that we have some control over, like infrastructure and our own software, but also on data, which we have much less control over. Therefore, in addition to monitoring standard metrics like latency, traffic, errors and saturation, we also need to monitor model prediction performance. An obvious challenge with monitoring model performance is that we usually don’t have a verified label to compare our model’s predictions to, since the model works on new data. In some cases we might have some indirect way of assessing the model’s effectiveness, for example by measuring click rate for a recommendation model. In other cases, we might have to rely on comparisons between time periods, for example by calculating a percentage of positive classifications hourly and alerting if it deviates by more than a few percent from the average for that time. Just like when validating the model, it’s also important to monitor metrics across slices, and not just globally, to be able to detect problems affecting specific segments. ML Ops: Machine Learning as an Engineering Discipline | Cristiano Breuel - Towards Data Science
|
|
|
|
A/B Testing
YouTube search... ...Google search
A/B testing (also known as bucket testing or split-run testing) is a user experience research methodology. A/B tests consist of a randomized experiment with two variants, A and B. It includes application of statistical hypothesis testing or "two-sample hypothesis testing" as used in the field of statistics. A/B testing is a way to compare two versions of a single variable, typically by testing a subject's response to variant A against variant B, and determining which of the two variants is more effective. A/B testing | Wikipedia
A randomized controlled trial (or randomized control trial; RCT) is a type of scientific (often medical) experiment that aims to reduce certain sources of bias when testing the effectiveness of new treatments; this is accomplished by randomly allocating subjects to two or more groups, treating them differently, and then comparing them with respect to a measured response. One group—the experimental group—receives the intervention being assessed, while the other—usually called the control group—receives an alternative treatment, such as a placebo or no intervention. The groups are monitored under conditions of the trial design to determine the effectiveness of the experimental intervention, and efficacy is assessed in comparison to the control. There may be more than one treatment group or more than one control group. Randomized controlled trial | Wikipedia
|
|
Scoring Deployed Models
YouTube search... ...Google search
|
|