Difference between revisions of "Decentralized: Federated & Distributed"
m |
m |
||
| Line 79: | Line 79: | ||
<youtube>jkkmBpJ-Eeo</youtube> | <youtube>jkkmBpJ-Eeo</youtube> | ||
<b>Distinguished Lecturer : Eric Xing - Strategies & Principles for Distributed Machine Learning | <b>Distinguished Lecturer : Eric Xing - Strategies & Principles for Distributed Machine Learning | ||
| − | </b><br>The rise of Big Data has led to new demands for Machine Learning (ML) systems to learn complex models with millions to billions of parameters that promise adequate capacity to digest massive datasets and offer powerful predictive analytics (such as high-dimensional latent features, intermediate representations, and decision functions) thereupon. In order to run ML algorithms at such scales, on a distributed cluster with 10s to 1000s of machines, it is often the case that significant engineering efforts are required --- and one might fairly ask if such engineering truly falls within the domain of ML research or not. Taking the view that Big ML systems can indeed benefit greatly from ML-rooted statistical and algorithmic insights --- and that ML researchers should therefore not shy away from such systems design --- we discuss a series of principles and strategies distilled from our resent effort on industrial-scale ML solutions that involve a continuum from application, to engineering, and to theoretical research and development of Big ML system and architecture, on how to make them efficient, general, and with convergence and scaling guarantees. These principles concern four key questions which traditionally receive little attention in ML research: How to distribute an ML program over a cluster? How to bridge ML computation with inter-machine [[Agents#Communication | communication]]? How to perform such [[Agents#Communication | communication]]? What should be communicated between machines? By exposing underlying statistical and algorithmic characteristics unique to ML programs but not typical in traditional computer programs, and by dissecting successful cases of how we harness these principles to design both high-performance distributed ML software and general-purpose ML framework, we present opportunities for ML researchers and practitioners to further shape and grow the area that lies between ML and systems. This is joint work with the CMU Petuum Team. | + | </b><br>The rise of Big Data has led to new demands for Machine Learning (ML) systems to learn complex models with millions to billions of parameters that promise adequate capacity to digest massive datasets and offer powerful predictive analytics (such as high-dimensional latent features, intermediate representations, and decision functions) thereupon. In order to run ML algorithms at such scales, on a distributed cluster with 10s to 1000s of machines, it is often the case that significant engineering efforts are required --- and one might fairly ask if such engineering truly falls within the domain of ML research or not. Taking the view that Big ML systems can indeed benefit greatly from ML-rooted statistical and algorithmic insights --- and that ML researchers should therefore not shy away from such systems design --- we discuss a series of principles and strategies distilled from our resent effort on industrial-scale ML solutions that involve a continuum from application, to engineering, and to theoretical research and [[development]] of Big ML system and architecture, on how to make them efficient, general, and with convergence and scaling guarantees. These principles concern four key questions which traditionally receive little attention in ML research: How to distribute an ML program over a cluster? How to bridge ML computation with inter-machine [[Agents#Communication | communication]]? How to perform such [[Agents#Communication | communication]]? What should be communicated between machines? By exposing underlying statistical and algorithmic characteristics unique to ML programs but not typical in traditional computer programs, and by dissecting successful cases of how we harness these principles to design both high-performance distributed ML software and general-purpose ML framework, we present opportunities for ML researchers and practitioners to further shape and grow the area that lies between ML and systems. This is joint work with the CMU Petuum Team. |
|} | |} | ||
|<!-- M --> | |<!-- M --> | ||
| Line 209: | Line 209: | ||
* [https://dodcio.defense.gov/Portals/0/Documents/DoD-C3-Strategy.pdf Command, Control, and Communications (C3) Modernization Strategy | Department of] [[Defense]] | * [https://dodcio.defense.gov/Portals/0/Documents/DoD-C3-Strategy.pdf Command, Control, and Communications (C3) Modernization Strategy | Department of] [[Defense]] | ||
| − | [[Microsoft]]’s SharePoint, Exchange, and Office 365 products run on Azure and Azure Stack, as do [[Microsoft]]’s database, e-commerce, and software development products. Extend Azure services and capabilities to your environment of choice—from the datacenter to edge locations and remote offices—with Azure Stack. Build, deploy, and run hybrid and edge computing apps consistently across your IT ecosystem, with flexibility for diverse workloads. The Azure Stack Hub architecture lets you provide Azure services at the edge for remote locations or intermittent connectivity, disconnected from the internet. You can create hybrid solutions that process data locally in Azure Stack Hub and then aggregate it in Azure for additional processing and analytics [https://www.nextplatform.com/2017/09/22/azure-stack-finally-takes-microsoft-public-cloud-private/#:~:text=Microsoft's%20SharePoint%2C%20Exchange%2C%20and%20Office,commerce%2C%20and%20software%20development%20products. Azure Stack Finally Takes Microsoft Public Cloud Private | Paul Teich - The Next Platform] | + | [[Microsoft]]’s SharePoint, Exchange, and Office 365 products run on Azure and Azure Stack, as do [[Microsoft]]’s database, e-commerce, and software [[development]] products. Extend Azure services and capabilities to your environment of choice—from the datacenter to edge locations and remote offices—with Azure Stack. Build, deploy, and run hybrid and edge computing apps consistently across your IT ecosystem, with flexibility for diverse workloads. The Azure Stack Hub architecture lets you provide Azure services at the edge for remote locations or intermittent connectivity, disconnected from the internet. You can create hybrid solutions that process data locally in Azure Stack Hub and then aggregate it in Azure for additional processing and analytics [https://www.nextplatform.com/2017/09/22/azure-stack-finally-takes-microsoft-public-cloud-private/#:~:text=Microsoft's%20SharePoint%2C%20Exchange%2C%20and%20Office,commerce%2C%20and%20software%20development%20products. Azure Stack Finally Takes Microsoft Public Cloud Private | Paul Teich - The Next Platform] |
<blockquote style="border: 2px solid #666; padding: 10px; background-color: #ccc;"> Sometimes, this kind of environment is also referred to as a 'submarine' scenario. - [[Microsoft]]</blockquote> | <blockquote style="border: 2px solid #666; padding: 10px; background-color: #ccc;"> Sometimes, this kind of environment is also referred to as a 'submarine' scenario. - [[Microsoft]]</blockquote> | ||
Revision as of 14:43, 17 March 2023
Youtube search... ...Google search
- Case Studies
- Blockchain
- Privacy
- OpenMined
- Platforms: AI/Machine Learning as a Service (AIaaS/MLaaS)
- Computer Networks
Contents
Centralized vs. Decentralized vs. Distributed
Youtube search... ...Google search
- Social Media Alternatives Series, EP. 1: What You NEED to Know | Fliphodl
- Exclusive: We tried Ultra and it will kill Steam | Fliphodl
Federated
Youtube search... ...Google search
- Federated Learning | Wikipedia
- Federated Machine Learning: Concept and Applications | Q. Yang, Y. Liu, T. Chen and Y. Tong
- Federated Learning: The Future of Distributed Machine Learning | Mi Zhang - Medium
- Chained Anomaly Detection Models for Federated Learning: An Intrusion Detection Case Study | D. Preuveneers, V. Rimmer, I. Tsingenopoulos, J. Spooren, W. Joosen and E. Ilie-Zudor
- A Beginners Guide to Federated Learning | Santanu Bhattacharya
- Watch me Build a Cybersecurity Startup
- Federated Learning Frameworks:
- OpenMined Pysyft ...GitHub
- TensorFlow Federated Learning ...* TensorFlow Federated helps train AI models on data from different locations | Kyle Wiggers - VentureBeat
- Federated AI Ecosystem (FATE) | FEDAI.org
- Clara Federated Learning | NVIDIA
Distributed
Youtube search... ...Google search
- Distributed Deep Reinforcement Learning (DDRL)
- Distributed hyperparameter tuning for machine learning models | Microsoft
Distributed machine learning refers to multi-node machine learning algorithms and systems that are designed to improve performance, increase accuracy, and scale to larger input data sizes. Increasing the input data size for many algorithms can significantly reduce the learning error and can often be more effective than using more complex methods [8]. Distributed machine learning allows companies, researchers, and individuals to make informed decisions and draw meaningful conclusions from large amounts of data. Many systems exist for performing machine learning tasks in a distributed environment. These systems fall into three primary categories: database, general, and purpose-built systems. Each type of system has distinct advantages and disadvantages, but all are used in practice depending upon individual use cases, performance requirements, input data sizes, and the amount of implementation effort. | SpringerLink
|
|
|
|
Peer-to-Peer
Youtube search... ...Google search
- Hopfield Network (HN)
- Feedback Loop - Peer Learning
- Developing Machine Intelligence within P2P Networks Using a Distributed Associative Memory | A. Amir, A. Amin, and A. Khan - SpingerLink ...Distributed Associative Memory Tree (DASMET), to deal with multi-feature recognition in a peer-to-peer (P2P)-based system.
- BrainTorrent: A Peer-to-Peer Environment for Decentralized Federated Learning | A. G. Roy, S. Siddiqui, S. Pölsterl, N. Navab, and C. Wachinger
- Role of Peer to Peer Networks In Creating Transparency And Increased Usage of AI | Imarticus
- A Reinforcement Learning based Distributed Search Algorithm For Hierarchical Peer-to-Peer Information Retrieval Systems | Haizheng Zhang and Victor Lesser
|
|
Proxy
Youtube search... ...Google search
|
|
Submarine Scenario
Youtube search... ...Google search
- Azure Stack
- Azure hybrid options | Microsoft
- Azure Stack Hub
- Datacenter integration planning considerations for Azure Stack Hub integrated systems | Microsoft
- Azure disconnected deployment planning decisions for Azure Stack Hub integrated systems Microsoft ... Features that are impaired or unavailable in disconnected deployments
- Hybrid relay connection in Azure and Azure Stack Hub | Microsoft
- AI at the edge with Azure Stack Hub | Microsoft
- Stretched clusters overview | Microsoft
- Hybrid availability and performance monitoring | Microsoft
- Azure Automation in a hybrid environment | Microsoft
- Use Azure file shares in a hybrid environment | Microsoft
- Command, Control, and Communications (C3) Modernization Strategy | Department of Defense
Microsoft’s SharePoint, Exchange, and Office 365 products run on Azure and Azure Stack, as do Microsoft’s database, e-commerce, and software development products. Extend Azure services and capabilities to your environment of choice—from the datacenter to edge locations and remote offices—with Azure Stack. Build, deploy, and run hybrid and edge computing apps consistently across your IT ecosystem, with flexibility for diverse workloads. The Azure Stack Hub architecture lets you provide Azure services at the edge for remote locations or intermittent connectivity, disconnected from the internet. You can create hybrid solutions that process data locally in Azure Stack Hub and then aggregate it in Azure for additional processing and analytics Azure Stack Finally Takes Microsoft Public Cloud Private | Paul Teich - The Next Platform
Sometimes, this kind of environment is also referred to as a 'submarine' scenario. - Microsoft
In 2021, the DoD CIO designated the Department of Navy CIO as the executive agent to lead a cross-service joint working group focused on Denied-Disconnected, Intermittent, and Low bandwidth (D-DIL)... Network server software and hardware exist at the tactical edge to provide critical IT services and data in these DDIL environments, along with a variety of spectrum communications and unclassified & classified network transports leveraging satellite links and low-Earth Orbit (LEO), Wi-Fi, cellular/4G LTE, millimeter wave/5G and others. The working group has teamed up with industry to refine DoD-unique requirements and use cases, resulting in the development of standardized architectures and solutions for the relevant collaboration and productivity tools (email, chat, voice and video, file management). These tools operate as a hybrid capability, which will allow users access to the full feature set when cloud connectivity is available, but remain productive locally within the DDIL environment. DoD working with Industry to Adapt Cloud Tools for the Tactical Edge - DON CIO
Additionally, as DoD enterprise IT moves to the cloud, tactical networks must unify access to data and applications from the enterprise level to the tactical edge. This means deploying cloudlike services at the tactical edge of the network, so that data is available at the edge even when WAN connectivity is unavailable.Four future trends In tactical network modernization | US Army
|
|
|
|