Decentralized: Federated & Distributed

From
Jump to: navigation, search

YouTube ... Quora ...Google search ...Google News ...Bing News


Centralized vs. Decentralized vs. Distributed

Youtube search... ...Google search

Centralised-decentralised-distributed.png


Federated

Youtube search... ...Google search




Distributed

Youtube search... ...Google search

Distributed machine learning refers to multi-node machine learning algorithms and systems that are designed to improve performance, increase accuracy, and scale to larger input data sizes. Increasing the input data size for many algorithms can significantly reduce the learning error and can often be more effective than using more complex methods [8]. Distributed machine learning allows companies, researchers, and individuals to make informed decisions and draw meaningful conclusions from large amounts of data. Many systems exist for performing machine learning tasks in a distributed environment. These systems fall into three primary categories: database, general, and purpose-built systems. Each type of system has distinct advantages and disadvantages, but all are used in practice depending upon individual use cases, performance requirements, input data sizes, and the amount of implementation effort. | SpringerLink

Distinguished Lecturer : Eric Xing - Strategies & Principles for Distributed Machine Learning
The rise of Big Data has led to new demands for Machine Learning (ML) systems to learn complex models with millions to billions of parameters that promise adequate capacity to digest massive datasets and offer powerful Predictive Analytics (such as high-dimensional latent features, intermediate representations, and decision functions) thereupon. In order to run ML algorithms at such scales, on a distributed cluster with 10s to 1000s of machines, it is often the case that significant engineering efforts are required --- and one might fairly ask if such engineering truly falls within the domain of ML research or not. Taking the view that Big ML systems can indeed benefit greatly from ML-rooted statistical and algorithmic insights --- and that ML researchers should therefore not shy away from such systems design --- we discuss a series of principles and strategies distilled from our resent effort on industrial-scale ML solutions that involve a continuum from application, to engineering, and to theoretical research and development of Big ML system and architecture, on how to make them efficient, general, and with convergence and scaling guarantees. These principles concern four key questions which traditionally receive little attention in ML research: How to distribute an ML program over a cluster? How to bridge ML computation with inter-machine communication? How to perform such communication? What should be communicated between machines? By exposing underlying statistical and algorithmic characteristics unique to ML programs but not typical in traditional computer programs, and by dissecting successful cases of how we harness these principles to design both high-performance distributed ML software and general-purpose ML framework, we present opportunities for ML researchers and practitioners to further shape and grow the area that lies between ML and systems. This is joint work with the CMU Petuum Team.

Distributed TensorFlow training (Google I/O '18)
To efficiently train machine learning models, you will often need to scale your training to multiple GPUs, or even multiple machines. TensorFlow now offers rich functionality to achieve this with just a few lines of code. Join this session to learn how to set this up.

Distribution Strategy API: https://goo.gl/F9vXqQ https://goo.gl/Zq2xvJ

ResNet50 Model Garden example with MirroredStrategy API: https://goo.gl/3UWhj8

Performance Guides: https://goo.gl/doqGE7 https://goo.gl/NCnrCn

Commands to set up a GCE instance and run distributed training: https://goo.gl/xzwN4C

Multi-machine distributed training with train_and_evaluate: https://goo.gl/kyikAC

Watch more TensorFlow sessions from I/O '18 here → https://goo.gl/GaAnBR See all the sessions from Google I/O '18 here → https://goo.gl/q1Tr8x

Subscribe to the TensorFlow channel → https://goo.gl/ht3WGe

Demystifying Parallel and Distributed Deep Learning: An In-Depth Concurrency Analysis
In this video from 2018 Swiss HPC Conference, Torsten Hoefler from (ETH) Zürich presents: Demystifying Parallel and Distributed Deep Learning: An In-Depth Concurrency Analysis. "Deep Neural Networks (DNN) are becoming an important tool in modern computing applications. Accelerating their training is a major challenge and techniques range from distributed algorithms to low-level circuit design. In this talk, we describe the problem from a theoretical perspective, followed by approaches for its parallelization. Specifically, we present trends in DNN architectures and the resulting implications on parallelization strategies. We discuss the different types of concurrency in DNNs; synchronous and asynchronous stochastic gradient descent; distributed system architectures; communication schemes; and performance modeling. Based on these approaches, we extrapolate potential directions for parallelism in deep learning." Learn more: https://hpcadvisorycouncil.com Sign up for our insideHPC Newsletter: https://insidehpc.com/newsletter

Machine Learning Systems for Highly Distributed and Rapidly Growing Data
Microsoft Research The usability and practicality of machine learning are largely influenced by two critical factors: low latency and low cost. However, achieving low latency and low cost is very challenging when machine learning depends on real-world data that are rapidly growing and highly distributed (e.g., training a face recognition model using pictures stored across many data centers globally). In this talk, I will present my work on building low-latency and low-cost machine learning systems that enable efficient processing of real-world, large-scale data. I will describe a system-level approach that is inspired by the general characteristics of machine learning algorithms, machine learning model structures, and machine learning training/serving data. In line with this approach, I will first present a system that provides both low-latency and low-cost machine learning serving (inferencing) over large-scale continuously-growing datasets (e.g. videos). Shifting the focus to model training, I will then present a system that makes machine learning training over geo-distributed datasets as fast as training within a single data center. Finally, I will discuss our ongoing efforts to tackle a fundamental and largely overlooked problem: machine learning training over skewed data partitions (e.g., facial images collected by cameras in different countries).

Peer-to-Peer

Youtube search... ...Google search



Introduction to Decentralized P2P Apps
Most people think peer-to-peer (P2P) networks are just for file sharing, but it turns out you can also build other types of applications on P2P networks with advantages like enhanced privacy and security. We’ll walk through the process of building an increasingly complex P2P cloud storage system (think Dropbox), and touch on the challenges you’d run into and some of their possible solutions. Topics include efficiently locating data within a large network and building a system where we can trust random people on the internet with our personal files. EVENT: SFNode Meetup July 2018 SPEAKER: Dylan Barnard PERMISSIONS: SFNode Meetup Organizer provided Coding Tech with the permission to republish this video.

What is a Peer to Peer Network? Blockchain P2P Networks Explained
A peer to peer network, often referred to as p2p network, is one of the key aspects of blockchain technology. In this video, we break down the complexity of peer to peer networks by first defining what a network is and how p2p networks differ from traditional networks. Learn more about P2P Networks

Proxy

Youtube search... ...Google search

Proxy vs. Reverse Proxy (Explained by Example)
Hussein Nasser In this episode we explain the difference between a Proxy (Forward proxy) and Reverse Proxy by example, and list all the benefits of each server. Online diagram tool used in this video: Http://www.gliffy.com

Proxy vs. Peer-to-Peer (P2P) Connections | remote.it webinar
In this webinar, we will explain how each connection type works, and in what applications you may prefer to use one or the other. You will learn how to use remote.it on Windows or macOS and on mobile (iOS/Android) apps to make P2P connections while we present the advantages of P2P versus traditional proxy connections from the remote.it web portal. All while making port forwardless connections.

Submarine Scenario

Youtube search... ...Google search

Microsoft’s SharePoint, Exchange, and Office 365 products run on Azure and Azure Stack, as do Microsoft’s database, e-commerce, and software development products. Extend Azure services and capabilities to your environment of choice—from the datacenter to edge locations and remote offices—with Azure Stack. Build, deploy, and run hybrid and edge computing apps consistently across your IT ecosystem, with flexibility for diverse workloads. The Azure Stack Hub architecture lets you provide Azure services at the edge for remote locations or intermittent connectivity, disconnected from the internet. You can create hybrid solutions that process data locally in Azure Stack Hub and then aggregate it in Azure for additional processing and analytics Azure Stack Finally Takes Microsoft Public Cloud Private | Paul Teich - The Next Platform

Sometimes, this kind of environment is also referred to as a 'submarine' scenario. - Microsoft

In 2021, the DoD CIO designated the Department of Navy CIO as the executive agent to lead a cross-service joint working group focused on Denied-Disconnected, Intermittent, and Low bandwidth (D-DIL)... Network server software and hardware exist at the tactical edge to provide critical IT services and data in these DDIL environments, along with a variety of spectrum communications and unclassified & classified network transports leveraging satellite links and low-Earth Orbit (LEO), Wi-Fi, cellular/4G LTE, millimeter wave/5G and others. The working group has teamed up with industry to refine DoD-unique requirements and use cases, resulting in the development of standardized architectures and solutions for the relevant collaboration and productivity tools (email, chat, voice and video, file management). These tools operate as a hybrid capability, which will allow users access to the full feature set when cloud connectivity is available, but remain productive locally within the DDIL environment. DoD working with Industry to Adapt Cloud Tools for the Tactical Edge - DON CIO

Additionally, as DoD enterprise IT moves to the cloud, tactical networks must unify access to data and applications from the enterprise level to the tactical edge. This means deploying cloudlike services at the tactical edge of the network, so that data is available at the edge even when WAN connectivity is unavailable.Four future trends In tactical network modernization | US Army

Azure Stack for hybrid compute and disconnected scenarios
Latest updates to hybrid compute using the power of Microsoft's Azure cloud on-premises with Azure Stack. Partner Director Program Manager, Natalia Mackevicius will show you how to run hybrid apps and process data between Azure Stack and your data center and in the Azure public cloud to meet your regulatory and policy requirements. Plus, see how you can use Azure Stack as a powerful control plane for disconnected scenarios to harness insights from your IoT devices running on the Edge.

Introduction to Microsoft 365 | Windows, EMS, Azure, Azure Stack | Versions and Deployment Models
Azure Stack is an Azure addon that may be used in datacenter and hybrid environments. This lets you execute apps on the same platform on the edge, which Microsoft calls hybrid cloud and fully disconnected solutions. It lets you operate cloud applications that connect to on-premises data for regulatory needs while using a single DevOps and maintenance stack. So developers can code for Azure and IT pros can manage it. They have a single platform to manage cloud and on-premises data. Office 365 and Azure both run in the public cloud, whereas private and hybrid clouds employ Azure Stack.

Can I Use Office 365 Offline?
Office 365 is Microsoft's suite of cloud services for office productivity, so you may be thinking it can only be used online. However, Office 365 comes with the offline capability to make sure you can stay productive at any time, whether or not you’re connected to the internet.

Azure AD Connect Sync and Cloud Sync, What’s the Difference?
Many organizations use Azure AD Connect Sync to synchronize hybrid identities from Windows AD to Azure AD DS. Microsoft recently announces a new service, Azure AD Connect Cloud Sync, that also synchronizes Windows AD identities to Azure AD. In this video, we go over how they are similar, features that are different, and when to use one or the other.