Workshop
https://www.youtube.com/watch?v=gfGTd7PXK54&list=PLMPUUgLIYH1ZJVEXTTZT82ipTprSu1hn8&index=1
Learning at the Wireless Edge
Vince Poor (Princeton University)
Two aspects
Using ML to optimize communication networks
Learning on mobile devices (the focus of today's talk)
Today's talk: focus on federated learning
Motivation
Federated learning over wireless channels (scheduling)
Privacy protection in federated learning (differential privacy)
Some research issues
ML
Tremendous progress in recent years (more data, increase in computational power)
Standard ML: implement in centralized manner (data center, cloud), full access to the data
SOTA models:
Standard software tools, specialized hardware
Wireless edge
Centralized ML are not suitable for many emerging applications
Self-driving cars, first responder networks, healthcare networks
What makes the application different:
Data is born at the edge (phone, IoT devices)
Limited capacity uplinks
Low latency & high reliability
Data privacy / security
Scalability & locality
Motivate moving learning closer to the network edge
Federated learning over wireless channels (scheduling)
Wireless: communication to the AP needs to go through wireless channels
Shared, resource-constrained
Only limited number of devices can be selected in each update round
Transmissions are not reliable due to interference
Questions
How should we schedule devices to update the trained weights?
How does the interference affect the training?
Scheduling mechanisms
Random scheduling: aggregator select N out of K users at random
Round Robin: divide into group
Proportional Fair: strongest SNRs
Design metric: age of information (AoI)
Age-based scheduling scheme for federated learning in mobile edge networks
Optimization algorithm in each iteration round
Wireless round robin
Privacy in federated learning
"privacy preserving": data remains on end-user devices
But end-user data can be inferred from the parameter (or gradient) updates
Approach: use differential privacy to protect end-user data
Refers to a type of privacy in which two datasets, one with private information and one without it, but otherwise identical, cannot be distinguished by a statistical query (with high probability)
Trade-off between privacy and accuracy
Other issues
Model efficiency
Resources on end-user devices are limited (e.g., energy, storage, computational power)
Trade-offs between # of layers, # of neurons per layer, accuracy
Communication efficiency
Limited data at the edge
Local data is sparse
Incorporating domain and physical knowledge
Security & Privacy
Robustness to malicious end-user devices & adversarial training examples
Other approaches to end-user privacy
Challenges in ML and the way forward
Ariela Zeira, Intel Labs
Challenges in DL
Compute efficiency
Memory overhead
Data efficiency
Online learning
Robustness
Knowledge Representation
Hyper Dimensional Computing
New paradigm for energy-efficient, noise-robust and fast alternatives to standard ML
Adventures in Learning-Based Rate Control
Brighten Godfrey, UIUC
TCP protocols: point solutions designed for specific environments, and far from optimal
Why does traditional CC architecture struggle?
TCP Reno, CUBIC, FAST, Scalable, HTCP
"Hardwired" control actions
Underlying conditions --> ideal control action
Not enough information about what's happening in the network
What is the right rate to send?
Network is a blackbox: but we can send at some rate and see what happens
Collect observations: throughput, loss rate, latency. We can summarize that in a utility function
A change in perspective
Traditional perspective: simple network model + well-crafted rules --> predictable results
"Black-box" perspective: the world is complex. Quantifying goal and observing effect of actions yields good decisions
A fit for learning!
Diverse, opaque environments
Only a trickle of information
Infer good action at millisecond timescales
Software components
Paper: PCC
Control algorithm: heuristic hill-climbing algorithm
Noise in measurement? randomized controlled trials
Where is the congestion control?
Equilibrium depends on utility function
Selfish utility-maximizing decision --> non-cooperative game
Promising performance
Upgrade: PCC Vivace (NSDI 2018)
Leveraging powerful tools from online learning theory
New utility function framework
Latency-awareness
Strictly concave --> equilibrium guarantee
Weighted fairness among senders
New control algorithm
Gradient-ascent --> convergence speed / stability
Deals with measurement noise
Performance: great improvement in latency & responsiveness, but still suboptimal in extremely dynamic networks (i.e. wireless)
Deep RL on congestion control (ICML 2019)
scale-free values to aid robustness
History length: what lengths work well
Training
Simulated environment
Order of magnitude faster than emulation
Each episode chooses link parameters from a range
Setting appropriate discount factor
Maximize expected cumulative discounted return
Future
Multi-gent scenarios: training, competition
Online training: challenge is to improve outcomes with limited additional training data
New uses
Scavenger transport
SIGCOMM 2020: Proteus: Scavenger Transport and Beyond
Different applications: software updates, CDN warmup, cloud storage replication, online video, real-time streaming, search
Elastic timing, inelastic timing
Scavenger design goals
Yielding: minimally impact primary flows
Performance: high utilization, low latency when only scavengers exist
Flexibility: dynamically switch, avoid separate implementation
Utility functions
Primary, Scavenger, Hybrid
RTT deviation as a competition indicator
Definition: standard deviation of observed RTT samples
Intuition: earlier signal of dynamics of buffer occupancy during flow competition
Proteus: yields more effectively
Cross-layer design
Dynamic threshold based on the application requirement (buffer occupancy)
Hybrid mode: to get the bandwidth when they need it
Improving QoE (rebuffer raito)
Rate control robustness (HotNets 2019)
Keeping up with rapid change
Recent acceleration of innovation in rate control
Approach: ML as an adversary
It gets rewarded when it finds environmental parameters that cause this algorithm under test to perform poorly (looking for the hard cases)
Do this carefully, rewarded for suboptimal performance of the algorithm, and smoothing to make it more useful results to see
Reward = -1 * protocol score + optimal score - smoothing penalty
Implement: ABR video, CC
Lessons learned
What worked
Modular architecture
New control algorithms
New utility functions open new issues
Learning-based control can improve performance over traditional protocols
Industry implementations
Open challenges & opportunities
Performance: fast decisions v.s careful decisions
Even one RTT can be a long time - especially in fluctuating wireless environments
Understanding protocol robustness
Existing opportunities in system design
Complexity in systems, environments, and application needs lead to opportunity for inference
Restructure systems for a learning mindset
Last updated
Was this helpful?