A Deep Reinforcement Learning Perspective on Internet Congestion Control
http://proceedings.mlr.press/v97/jay19a/jay19a.pdf
Internet congestion control: one of the most fundamental and challenging problems in communication networks
Determines what you get out of the internet
At what rate data goes in
Always running on every connection
With no prior knowledge
Water pipe
How fast do we send the data?
Too slow: harm performance
Too fast: overflow the pipe
Internet CC
Very limited information, massive agent churn, enormous, dynamic, complicated network
CC revisited
Default algorithm: transmission control algorithm (TCP)
Motivating deep RL
CC protocol
Blackbox
Locally-perceived history of feedback: loss rates, latency
On the other end: next sending rate
Hypothesis: this feedback contains information about patterns that can improve the choice of sending rates
RL formulation
Goal: maximize the cumulative reward
Agent observe the state, new action in the environment, output a reward
NN: agent, and simulator for the environment
Observe state, based on locally-perceived feedback
Aurora
Monitor interval: send ratio, latency ration, latency inflation
Keep design choice: scale-free observations affect robustness
3-layer NN
Rate change factor
Training
Very simple simulated network
Each episode chooses different link parameters
Entire training platform: PCC-RL
Real-World testing setup
Real packets in linux kernel network emulation
Using inference only
Allowing faster adaptation
Still producing state-of-the-art results
Last updated
Was this helpful?