%0 Conference Proceedings %T M-DRL: Deep Reinforcement Learning Based Coflow Traffic Scheduler with MLFQ Threshold Adaption %+ School of Computer Science and Engineering [Beijing] %A Chen, Tianba %A Li, Wei %A Sun, Yukang %A Li, Yunchun %Z Part 2: AI %< avec comité de lecture %( Lecture Notes in Computer Science %B 17th IFIP International Conference on Network and Parallel Computing (NPC) %C Zhengzhou, China %Y Xin He %Y En Shao %Y Guangming Tan %I Springer International Publishing %3 Network and Parallel Computing %V LNCS-12639 %P 80-91 %8 2020-09-28 %D 2020 %R 10.1007/978-3-030-79478-1_7 %K Coflow %K Datacenter network %K Deep reinforcement learning %Z Computer Science [cs]Conference papers %X The coflow scheduling in data-parallel clusters can improve application-level communication performance. The existing coflow scheduling method without prior knowledge usually uses Multi-Level Feedback Queue (MLFQ) with fixed threshold parameters, which is insensitive to coflow traffic characteristics. Manual adjustment of the threshold parameters for different application scenarios often has long optimization period and is coarse in optimization granularity. We propose M-DRL, a deep reinforcement learning based coflow traffic scheduler by dynamically setting thresholds of MLFQ to adapt to the coflow traffic characteristics, and reduces the average coflow completion time. Trace-driven simulations on the public dataset show that coflow communication stages using M-DRL complete 2.08$$\times $$×(6.48$$\times $$×) and 1.36$$\times $$×(1.25$$\times $$×) faster on average coflow completion time (95-th percentile) in comparison to per-flow fairness and Aalo, and is comparable to SEBF with prior knowledge. %G English %Z TC 10 %Z WG 10.3 %2 https://inria.hal.science/hal-03768743/document %2 https://inria.hal.science/hal-03768743/file/511910_1_En_7_Chapter.pdf %L hal-03768743 %U https://inria.hal.science/hal-03768743 %~ IFIP-LNCS %~ IFIP %~ IFIP-TC %~ IFIP-TC10 %~ IFIP-NPC %~ IFIP-WG10-3 %~ IFIP-LNCS-12639