%0 Conference Proceedings %T LCache: Machine Learning-Enabled Cache Management in Near-Data Processing-Based Solid-State Disks %+ Anhui University [Hefei] %+ Huazhong University of Science and Technology [Wuhan] (HUST) %A Sun, Hui %A Dai, Shangshang %A Cui, Qiao %A Huang, Jianzhong %Z Part 2: AI %< avec comité de lecture %( Lecture Notes in Computer Science %B 17th IFIP International Conference on Network and Parallel Computing (NPC) %C Zhengzhou, China %Y Xin He %Y En Shao %Y Guangming Tan %I Springer International Publishing %3 Network and Parallel Computing %V LNCS-12639 %P 128-139 %8 2020-09-28 %D 2020 %R 10.1007/978-3-030-79478-1_11 %K Near-data processing %K Machine learning %K Cache management solid state disks %Z Computer Science [cs]Conference papers %X In the era of big-data, large-scale storage systems use NAND Flash-based solid-state disks (SSDs). Some upper-level applications put higher requirements on the performance of SSD-based storage systems. SSDs typically exploit a small amount of DRAM as device side cache, yet the limitation of the DRAM inside an SSD makes a better performance difficult to achieve. The wide application of the existing cache management schemes (e.g., LRU, CFLRU) provides a solution to this problem. With the popularity of near-data processing paradigm in storage systems, the near-data processing-based SSDs are designed to improve the performance of the overall system. In this work, a new cache management strategy named LCache is proposed based on NDP-enabled SSD using a machine learning algorithm. LCache determines whether I/O requests will be accessed in a period by trained machine learning model (e.g., decision tree algorithm model) based on characteristics of I/O requests. When the infrequently accessed I/Os that are not intensive are directly flushed into the flash memory, LCahe enables to update the dirty data that has not been accessed in the cache to the flash memory. Thus, LCache can generate clean data while replace cached data with priority to minimize the cost of data evicting. LCache also can effectively achieve the threefold benefits: (1) reducing the data access frequency to frequently access data pages in flash memory, (2) improving the response time by 59.8%, 60% and 14.81% compared with LRU, CFLRU, and MQSim, respectively, and (3) optimizing the performance cliff by 68.2%, 68%, and 30.2%, respectively. %G English %Z TC 10 %Z WG 10.3 %2 https://inria.hal.science/hal-03768746/document %2 https://inria.hal.science/hal-03768746/file/511910_1_En_11_Chapter.pdf %L hal-03768746 %U https://inria.hal.science/hal-03768746 %~ IFIP-LNCS %~ IFIP %~ IFIP-TC %~ IFIP-TC10 %~ IFIP-NPC %~ IFIP-WG10-3 %~ IFIP-LNCS-12639