计算机科学
可扩展性
再培训
杠杆(统计)
依赖关系(UML)
阻塞(统计)
分布式计算
方案(数学)
人工智能
计算机网络
数据库
数学
数学分析
业务
国际贸易
作者
Pengfei Li,Yu Hua,Jingnan Jia,Pengfei Zuo
出处
期刊:Proceedings of the VLDB Endowment
[VLDB Endowment]
日期:2021-10-01
卷期号:15 (2): 321-334
被引量:31
标识
DOI:10.14778/3489496.3489512
摘要
Index structures in memory systems become important to improve the entire system performance. The promising learned indexes leverage deep-learning models to complement existing index structures and obtain significant performance improvements. Existing schemes rely on a delta-buffer to support the scalability, which however incurs high overheads when a large number of data are inserted, due to the needs of checking both learned indexes and extra delta-buffer. The practical system performance also decreases since the shared delta-buffer quickly becomes large and requires frequent retraining due to high data dependency. To address the problems of limited scalability and frequent retraining, we propose a FINE-grained learned index scheme with high scalability, called FINEdex, which constructs independent models with a flattened data structure (i.e., the data arrays with low data dependency) under the trained data array to concurrently process the requests with low overheads. By further efficiently exploring and exploiting the characteristics of the workloads, FINEdex processes the new requests in-place with the support of non-blocking retraining, hence adapting to the new distributions without blocking the systems. We evaluate FINEdex via YCSB and real-world datasets, and extensive experimental results demonstrate that FINEdex improves the performance respectively by up to 1.8× and 2.5× than state-of-the-art XIndex and Masstree. We have released the open-source codes of FINEdex for public use in GitHub.
科研通智能强力驱动
Strongly Powered by AbleSci AI