TY - GEN
T1 - A Bayesian Reinforcement Learning Framework for Online Index Tuning
AU - Hasan, Md Rakibul
AU - Wu, Xiaoying
AU - Theodoratos, Dimitri
N1 - Publisher Copyright:
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2026.
PY - 2026
Y1 - 2026
N2 - Efficient index tuning is critical to maintain high query performance in database systems with dynamic workloads. Traditional off-line and heuristic-driven tuning methods often incur high overhead due to frequent reconfiguration and fail to adapt to evolving workloads. To overcome these limitations, we address online index selection and formulate it as a sequential decision-making problem under uncertainty. We propose a Bayesian Reinforcement Learning framework that adaptively tunes index configurations based solely on the observed workload history, without requiring prior knowledge of the workload. Our framework leverages Q-learning with Thompson Sampling, a posterior distribution sampling method, to adaptively maintain and refine index configurations over time. The probabilistic mechanism of our approach enables the Q-learning agent to effectively balance exploration and exploitation, allowing it to traverse the vast exponential index configuration space. Our comprehensive experimental evaluation demonstrates that our algorithm excels at online index tuning across a diverse range of workloads on a standard benchmark dataset and outperforms other index tuning algorithms which are based on alternative learning methods.
AB - Efficient index tuning is critical to maintain high query performance in database systems with dynamic workloads. Traditional off-line and heuristic-driven tuning methods often incur high overhead due to frequent reconfiguration and fail to adapt to evolving workloads. To overcome these limitations, we address online index selection and formulate it as a sequential decision-making problem under uncertainty. We propose a Bayesian Reinforcement Learning framework that adaptively tunes index configurations based solely on the observed workload history, without requiring prior knowledge of the workload. Our framework leverages Q-learning with Thompson Sampling, a posterior distribution sampling method, to adaptively maintain and refine index configurations over time. The probabilistic mechanism of our approach enables the Q-learning agent to effectively balance exploration and exploitation, allowing it to traverse the vast exponential index configuration space. Our comprehensive experimental evaluation demonstrates that our algorithm excels at online index tuning across a diverse range of workloads on a standard benchmark dataset and outperforms other index tuning algorithms which are based on alternative learning methods.
UR - https://www.scopus.com/pages/publications/105017367932
UR - https://www.scopus.com/pages/publications/105017367932#tab=citedBy
U2 - 10.1007/978-3-032-02215-8_21
DO - 10.1007/978-3-032-02215-8_21
M3 - Conference contribution
AN - SCOPUS:105017367932
SN - 9783032022141
T3 - Lecture Notes in Computer Science
SP - 259
EP - 267
BT - Big Data Analytics and Knowledge Discovery - 27th International Conference, DaWaK 2025, Proceedings
A2 - Leung, Carson K.
A2 - Dignös, Anton
A2 - Kotsis, Gabriele
A2 - Khalil, Ismail
A2 - Tjoa, A. Min
PB - Springer Science and Business Media Deutschland GmbH
T2 - 27th International Conference on Big Data Analytics and Knowledge Discovery, DaWaK 2025
Y2 - 25 August 2025 through 27 August 2025
ER -