Abstract
High-Dimensional and Incomplete (HDI) data are commonly seen in various big-data-related applications concerning the inherent non-negativity interactions among numerous nodes. A Non-negative Latent Factor Analysis (NLFA) model performs efficient representation learning to such HDI data. However, existing NLFA models all adopt a static divergence metric like Euclidean distance or α-β divergence to build its learning objective, which evidently restricts its scalability in representing HDI data from different domains. Aiming at addressing this critical issue, this study proposes an Adaptive Divergence-based Non-negative Latent-factor-analysis (ADNL) model with three-fold ideas: a) generalizing the objective function with the α-β-divergence to expand its potential of representing various HDI data; b) facilitating a smooth non-negative bridging function to connect the optimization variables with output latent factors for keeping non-negativity; and c) making the divergence parameters adaptive through position-transitional particle swarm optimization, thereby facilitating adaptive divergence in the learning objective to achieve high scalability. Empirical studies on six HDI datasets from real applications demonstrate that an ADNL model outperforms the state-of-the-art models in both estimation accuracy and computational efficiency for missing data of an HDI matrix.
Original language | English (US) |
---|---|
Pages (from-to) | 1209-1222 |
Number of pages | 14 |
Journal | IEEE Transactions on Emerging Topics in Computational Intelligence |
Volume | 8 |
Issue number | 2 |
DOIs | |
State | Published - Apr 1 2024 |
All Science Journal Classification (ASJC) codes
- Computer Science Applications
- Control and Optimization
- Computational Mathematics
- Artificial Intelligence
Keywords
- Adaptive divergence
- big data
- generalized divergence
- high-dimensional and incomplete data
- non-negative latent factor analysis
- stochastic momentum methods