Non-negativity is vital for latent factor models to preserve the important feature of most high-dimensional and sparse (HiDS) matrices, e.g., none of their entries is negative. With the consideration of non-negativity, the training process of a latent factor model should be specified to achieve constraints-incorporated learning schemes. However, such schemes are neither flexible nor extensible. This work investigates algorithms of inherently non-negative latent factor analysis, which separates non-negativity constraints from the training process. Based on a deep investigation into the learning objective of a non-negative latent factor model, we separate the desired latent factors from decision variables involved in the training process via a single-element-dependent mapping function that makes the output factors inherently non-negative. Then we theoretically prove that the resultant model is able to represent the original one effectively. As a result, we design a highly efficient algorithm to bring the Inherently Non-negative Latent Factor model into practice. Experimental results on three HiDS matrices from industrial recommender systems show that compared with state-of-the-art non-negative latent factor models, the proposed one is able to obtain advantage in prediction accuracy with comparable or higher computational efficiency. Moreover, such high performance is achieved through its unconstrained optimization process on the premise of fulfilling the non-negativity constraints. Hence, the proposed model is highly valuable for industrial applications required to handle HiDS matrices subject to non-negativity constraints.