Abstrakt: |
Recommendation Systems have obtained huge attention with notion to assist users in determining their interests by prognosticating their ratings or preferences on specific item. Concurrently, the unique capability of RL (Reinforcement Learning) agent to learn from environment for reward without training the data makes it specifically suitable approach for such systems. Due to such ability, traditional works have considered DRL (Deep RL) for recommendation system. However, existing studies faced several challenges like scalability issues, probability for overlapping of numerous values and information loss while passing into a NN (Neural Network) and improper model training which lead to incorrect recommendations. Hence, this study intends to resolve these existing pitfalls. To accomplish this, the research proposes a DRR (DRL based Recommendation) framework in accordance with actor-critic learning. In actor network, DWL-FA (Deep Weighted Likelihood-Factor Analysis) is proposed for modifying the existing DNN (Deep Neural Network) to new-environment for compensating vector through the removal of unwanted regions in network results. Attention mechanism considered in this process affords decoder with suitable information from each hidden-states of the encoder. This attention mechanism along with DWL-FA model is further capable of selectively concentrating on valuable input sequences thereby effectively learning association amongst them. This assists the trained model to learn better. Subsequently, in critic network, HMP-WU (Hidden Markov Probability-Weight Updation) is proposed for optimizing the interactions amongst users with their preferences for the recommended items (environment) and recommender system (agent). In this case, weight updation process assists in comprehending related sequences thereby resolving incorrect predictions. These proposed processes have made the system explore better results with an increase of 5.74% with regard to average of p-value. [ABSTRACT FROM AUTHOR] |