Recommendation system based on user actions in the social network
https://doi.org/10.15514/ISPRAS-2020-32(3)-9
Abstract
Currently, a large number of people use various photo hosting services, social networks, online services, and so on. At the same time, users leave a lot of information about themselves on the Internet. These can be photos, comments, geotags, and so on. This information can be used to create a system that can identify different target groups of users. In the future, you can run ad campaigns based on target groups, create recommendation ads, and so on. This article will discuss a system that allows users to identify their interests based on their actions in a social network. The following features were selected for analysis: published photos and text, comments on posts, information about favorite publications, and geotags. To identify target groups, the task was to analyze images in photos and analyze text. Image analysis involves object recognition, and text analysis involves highlighting the main theme of the text and analyzing the tone of the text. The analysis data is combined using a unique identifier with the rest of the information and allows you create a data showcase that can be used to select target groups using a simple SQL-query.
About the Authors
Vitaly Viktorovich MONASTYREVRussian Federation
Student
Pavel Dmitrievich DROBINTSEV
Russian Federation
Ph.D., Associate Professor
References
1. Recommendation Technology ‘Disco’. URL: https://yandex.com/company/technologies/disco/.
2. Covington P., Adams J., Sargin E. Deep Neural Networks for YouTube Recommendations. In Proc. of the 10th ACM Conference on Recommender Systems - RecSys ’16, 2016, pp. 191-198.
3. Gomez-Uribe C.A., Hunt N. The Netflix Recommender System. ACM Transactions on Management Information Systems, vol. 6, issue 4, 2015, pp. 1-19.
4. VK Experts. URL: https://vk.com/press/theme-feeds.
5. Matrixnet. URL: https://yandex.ru/company/technologies/matrixnet/.
6. Joel Murach. Murach's MySQL. Mike Murach & Associates, 2012, 612 p.
7. TensorFlow models, GitHub. URL: https://github.com/tensorflow/models.
8. Szegedy C., Vanhoucke V., Ioffe S., Shlens J., Wojna Z. Rethinking the Inception Architecture for Computer Vision. In Proc. of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 2818-2826.
9. synsets for Task 2 (same as in ILSVRC2012). URL: http://image-net.org/challenges/LSVRC/2014/browse-synsets.
10. Bíró I., Szabó J. Latent Dirichlet Allocation for Automatic Document Categorization. Lecture Notes in Computer Science, vol. 5782, 2009, pp. 430-441.
11. Gensim project page. URL: https://pypi.org/project/gensim/.
12. NLTK project page. URL: https://www.nltk.org/.
13. pyLDAvis project page. URL: https://www.nltk.org/.
14. Thematic modeling using Gensim (Python). URL: https://webdevblog.ru/tematicheskoe-modelirovanie-s-pomoshhju-gensim-python/.
15. Jin R., Lu L., Lee J., Usman A. Multi‐representational convolutional neural networks for text classification. Computational Intelligence, vol. 35, issue 3, ,2019, 599–609.
16. Text tonality analysis using convolutional neural networks. URL: https://habr.com/ru/company/mailru/blog/417767/.
17. Cliche M. BB_twtr at SemEval-2017 Task 4: Twitter Sentiment Analysis with CNNs and LSTMs. In Proc. of the 11th International Workshop on Semantic Evaluation (SemEval-2017), 2017, pp. 573-580.
18. Zhang Y., Wallace B.A Sensitivity Analysis of (and Practitioners' Guide to) Convolutional Neural Networks for Sentence Classification. arXiv preprint arXiv:1510.03820, 2015.
19. Rubtsova Y.V. (2015). Constructing a corpus for sentiment classification training. Programmnye produkty i sistemy, no. 27, 2015, pp. 72-78 (in Russian) / Рубцова Ю.А. Построение корпуса текстов для настройки тонового классификатора. Программные продукты и системы, no. 27, 2015 г., стр. 72-78.
20. Keras project page. URL: https://keras.io/.
Review
For citations:
MONASTYREV V.V., DROBINTSEV P.D. Recommendation system based on user actions in the social network. Proceedings of the Institute for System Programming of the RAS (Proceedings of ISP RAS). 2020;32(3):101-108. https://doi.org/10.15514/ISPRAS-2020-32(3)-9