LEARNING DYADIC DATA AND PREDICTING UNACCOMPLISHED CO-OCCURRENT VALUES BY MIXTURE MODEL

Authors

  • L Nguyen Loc Nguyen’s Academic Network, Vietnam
  • MH Lanuza Philippine Normal University, Manila, Philippines

DOI:

https://doi.org/10.17501/26307413.2023.6112

Keywords:

dyadic data, co-occurrence data, expectation maximization (EM) algorithm, mixture model

Abstract

Dyadic data which is also called co-occurrence data (COD) contains co-occurrences of objects where these objects are indexed and grouped into two finite sets. It is necessary to model dyadic data by applied mathematical tools because dyadic data analysis is interesting and important to many applications relating to indexed two-dimensional data such as image processing and recommendation collaborative filtering. Fortunately, finite mixture model is a solid statistical model to learn and make inference on dyadic data because mixture model is built smoothly and reliably by expectation maximization (EM) algorithm which is suitable to inherent spareness of dyadic data. This research summarizes mixture models for dyadic data, in which there are three well-known models such as symmetric mixture model (SMM), asymmetric mixture model (AMM), and product-space mixture model (PMM) which are described by beautiful mathematical proofs and explanations derived from EM algorithm. Objects in traditional dyadic data are indexed as categories and so their potential real values are concerned because of potential applications and extensions of dyadic data analysis. For instance, when each co-occurrence in dyadic data is associated with a real value, there are many unaccomplished values because a lot of co-occurrences are inexistent. In the research, these unaccomplished values are estimated as mean (expectation) of random variable given partial probabilistic distributions inside dyadic mixture model. This estimation result is solid due to support of EM algorithm.

Downloads

Download data is not yet available.

References

Hofmann, T. (2004, January). Latent Semantic Models for Collaborative Filtering. ACM Transactions on Information Systems (TOIS), 22(1), 89-115. doi:10.1145/963770.963774

Hofmann, T., & Puzicha, J. (1998). Statistical Models for Co-occurrence Data. Massachusetts Institute of Technology, Artificial Intelligence

Laboratory. MIT Publisher. Retrieved from https://dspace.mit.edu/bitstream/handle/1721.1/7253/AIM-1625.pdf?sequence=2

Hofmann, T., & Puzieha, J. (1999). Latent Class Models for Collaborative Filtering. In T. Dean (Ed.), Proceedings of the Sixteenth International Joint Conference on Artificial Intelligence (IJCAI '99) (pp. 688-693). San Francisco, CA, USA: Morgan Kaufmann. Retrieved from https://dl.acm.org/citation.cfm?id=687583

Hofmann, T., Puzicha, J., & Jordan, M. I. (1998). Learning from Dyadic Data. In M. J. Kearns, S. A. Solla, & D. A. Cohn (Ed.), Advances in Neural Information Processing Systems 11 (NIPS 1998). 11, pp. 466-472. Denver: MIT Press. Retrieved from https://papers.nips.cc/paper/1503-learning-from-dyadic-data

Nguyen, L. (2020). Tutorial on EM algorithm. MDPI. Preprints. doi:10.20944/preprints201802.0131.v8

Downloads

Published

2023-12-02

How to Cite

Nguyen, L., & Lanuza, M. (2023). LEARNING DYADIC DATA AND PREDICTING UNACCOMPLISHED CO-OCCURRENT VALUES BY MIXTURE MODEL. Proceedings of the International Conference on Future of Education, 6(1), 152–171. https://doi.org/10.17501/26307413.2023.6112