Skip to main content
Log in

Co-Adjustment Learning for Co-Clustering

  • Published:
Cognitive Computation Aims and scope Submit manuscript

Abstract

Co-clustering simultaneously performs clustering on the sample and feature dimensions of the data matrix, so it can obtain better insight into the data than traditional clustering. Adjustment learning extracts valuable information from chunklets for unsupervised cluster learning in specific scenarios, but in fact it can be easily extended to semi-supervised and supervised learning situations. In this paper, we propose a novel co-clustering framework, named co-adjustment learning for co-clustering (CALCC), and CALCC can be simultaneously used in unsupervised, semi-supervised and supervised learning situations. A novel co-adjustment learning (CAL) model is proposed to extract meaningful representations in both sample space and feature space for co-clustering. CAL can not only perform the sample projection as well as feature projection under the guidance of chunklet information, it can also transform the original data into another space with improved separability. We can obtain the row partition matrix and column partition matrix by performing the clustering process on the representations learned by the CAL model. In order to prove the availability of our framework, an unsupervised case of CALCC is introduced to make an extensive comparison with several related methods (specifically including the classic co-clustering methods and the state-of-the-art methods closely related to our work) on several image and real data sets. The experimental results show the superior performance of the CAL model in discovering discriminative representations and demonstrate the effectiveness of the CALCC framework. The proposed CALCC framework, as demonstrated in the experiments, is more effective superior to the related methods. In addition, the chunklet information can be effective to enhance the expression ability of the learned representations.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5

Similar content being viewed by others

References

  1. Abdullah A, Hussain A. A cognitively inspired approach to two-way cluster extraction from one-way clustered data. Cogn Comput. 2014;7(1):161–82.

    Article  Google Scholar 

  2. Banerjee A, Dhillon I, Ghosh J, Merugu S, Modha DS. A generalized maximum entropy approach to bregman co-clustering and matrix approximation. J Mach Learn Res. (Aug) 2007;8:1919–86.

  3. Bekkerman R, Ran EY, Mccallum A. Multi-way distributional clustering via pairwise interactions. International Conference on Machine Learning. 2005:41–8.

  4. Blake C. Uci repository of machine learning databases. Department of Information and Computer Science: University of California; 1998.

    Google Scholar 

  5. Busygin S. Biclustering in data mining. Computers and Operations Research. 2008;35(9):2964–87.

    Article  MathSciNet  Google Scholar 

  6. Cai D, He X, Han J, Huang TS. Graph regularized nonnegative matrix factorization for data representation. IEEE Trans Pattern Anal Mach Intell. 2011;33(8):1548–60.

    Article  Google Scholar 

  7. Chen LC, Yu PS, Tseng VS. Wf-msb: a weighted fuzzy-based biclustering method for gene expression data. Int J Data Min Bioin. 2011;5(1):89–109.

    Article  Google Scholar 

  8. Chen, X., Ritter, A., Gupta, A., Mitchell, T.: Sense discovery via co-clustering on images and text. In: IEEE Conference on Computer Vision and Pattern Recognition. 2015:5298–5306.

  9. Chen Y, Wang L, Dong M. Non-negative matrix factorization for semisupervised heterogeneous data coclustering. IEEE Transactions on Knowledge Data Engineering. 2010;22(10):1459–74.

    Article  Google Scholar 

  10. Cheng Y, Church GM. Biclustering of expression data. In: Eighth International Conference on Intelligent Systems for Molecular Biology. 2000:93–103.

  11. Dhillon I, Mallela S, Modha D. Information-theoretic co-clustering. In: ACM SIGKDD International Conference on Knowledge Discovery Data Mining. 2003:89–98.

  12. Dhillon IS. Co-clustering documents and words using bipartite spectral graph partitioning. In: ACM SIGKDD International Conference on Knowledge Discovery Data Mining. 2001:269–274.

  13. Feldman DD, Griffiths LJ. A constraint projection approach for robust adaptive beamforming. In: Acoustics, Speech, and Signal Processing, 1991. Icassp-91., 1991 International Conference. 1991:1381–84.

  14. Gao C, Mcdowell IC, Zhao S, Brown CD, Engelhardt BE. Context specific and differential gene co-expression networks via bayesian biclustering: Plos Computational Biology. 2016;12(7):e1004791.

    Google Scholar 

  15. Garcia-Pedrajas N, Maudes-Raedo J, Garcia-Osorio C, Rodriguez-Diez JJ. Supervised subspace projections for constructing ensembles of classifiers. Inf Sci. 2012;193(11):1–21.

    Article  Google Scholar 

  16. Hartigan JA. Direct clustering of a data matrix. Publications of the American Statistical Association. 1978;67(337):123–9.

    Article  Google Scholar 

  17. Herlocker JL, Konstan JA, Terveen L, Riedl J. Evaluating collaborative filtering recommender systems. ACM Transactions on Information Systems. 2004;22(1):5–53.

    Article  Google Scholar 

  18. Huang S, Wang H, Li T, Li T, Xu Z. Robust graph regularized nonnegative matrix factorization for clustering. Data Mining and Knowledge Discovery. 2018;32(2):483–503.

    Article  MathSciNet  Google Scholar 

  19. Huang S, Wang H, Li T, Yang Y, Li T. Constraint co-projections for semi-supervised co-clustering. IEEE Transactions on Cybernetics. 2015;46(12):3047–58.

    Article  Google Scholar 

  20. Huang S, Xu Z, Lv J. Adaptive local structure learning for document co-clustering. Knowledge-Based Systems. 2018;148:74–84.

    Article  Google Scholar 

  21. Jain AK, Murty MN, Flynn PJ. Data clustering: a review. ACM Computing Surveys. 1999;31(3):264–32323.

    Article  Google Scholar 

  22. Klein D, Kamvar SD, Manning CD. From instance-level constraints to space-level constraints: Making the most of prior knowledge in data clustering. In: Proceedings of the Nineteenth International Conference on Machine Learning. 2020:307–314.

  23. Kumar S, Gao X, Welch I. Learning under data shift for domain adaptation: A model-based co-clustering transfer learning solution. In: Pacific Rim Knowledge Acquisition Workshop. 2016:43–54.

  24. Leung WT, Lee DL, Lee WC. Clr:a collaborative location recommendation framework based on co-clustering. 2011:305–14.

  25. Li H, Wang M, Hua XS. Msra-mm 2.0: A large-scale web multimedia dataset. In: IEEE International Conference on Data Mining Workshops. 2009:164–69.

  26. Lu Z, Liu G, Wang S. Sparse neighbor constrained co-clustering via category consistency learning. Knowledge-Based Systems. 2020.

  27. Nie F, Wang X, Deng C, Huang H. Learning a structured optimal bipartite graph for co-clustering. In: Guyon I, Luxburg UV, Bengio S, Wallach H, Fergus R, Vishwanathan S, Garnett R (eds.) Advances in Neural Information Processing Systems 30. 2017:4129–38. Curran Associates, Inc. http://papers.nips.cc/paper/7001-learning-a-structured-optimal-bipartite-graph-for-co-clustering.pdf

  28. Peng X, Xu D. Structural regularized projection twin support vector machine for data classification. Information Sciences. 2014;279(279):416–32.

    Article  Google Scholar 

  29. Prats-Montalbn JM, Lopez F, Valiente JM, Ferrer A. Feature extraction and classification in surface grading application using multivariate statistical projection models. In: Eighth International Conference on Quality Control by Artificial Vision. 2007:63560N–63560N–11.

  30. Rodriguez A, Laio A. Clustering by fast search and find of density peaks. Machine Learning. 2014;344(6191):1492.

    Google Scholar 

  31. Shan H, Banerjee A. Bayesian co-clustering. In: Eighth IEEE International Conference on Data Mining. 2008:530–39.

  32. Shental N, Bar-Hillel A, Hertz T, Weinshall D. Gaussian mixture models with equivalence constraints. Constrained Clustering. 2009:33–58.

  33. Shental N, Hertz T, Weinshall D, Pavel M. Adjustment learning and relevant component analysis. In: European Conference on Computer Vision. 2002:776–92.

  34. Shi, X, Fan W, Yu PS. Efficient semi-supervised spectral co-clustering with constraints. In: IEEE International Conference on Data Mining. 2011:043–48.

  35. Wang P, Domeniconi C, Laskey KB. Nonparametric bayesian clustering ensembles. In: European Conference on Machine Learning and Knowledge Discovery in Databases. 2010:435–50.

  36. Webb AR, Copsey KD. Introduction to statistical pattern recognition. Academic Press, 1972:2133–43

  37. Whang JJ, Dhillon IS. Non-exhaustive, overlapping co-clustering. In: Proceedings of the 2017 ACM on Conference on Information and Knowledge Management. 2017;2367–70.

  38. Yan Y, Chen L, Tjhi WC. Fuzzy semi-supervised co-clustering for text documents. Fuzzy Sets and Systems. 2013;215(215):74–89.

    Article  MathSciNet  Google Scholar 

  39. Yu X, Yu G, Wang J, Domeniconi C. Co-clustering ensembles based on multiple relevance measures. IEEE Transactions on Knowledge and Data Engineering. 2019.

  40. Zhang, Daoqiang, Chen, Songcan, Zhou, ZhiHua, Yang, Qiang. Constraint projections for ensemble learning. National conference on artificial intelligence. 2008;758–763.

Download references

Acknowledgements

This work is partially supported by Key program for International S&T Cooperation of Sichuan Province, No. (2019YFH0097); Science and Technology Support Project of Sichuan Province under 290 Grant No. 2020YFG0045, 2020YFG0238.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Hongjun Wang.

Ethics declarations

Conflicts of Interest

The authors declare that they have no conflict of interest.

Ethical Approval

This article does not contain any studies with human participants or animals performed by any of the authors.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zhang, J., Wang, H., Huang, S. et al. Co-Adjustment Learning for Co-Clustering. Cogn Comput 13, 504–517 (2021). https://doi.org/10.1007/s12559-021-09827-8

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s12559-021-09827-8

Keywords

Navigation