skip to main content
10.1145/1076034.1076082acmconferencesArticle/Chapter ViewAbstractPublication PagesirConference Proceedingsconference-collections
Article

Multi-labelled classification using maximum entropy method

Authors Info & Claims
Published:15 August 2005Publication History

ABSTRACT

Many classification problems require classifiers to assign each single document into more than one category, which is called multi-labelled classification. The categories in such problems usually are neither conditionally independent from each other nor mutually exclusive, therefore it is not trivial to directly employ state-of-the-art classification algorithms without losing information of relation among categories. In this paper, we explore correlations among categories with maximum entropy method and derive a classification algorithm for multi-labelled documents. Our experiments show that this method significantly outperforms the combination of single label approach.

References

  1. Benson, S. J., McInnes, L. C., Moré, J., & Sarich, J. (2004). TAO user manual (revision 1.7) (Technical Report ANL/MCS-TM-242). Mathematics and Computer Science Division, Argonne National Laboratory. http://www.mcs.anl.gov/tao.]]Google ScholarGoogle Scholar
  2. Cai, L., & Hofmann, T. (2004). Hierarchical document categorization with support vector machines CIKM '04: Proceedings of the Thirteenth ACM conference on Information and knowledge management (pp. 78--87). Washington, D.C., USA: ACM Press.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Chen, S. F., & Rosenfeld, R. (1999). A Gaussian prior for smoothing maximum entropy models (Technical Report CMU-CS-99-108). School of Computer Science Carnegie Mellon University.]]Google ScholarGoogle ScholarCross RefCross Ref
  4. Clare, A., & King, R. D. (2001). Knowledge discovery in multi-label phenotype data. PKDD '01: Proceedings of the 5th European Conference on Principles of Data Mining and Knowledge Discovery (pp. 42--53). Springer-Verlag.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Comite, F. D., Gilleron, R., & Tommasi, M. (2001). Learning multi-label alternating decision trees and applications. Proceedings of CAP'01 (pp. 195--210).]]Google ScholarGoogle Scholar
  6. Crammer, K., & Singer, Y. (2002). A new family of online algorithms for category ranking. Proceedings of the 25th annual international ACM SIGIR conference on Research and development in information retrieval (pp. 151--158). Tampere, Finland: ACM Press.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Della Pietra, S., Della Pietra, V. J., & Lafferty, J. D. (1997). Inducing features of random fields. IEEE Transactions on Pattern Analysis and Machine Intelligence, 19, 380--393.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Elisseeff, A., & Weston, J. (2002). A kernel method for multi-labelled classification. Advances in Neural Information Processing Systems 14 (pp. 681--687). Cambridge, MA: MIT Press.]]Google ScholarGoogle Scholar
  9. Gao, S., Wu, W., Lee, C.-H., & Chua, T.-S. (2004). A mfom learning approach to robust multiclass multi-label text categorization. ICML '04: Twenty-first international conference on Machine learning. Banff, Alberta, Canada: ACM Press.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Godbole, S., & Sarawagi, S. (2004). Discriminative methods for multi-labeled classification. PAKDD.]]Google ScholarGoogle Scholar
  11. Har-Peled, S., Roth, D., & Zimak, D. Constraint classification for multiclass classification and ranking. In S. T. S. Becker and K. Obermayer (Eds.), Advances in neural information processing systems 15. MIT Press.]]Google ScholarGoogle Scholar
  12. Jaynes, E. T. (1957). Information theory and statistical mechanics. Physical Review, 106, 620--630.]]Google ScholarGoogle ScholarCross RefCross Ref
  13. Malouf, R. (2002). A comparison of algorithms for maximum entropy parameter estimation. Proc. of the sixth CoNLL.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. McCallum, A. (1999). Multi-label text classification with a mixture model trained by EM. AAAI'99 Workshop on Text Learning.]]Google ScholarGoogle Scholar
  15. Nigam, K., Lafferty, J., & McCallum, A. (1999). Using maximum entropy for text classification. IJCAI-99 Workshop on Machine Learning for Information Filtering (pp. 61--67).]]Google ScholarGoogle Scholar
  16. Schapire, R. E., & Singer, Y. (2000). Boostexter: A boosting-based system for text categorization. Machine Learning, 39, 135--168.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Ueda, N., & Saito, K. Parametric mixture models for multi-labeled text. Advances in Neural Information Processing Systems 15. MIT Press.]]Google ScholarGoogle Scholar
  18. Wilcoxon, F. (1945). Individual comparisons by ranking methods. Biometrics, 1, 80--93.]]Google ScholarGoogle ScholarCross RefCross Ref
  19. Yang, Y., & Liu, X. (1999). A re-examination of text categorization methods. Proceedings of the 22nd Annual International Conference on Research and Development in Information Retrieval (SIGIR'99) (pp. 42--49). Berkley: ACM Press.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. Zhang, T., & Oles, F. J. (2001). Text categorization based on regularized linear classification methods. Inf. Retr., 4, 5--31.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. Zhu, J., & Hastie, T. (2003). Classification of gene microarrays by penalized logistic regression. Biostatistics.]]Google ScholarGoogle Scholar

Index Terms

  1. Multi-labelled classification using maximum entropy method

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      SIGIR '05: Proceedings of the 28th annual international ACM SIGIR conference on Research and development in information retrieval
      August 2005
      708 pages
      ISBN:1595930345
      DOI:10.1145/1076034

      Copyright © 2005 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 15 August 2005

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • Article

      Acceptance Rates

      Overall Acceptance Rate792of3,983submissions,20%

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader