skip to main content
10.1145/1321440.1321498acmconferencesArticle/Chapter ViewAbstractPublication PagescikmConference Proceedingsconference-collections
research-article

A two-stage approach to domain adaptation for statistical classifiers

Authors Info & Claims
Published:06 November 2007Publication History

ABSTRACT

In this paper, we consider the problem of adapting statistical classifiers trained from some source domains where labeled examples are available to a target domain where no labeled example is available. One characteristic of such a domain adaptation problem is that the examples in the source domains and the target domain are known to follow different distributions. Thus a regular classification method would tend to overfit the source domains. We present a two-stage approach to domain adaptation, where at the first <generalization stage, we look for a set of features generalizable across domains, and at the second adaptation stage, we pick up useful features specific to the target domain. Observing that the exact objective function is hard to optimize, we then propose a number of heuristics to approximately achieve the goal of generalization and adaptation. Our experiments on gene name recognition using a real data set show the effectiveness of our general framework and the heuristics.

References

  1. R. K. Ando and T. Zhang. A framework for learning predictive structures from multiple tasks and unlabeled data. Journal of Machine Learning Research, 6:1817--1853, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. S. Ben-David, J. Blitzer, K. Crammer, and F. Pereira. Analysis of representations for domain adaptation. In Advances in Neural Information Processing Systems 19, 2007.Google ScholarGoogle Scholar
  3. J. C. Bezdek and R. J. Hathaway. Some notes on alternating optimization. In Proceedings of the 2002 AFSS International Conference on Fuzzy Systems, pages 288--300, 2002. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. J. Blitzer, R. McDonald, and F. Pereira. Domain adaptation with structural correspondence learning. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, pages 120--128, 2006. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Y. S. Chan and H. T. Ng. Estimating class priors in domain adaptation for word sense disambiguation. In Proceedings of the 21st International Conference on Computational Linguistics and 44th Annual Meeting of the Association for Computational Linguistics, pages 89--96, 2006. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. C. Chelba and A. Acero. Adaptation of maximum entropy capitalizer: Little data can help a lot. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, pages 285--292, 2004.Google ScholarGoogle Scholar
  7. H. Daumé III and D. Marcu. Domain adaptation for statistical classifiers. Journal of Artificial Intelligence Research, 26:101--126, 2006. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. J. Finkel, S. Dingare, C. D. Manning, M. Nissim, B. Alex, and C. Grover. Exploring the boundaries: gene and protein identification in biomedical text. BMC Bioinformatics, 6(Suppl 1):S5, 2005.Google ScholarGoogle ScholarCross RefCross Ref
  9. R. Florian, H. Hassan, A. Ittycheriah, H. Jing, N. Kambhatla, X. Luo, N. Nicolov, and S. Roukos. A statistical model for multilingual entity detection and tracking. In Proceedings of the Human Language Technology Conference of the North American Chapter of the Association for Computational Linguistics, pages 1--8, 2004.Google ScholarGoogle ScholarCross RefCross Ref
  10. D. W. Hosmer and S. Lemeshow. Applied Logistic Regression. Wiley Series in Probability and Statistics. John Wiley &amp; Sons, Inc., 2000.Google ScholarGoogle Scholar
  11. J. Jiang and C. Zhai. Exploiting domain structure for named entity recognition. In Proceedings of The Human Language Technology Conference of the North American Chapter of the Association for Computational Linguistics, pages 74--81, 2006. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. X. Li and J. Bilmes. A Bayesian divergence prior for classifier adaptation. In Proceedings of the 11th International Conference on Artificial Intelligence and Statistics, 2007.Google ScholarGoogle Scholar
  13. B. Roark and M. Bacchiani. Supervised and unsupervised PCFG adaptatin to novel domains. In Proceedings of the Human Language Technology Conference of the North American Chapter of the Association for Computational Linguistics, pages 126--133, 2003. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. V. N. Vapnik. The Nature of Statistical Learning Theory. Springer-Verlag New York, Inc., 1995. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. X. Zhu. Semi-supervised learning literature survey. Technical Report 1530, University of Wisconsin, 2005.Google ScholarGoogle Scholar

Index Terms

  1. A two-stage approach to domain adaptation for statistical classifiers

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader