ABSTRACT
In this paper, we consider the problem of adapting statistical classifiers trained from some source domains where labeled examples are available to a target domain where no labeled example is available. One characteristic of such a domain adaptation problem is that the examples in the source domains and the target domain are known to follow different distributions. Thus a regular classification method would tend to overfit the source domains. We present a two-stage approach to domain adaptation, where at the first <generalization stage, we look for a set of features generalizable across domains, and at the second adaptation stage, we pick up useful features specific to the target domain. Observing that the exact objective function is hard to optimize, we then propose a number of heuristics to approximately achieve the goal of generalization and adaptation. Our experiments on gene name recognition using a real data set show the effectiveness of our general framework and the heuristics.
- R. K. Ando and T. Zhang. A framework for learning predictive structures from multiple tasks and unlabeled data. Journal of Machine Learning Research, 6:1817--1853, 2005. Google ScholarDigital Library
- S. Ben-David, J. Blitzer, K. Crammer, and F. Pereira. Analysis of representations for domain adaptation. In Advances in Neural Information Processing Systems 19, 2007.Google Scholar
- J. C. Bezdek and R. J. Hathaway. Some notes on alternating optimization. In Proceedings of the 2002 AFSS International Conference on Fuzzy Systems, pages 288--300, 2002. Google ScholarDigital Library
- J. Blitzer, R. McDonald, and F. Pereira. Domain adaptation with structural correspondence learning. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, pages 120--128, 2006. Google ScholarDigital Library
- Y. S. Chan and H. T. Ng. Estimating class priors in domain adaptation for word sense disambiguation. In Proceedings of the 21st International Conference on Computational Linguistics and 44th Annual Meeting of the Association for Computational Linguistics, pages 89--96, 2006. Google ScholarDigital Library
- C. Chelba and A. Acero. Adaptation of maximum entropy capitalizer: Little data can help a lot. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, pages 285--292, 2004.Google Scholar
- H. Daumé III and D. Marcu. Domain adaptation for statistical classifiers. Journal of Artificial Intelligence Research, 26:101--126, 2006. Google ScholarDigital Library
- J. Finkel, S. Dingare, C. D. Manning, M. Nissim, B. Alex, and C. Grover. Exploring the boundaries: gene and protein identification in biomedical text. BMC Bioinformatics, 6(Suppl 1):S5, 2005.Google ScholarCross Ref
- R. Florian, H. Hassan, A. Ittycheriah, H. Jing, N. Kambhatla, X. Luo, N. Nicolov, and S. Roukos. A statistical model for multilingual entity detection and tracking. In Proceedings of the Human Language Technology Conference of the North American Chapter of the Association for Computational Linguistics, pages 1--8, 2004.Google ScholarCross Ref
- D. W. Hosmer and S. Lemeshow. Applied Logistic Regression. Wiley Series in Probability and Statistics. John Wiley & Sons, Inc., 2000.Google Scholar
- J. Jiang and C. Zhai. Exploiting domain structure for named entity recognition. In Proceedings of The Human Language Technology Conference of the North American Chapter of the Association for Computational Linguistics, pages 74--81, 2006. Google ScholarDigital Library
- X. Li and J. Bilmes. A Bayesian divergence prior for classifier adaptation. In Proceedings of the 11th International Conference on Artificial Intelligence and Statistics, 2007.Google Scholar
- B. Roark and M. Bacchiani. Supervised and unsupervised PCFG adaptatin to novel domains. In Proceedings of the Human Language Technology Conference of the North American Chapter of the Association for Computational Linguistics, pages 126--133, 2003. Google ScholarDigital Library
- V. N. Vapnik. The Nature of Statistical Learning Theory. Springer-Verlag New York, Inc., 1995. Google ScholarDigital Library
- X. Zhu. Semi-supervised learning literature survey. Technical Report 1530, University of Wisconsin, 2005.Google Scholar
Index Terms
- A two-stage approach to domain adaptation for statistical classifiers
Recommendations
Mutual Domain Adaptation
Highlights- We tackle a realistic problem setting of domain adaptation, where most domains are label-deficient and need to be helped and recent data become more sparsely labeled which makes the learning even more difficult.
- To tackle this problem, ...
AbstractTo solve the label sparsity problem, domain adaptation has been well-established, suggesting various methods such as finding a common feature space of different domains using projection matrices or neural networks. Despite recent advances, domain ...
Graph Adaptive Knowledge Transfer for Unsupervised Domain Adaptation
Computer Vision – ECCV 2018AbstractUnsupervised domain adaptation has caught appealing attentions as it facilitates the unlabeled target learning by borrowing existing well-established source domain knowledge. Recent practice on domain adaptation manages to extract effective ...
Domain Adaptation Problems: A DASVM Classification Technique and a Circular Validation Strategy
This paper addresses pattern classification in the framework of domain adaptation by considering methods that solve problems in which training data are assumed to be available only for a source domain different (even if related) from the target domain ...
Comments