ABSTRACT
We extend tree-based methods to the prediction of structured outputs using a kernelization of the algorithm that allows one to grow trees as soon as a kernel can be defined on the output space. The resulting algorithm, called output kernel trees (OK3), generalizes classification and regression trees as well as tree-based ensemble methods in a principled way. It inherits several features of these methods such as interpretability, robustness to irrelevant variables, and input scalability. When only the Gram matrix over the outputs of the learning sample is given, it learns the output kernel as a function of inputs. We show that the proposed algorithm works well on an image reconstruction task and on a biological network inference problem.
- Blockeel, H., Bruynooghe, M., Dzeroski, S., Ramon, J., & Struyf, J. (2002). Hierarchical multi-classification. KDD-2002 Workshop Notes: MRDM 2002, Workshop on Multi-Relational Data Mining (pp. 21--35).]]Google Scholar
- Blockeel, H., De Raedt, L., & Ramon, J. (1998). Top-down induction of clustering trees. Proc. of ICML 1998 (pp. 55--63).]] Google ScholarDigital Library
- Breiman, L. (1996). Bagging predictors. Machine Learning, 24, 123--140.]] Google ScholarDigital Library
- Breiman, L., Friedman, J., Olsen, R., & Stone, C. (1984). Classification and regression trees. Wadsworth International (California).]]Google Scholar
- Cortes, C., Mehryar, M., & Weston, J. (2005). A general regression technique for learning transductions. Proceedings of ICML 2005 (pp. 153--160).]] Google ScholarDigital Library
- Dhillon, I., Guan, Y., & Kullis, B. (2004). Kernel k-means, spectral clustering and normalized cuts (Technical Report). University of Austin, Texas.]]Google Scholar
- Geurts, P., Ernst, D., & Wehenkel, L. (2006). Extremely randomized trees. Machine Learning Journal (advance access: DOI 10.1007/s10994-006-6226-1).]] Google ScholarDigital Library
- Kondor, R., & Lafferty, J. (2002). Diffusion kernels on graphs and other discrete input. Proc. of ICML 2002 (pp. 315--322).]] Google ScholarDigital Library
- Ralaivola, L., & d'Alchéé-Buc, F. (2003). Dynamical modeling with kernels for nonlinear time series prediction. Advances in Neural Information Processing Systems.]]Google Scholar
- Scholkopf, B., & Smola, A. (2002). Learning with kernels. MIT Press.]]Google Scholar
- Segal, M. (1992). Tree structured methods for longitudinal data. Journal of the American Statistical Association, 87, 407--418.]]Google ScholarCross Ref
- Taskar, B., Chatalbashev, V., Koller, D., & Guestrin, C. (2005). Learning structured prediction models: A large margin approach. Proc. of the 22nd International Conference on Machine Learning (pp. 897--904).]] Google ScholarDigital Library
- Todorovski, L., Blockeel, H., & Dzeroski, S. (2002). Ranking with predictive clustering trees. Proc. of the 13th European Conference on Machine Learning (pp. 444--456).]] Google ScholarDigital Library
- Tsochantaridis, I., Joachims, T., Hofmann, T., & Altun, Y. (2005). Large margin methods for structured and interdependent output variables. Journal of Machine Learning Research, 6, 1453--1484.]] Google ScholarDigital Library
- Tsuda, K., Akaho, S., & Asai, K. (2003). The em algorithm for kernel matrix completion with auxiliary data. Journal of machine learning research, 4, 67--81.]] Google ScholarDigital Library
- Vert, J.-P., & Yamanishi, Y. (2004). Supervised graph inference. Advances in Neural Information Processing Systems, 17, 1433--1440.]]Google Scholar
- Weston, J., Chapelle, O., Elisseeff, A., Schoelkopf, B., & Vapnik, V. (2002). Kernel dependency estimation. Advances in Neural Information Processing Systems, 15.]]Google Scholar
- Weston, J., Schoelkopf, B., & Bousquet, O. (2005). Joint kernel maps. Proceedings of the 8th International Work-Conference on Artificial Neural Networks (pp. 176--191).]] Google ScholarDigital Library
- Yamanishi, Y., Vert, J.-P., & Kanehisa, M. (2004). Protein network inference from multiple genomic data: a supervised approach. Bioinformatics, 20, i363--i370.]] Google ScholarDigital Library
- Yamanishi, Y., Vert, J.-P., & Kanehisa, M. (2005). Supervised enzyme network inference from the integration of genomic data and chemical information. Bioinformatics, 21, i468--i477.]] Google ScholarDigital Library
Index Terms
- Kernelizing the output of tree-based methods
Recommendations
Approximating the selected-internal Steiner tree
In this paper, we consider a variant of the well-known Steiner tree problem. Given a complete graph G=(V,E) with a cost function c:E->R^+ and two subsets R and R^' satisfying R^'@?R@?V, a selected-internal Steiner tree is a Steiner tree which contains (...
Computational Methods for Minimum Spanning Tree Algorithms
The well-known algorithms of Kruskal, Prim, and Sollin for constructing a minimum spanning tree are surveyed in this paper. We discuss various data structures that can facilitate the search and update operations of these algorithms. In particular, ...
Mutual information-based multi-output tree learning algorithm
A tree model with low time complexity can support the application of artificial intelligence to industrial systems. Variable selection based tree learning algorithms are more time efficient than existing Classification and Regression Tree (CART) ...
Comments