## Abstract and Applied Analysis

### A Fast Logdet Divergence Based Metric Learning Algorithm for Large Data Sets Classification

#### Abstract

Large data sets classification is widely used in many industrial applications. It is a challenging task to classify large data sets efficiently, accurately, and robustly, as large data sets always contain numerous instances with high dimensional feature space. In order to deal with this problem, in this paper we present an online Logdet divergence based metric learning (LDML) model by making use of the powerfulness of metric learning. We firstly generate a Mahalanobis matrix via learning the training data with LDML model. Meanwhile, we propose a compressed representation for high dimensional Mahalanobis matrix to reduce the computation complexity in each iteration. The final Mahalanobis matrix obtained this way measures the distances between instances accurately and serves as the basis of classifiers, for example, the $k$-nearest neighbors classifier. Experiments on benchmark data sets demonstrate that the proposed algorithm compares favorably with the state-of-the-art methods.

#### Article information

Source
Abstr. Appl. Anal., Volume 2014 (2014), Article ID 463981, 9 pages.

Dates
First available in Project Euclid: 6 October 2014

https://projecteuclid.org/euclid.aaa/1412606636

Digital Object Identifier
doi:10.1155/2014/463981

Zentralblatt MATH identifier
07022432

#### Citation

Mei, Jiangyuan; Hou, Jian; Chen, Jicheng; Karimi, Hamid Reza. A Fast Logdet Divergence Based Metric Learning Algorithm for Large Data Sets Classification. Abstr. Appl. Anal. 2014 (2014), Article ID 463981, 9 pages. doi:10.1155/2014/463981. https://projecteuclid.org/euclid.aaa/1412606636

#### References

• S. Yin, S. X. Ding, A. Haghani, H. Hao, and P. Zhang, “A comparison study of basic data-driven fault diagnosis and process monitoring methods on the benchmark tennessee eastman process,” Journal of Process Control, vol. 22, no. 9, pp. 1567–1581, 2012.
• S. Yin, G. Wang, and H. R. Karimi, “Data-driven design of robust fault detection system for wind turbines,” Mechatronics, 2013.
• S. Yin, S. X. Ding, A. H. Abandan Sari, and H. Hao, “Data-driven monitoring for stochastic systems and its application on batch process,” International Journal of Systems Science, vol. 44, no. 7, pp. 1366–1376, 2013.
• H. Gao, C. Ding, C. Song, and J. Mei, “Automated inspection of e-shaped magnetic core elements using k-tsl-center clustering and active shape models,” IEEE Transactions on Industrial Informatics, vol. 9, no. 3, pp. 1782–1789, 2013.
• M. Guillaumin, J. Verbeek, and C. Schmid, “Is that you? Metric learning approaches for face identification,” in Proceedings of the IEEE 12th International Conference on Computer Vision (ICCV '09), pp. 498–505, October 2009.
• E. P. Xing, M. I. Jordan, S. Russell, and A. Ng, “Distance metric learning with application to clustering with side-information,” in Advances in Neural Information Processing Systems, pp. 505–512, 2002.
• J. V. Davis, B. Kulis, P. Jain, S. Sra, and I. S. Dhillon, “Information-theoretic metric learning,” in Proceedings of the ACM 24th International Conference on Machine Learning (ICML '07), pp. 209–216, June 2007.
• P. Jain, B. Kulis, J. V. Davis, and I. S. Dhillon, “Metric and kernel learning using a linear transformation,” Journal of Machine Learning Research, vol. 13, pp. 519–547, 2012.
• K. Weinberger, J. Blitzer, and L. Saul, “Distance metric learning for large margin nearest neighbor classification,” in Advances in Neural Information Processing Systems, vol. 18, p. 1473, 2006.
• S. Yin, H. Luo, and S. Ding, “Real-time implementation of fault-tolerant control systems with performance optimization,” IEEE Transactions on Industrial Electronics, vol. 61, no. 5, pp. 2402–2411, 2012.
• J. V. Davis and I. S. Dhillon, “Differential entropic clustering of multivariate gaussians,” in Advances in Neural Information Processing, pp. 337–344, 2006.
• S. Yin, X. Yang, and H. R. Karimi, “Data-driven adaptive observer for fault diagnosis,” Mathematical Problems in Engineering, vol. 2012, Article ID 832836, 21 pages, 2012.
• S. Shalev-Shwartz, Y. Singer, and A. Y. Ng, “Online and batch learning of pseudo-metrics,” in Proceedings of the ACM 21st International Conference on Machine Learning (ICML '04), p. 94, July 2004.
• P. Jain, B. Kulis, I. S. Dhillon, and K. Grauman, “Online metric learning and fast similarity search,” in Proceedings of the 22nd Annual Conference on Neural Information Processing Systems (NIPS '08), pp. 761–768, December 2008.
• B. Kulis, M. Sustik, and I. Dhillon, “Learning low-rank kernel matrices,” in Proceedings of the 23rd International Conference on Machine Learning (ICML '06), pp. 505–512, June 2006.
• M. Sugiyama, “Dimensionality reduction of multimodal labeled data by local fisher discriminant analysis,” Journal of Machine Learning Research, vol. 8, pp. 1027–1061, 2007.
• Q. Wang, P. C. Yuen, and G. Feng, “Semi-supervised metric learning via topology preserving multiple semi-supervised assumptions,” Pattern Recognition, vol. 46, no. 9, pp. 2576–2587, 2013.
• J. V. Davis and I. S. Dhillon, “Structured metric learning for high dimensional problems,” in Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD '08), pp. 195–203, August 2008.
• H. Zha and H. D. Simon, “On updating problems in latent semantic indexing,” SIAM Journal on Scientific Computing, vol. 21, no. 2, pp. 782–791, 1999.
• H. Zhao, P. C. Yuen, and J. T. Kwok, “A novel incremental principal component analysis and its application for face recognition,” IEEE Transactions on Systems, Man, and Cybernetics B: Cybernetics, vol. 36, no. 4, pp. 873–886, 2006.
• J. Mei, M. Liu, H. R. Karimi, and H. Gao, “Logdet divergence based metric learning using triplet labels,” in Proceedings of the Workshop on Divergences and Divergence Learning ( ICML '13), 2013. \endinput