We apologize for a recent technical issue with our email system, which temporarily affected account activations. Accounts have now been activated. Authors may proceed with paper submissions. PhDFocusTM
CFP last date
20 December 2024
Reseach Article

Study of K-NN Evaluation for Text Categorization using Multiple Level Learning

by Monika, Rajender Singh Chhillar
International Journal of Computer Applications
Foundation of Computer Science (FCS), NY, USA
Volume 122 - Number 22
Year of Publication: 2015
Authors: Monika, Rajender Singh Chhillar
10.5120/21855-5151

Monika, Rajender Singh Chhillar . Study of K-NN Evaluation for Text Categorization using Multiple Level Learning. International Journal of Computer Applications. 122, 22 ( July 2015), 9-12. DOI=10.5120/21855-5151

@article{ 10.5120/21855-5151,
author = { Monika, Rajender Singh Chhillar },
title = { Study of K-NN Evaluation for Text Categorization using Multiple Level Learning },
journal = { International Journal of Computer Applications },
issue_date = { July 2015 },
volume = { 122 },
number = { 22 },
month = { July },
year = { 2015 },
issn = { 0975-8887 },
pages = { 9-12 },
numpages = {9},
url = { https://ijcaonline.org/archives/volume122/number22/21855-5151/ },
doi = { 10.5120/21855-5151 },
publisher = {Foundation of Computer Science (FCS), NY, USA},
address = {New York, USA}
}
%0 Journal Article
%1 2024-02-06T23:11:12.750482+05:30
%A Monika
%A Rajender Singh Chhillar
%T Study of K-NN Evaluation for Text Categorization using Multiple Level Learning
%J International Journal of Computer Applications
%@ 0975-8887
%V 122
%N 22
%P 9-12
%D 2015
%I Foundation of Computer Science (FCS), NY, USA
Abstract

Predefined category exists for text categorization. In a document, text may be of any type category like government, education or health etc. many methods exist in market invented by researchers for text categorization. One of them is k-NN (k nearest neighbor) algorithm. k play a role to define number of classes for categorization. A training set is generated for each type of category to check its performance than whole text categorized. There is a problem of missing information during training sets. After study recent years invention on k-NN, we find out a solution of this problem. Multiple-Level Learning will improve the performance of k-NN. So in this paper we study about k-NN and propose hybrid algorithm with combination of Multiple-Level Learning and k-NN.

References
  1. E. Fix, and J. Hodges, "Discriminatory analysis. Nonparametric discrimination: Consistency properties". Technical Report 4, USAF School of Aviation Medicine, Randolph Field, Texas, 1951
  2. T. M. Cover, and P. E. Hart, "Nearest neighbor patternclassification", IEEE Transactions on Information Theory, 13, pp. 21–27, 1967
  3. R. O. Duda, and P. E. Hart, Pattern classification andscene analysis, New York: Wiley, 1973.
  4. W. Yu, and W. Zhengguo, "A fast kNN algorithm for textcategorization", Proceedings of the Sixth International Conference on Machine Learning and Cybernetics, Hong Kong, pp. 3436-3441, 2007
  5. W. Yi, B. Shi, and W. Zhang'ou, "A Fast KNN Algorithm Applied to Web Text Categorization", Journalof The China Society for Scientific and TechnicalInformation, 26(1), pp. 60-64, 2007.
  6. K. G. Anil, "On optimum choice of k in nearest neighborclassification", Computational Statistics and DataAnalysis, 50, pp. 3113–3123, 2006
  7. E. Kushilevitz, R. Ostrovsky, and Y. Rabani, "Efficientsearch for approximate nearest neighbor in highdimensional spaces". SIAM Journal on Computing, 30, pp. 457–474, 2000
  8. M. Lindenbaum, S. Markovitch, and D. Rusakov, "Selective sampling for nearest neighbor classifiers", Machine Learning, 54, pp. 125–152, 2004, IJCSI International Journal of Computer Science Issues, Vol. 7, Issue 4, No 2, July 2010 21www. IJCSI. org
  9. C. Zhou, Y. Yan, and Q. Chen, "Improving nearestneighbor classification with cam weighted distance". Pattern Recognition, 39, pp. 635–645, 2006
  10. D. P. Muni, and N. R. D. Pal, "Genetic programming forsimultaneous feature selection and classifier design",IEEE Transactions on Systems Man and Cybernetics PartB – Cybernetics, 36, pp. 106–117, 2006.
  11. J. Neumann, C. Schnorr, and G. Steidl, "Combined SVMbasedfeature selection and classification", Machine Learning, 61, pp. 129–150, 2005
  12. Tanya Taneja, Balraj Sharma, "Text Classification Using PSO & Other Technique" International Journal of Recent Development in Engineering and TechnologyWebsite: www. ijrdet. com (ISSN 2347-6435(Online) Volume 3, Issue 1, July 2014)
  13. Zhang W. , Yoshida T. , and Tang X," Text classification using multi-word features", In proceedings of the IEEE international conference on Systems, Man and Cybernetics, pp. 3519 – 3524, 2007.
  14. Witten, I. H. and Frank, E. "Data Mining: Practical Machine Learning Tools and Techniques", (2nd Edition), Morgan Kaufmann, San Francisco, 2005.
Index Terms

Computer Science
Information Sciences

Keywords

Data Mining Text Classification k-NN algorithm Multiple-Level Learning