SETRED: Self-training with editing

153Citations
Citations of this article
65Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Self-training is a semi-supervised learning algorithm in which a learner keeps on labeling unlabeled examples and retraining itself on an enlarged labeled training set. Since the self-training process may erroneously label some unlabeled examples, sometimes the learned hypothesis does not perform well. In this paper, a new algorithm named SETRED is proposed, which utilizes a specific data editing method to identify and remove the mislabeled examples from the self-labeled data. In detail, in each iteration of the self-training process, the local cut edge weight statistic is used to help estimate whether a newly labeled example is reliable or not, and only the reliable self-labeled examples are used to enlarge the labeled training set. Experiments show that the introduction of data editing is beneficial, and the learned hypotheses of SETRED outperform those learned by the standard self-training algorithm. © Springer-Verlag Berlin Heidelberg 2005.

Cite

CITATION STYLE

APA

Li, M., & Zhou, Z. H. (2005). SETRED: Self-training with editing. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 3518 LNAI, pp. 611–621). Springer Verlag. https://doi.org/10.1007/11430919_71

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free