Active and semisupervised learning are related techniques aiming at reducing the effort of creating training sets for classification and regression tasks. In this work, we present a framework for combining these two techniques on the basis of Renyi entropy regularization, enabling a synergy effect. We build upon the existing semisupervised learning model which attempts to balance the likelihood of labeled examples and the entropy of putative object probabilities within the unlabeled pool. To enable efficient optimization of the model, we generalize the deterministic annealing expectation-maximization (DAEM) algorithm, originally designed for Shannon entropy, to accommodate the use of Renyi entropies. The Renyi-regularized model is then applied to expected error reduction (EER), an active learning approach based on minimizing the entropy of unlabeled object probabilities. We investigate object preselection with a greedy approximation of the object feature matrix as a means to reduce computational complexity. To assess the performance of the proposed framework, we apply it to two real-world remote sensing problems with significantly different input data characteristics: detecting dead trees from color infrared aerial images (2-D) and detecting dead trunk stems in ALS point clouds (3-D). Our results show that for small training sets, the semisupervised Renyi-regularized classifier improves the classification rate by up to 11% and 10% points compared to the unregularized baseline for ALS and image data, respectively. This gain carries over to active learning, where the regularized EER achieves 90% of the final classification performance using 50% and 70% of the number of queries required by standard EER.
«
Active and semisupervised learning are related techniques aiming at reducing the effort of creating training sets for classification and regression tasks. In this work, we present a framework for combining these two techniques on the basis of Renyi entropy regularization, enabling a synergy effect. We build upon the existing semisupervised learning model which attempts to balance the likelihood of labeled examples and the entropy of putative object probabilities within the unlabeled pool. To ena...
»