论文标题
微观监督干扰学习:表示概率分布的观点
Micro-supervised Disturbance Learning: A Perspective of Representation Probability Distribution
论文作者
论文摘要
在广泛的条件下,基于欧几里得距离的现有表示学习方法显示了不稳定性。此外,标签的稀缺性和高成本促使我们探索更具表现力的学习方法,这些学习方法取决于标签尽可能少。为了解决这些问题,首先基于表示概率分布在表示模型上引入小型扰动意识形态。仅取决于每个群集的两个标签的积极小扰动信息(SPI)用于刺激表示概率分布,然后提出了两个变体模型,以微调RBM的预期表示分布,即,微观监督干扰的干扰GRBM(Micro-DGRBM)(Micro-DGRBM)(Micro-DGRBM)和微渗透性扰动干扰RBM(micm micm mictro-drbm)模型。 SPI的Kullback-Leibler(KL)差异在同一群集中被最小化,以促进表示概率分布,以在对比性差异(CD)学习中变得更加相似。相比之下,在不同的簇中最大化SPI的KL差异,以强制执行表示概率分布,从而在CD学习中变得更加不同。为了在SPI的连续刺激下探索表示能力,我们提出了基于Micro-DGRBM和Micro-DRBM模型的深度小监督干扰学习(Micro-DL)框架,并将其与没有任何外部刺激的类似深层结构进行比较。实验结果表明,与基线方法相比,所提出的深层微型DL架构显示出更好的性能,即最相关的浅层模型和聚类的深框架。
The instability is shown in the existing methods of representation learning based on Euclidean distance under a broad set of conditions. Furthermore, the scarcity and high cost of labels prompt us to explore more expressive representation learning methods which depends on the labels as few as possible. To address these issues, the small-perturbation ideology is firstly introduced on the representation learning model based on the representation probability distribution. The positive small-perturbation information (SPI) which only depend on two labels of each cluster is used to stimulate the representation probability distribution and then two variant models are proposed to fine-tune the expected representation distribution of RBM, namely, Micro-supervised Disturbance GRBM (Micro-DGRBM) and Micro-supervised Disturbance RBM (Micro-DRBM) models. The Kullback-Leibler (KL) divergence of SPI is minimized in the same cluster to promote the representation probability distributions to become more similar in Contrastive Divergence(CD) learning. In contrast, the KL divergence of SPI is maximized in the different clusters to enforce the representation probability distributions to become more dissimilar in CD learning. To explore the representation learning capability under the continuous stimulation of the SPI, we present a deep Micro-supervised Disturbance Learning (Micro-DL) framework based on the Micro-DGRBM and Micro-DRBM models and compare it with a similar deep structure which has not any external stimulation. Experimental results demonstrate that the proposed deep Micro-DL architecture shows better performance in comparison to the baseline method, the most related shallow models and deep frameworks for clustering.