0


用于学习隐藏变量模型的可扩展张量分解

Scalable Tensor Decompositions for Learning Hidden Variable Models
课程网址: http://videolectures.net/onlinelearning2012_kakade_hidden_variabl...  
主讲教师: Sham M. Kakade
开课单位: 微软公司
开课时间: 2013-05-28
课程语种: 英语
中文简介:
在许多应用程序中,我们面临着对多个之间的交互进行建模的挑战观察结果。机器学习和人工智能的一种流行和成功的方法是假设某些潜在(或隐藏)原因的存在有助于解释其中的相关性观察数据。 (无监督)学习问题是仅用精确估计模型观察数据的样本。例如,在文档建模中,我们可能希望进行特征化文档中“词袋”的相关结构。这里,标准模型是假定的该文档涉及几个主题(隐藏变量)和每个活动主题确定文档中单词的出现次数。学习问题是,只使用在文档中观察到的单词(而不是隐藏的主题),来估计主题概率向量(即发现单词倾向于出现在不同topcis下的强度)。在实践中,广泛的潜变量模型通常适用于局部搜索启发式算法(例如EM算法)或基于采样的方法。本演讲将讨论标准线性代数工具的一般化(例如光谱方法)对于各种潜变量模型,张量为张量提供了可证明有效的估计方法(在适当的假设下),包括高斯模型的混合,隐马尔可夫模型,主题模型,潜在Dirichlet分配,潜在解析树模型(PCFG和依赖解析器)和社交网络中社区的模型。谈话也将简要介绍讨论矩阵和张量分解方法如何用于结构学习,确定某些隐藏原因和底层图形的存在的问题,这些隐藏原因与观察到的变量之间的结构。
课程简介: In many applications, we face the challenge of modeling the interactions between multiple observations. A popular and successful approach in machine learning and AI is to hypothesize the existence of certain latent (or hidden) causes which help to explain the correlations in the observed data. The (unsupervised) learning problem is to accurately estimate a model with only samples of the observed data. For example, in document modeling, we may wish to characterize the correlational structure of the "bag of words" in documents. Here, a standard model is to posit that documents are about a few topics (the hidden variables) and that each active topic determines the occurrence of words in the document. The learning problem is, using only the observed words in the documents (and not the hidden topics), to estimate the topic probability vectors (i.e. discover the strength by which words tend to appear under different topcis). In practice, a broad class of latent variable models is most often fit with either local search heuristics (such as the EM algorithm) or sampling based approaches. This talk will discuss how generalizations of standard linear algebra tools (e.g. spectral methods) to tensors provide provable and efficient estimation methods for various latent variable models (under appropriate assumptions), including mixtures of Gaussians models, hidden Markov models, topic models, latent Dirichlet allocation, latent parse tree models (PCFGs and dependency parsers), and models for communities in social networks. The talk will also briefly discuss how matrix and tensor decomposition methods can be used for the structure learning problem of determining both the existence of certain hidden causes and the underlying graphical structure between these hidden causes and the observed variables.
关 键 词: 机器学习; 人工智能; 概率向量
课程来源: 视频讲座网
最后编审: 2020-04-30:chenxin
阅读次数: 38