0


字典学习的样本复杂性

The Sample Complexity of Dictionary Learning
课程网址: http://videolectures.net/colt2011_vainsencher_dictionary/  
主讲教师: Daniel Vainsencher
开课单位: 以色列理工学院
开课时间: 2011-08-02
课程语种: 英语
中文简介:
有时可以使用字典稀疏地描述大量信号,即,每个元素可以表示为来自字典的少数元素的线性组合。用于各种信号处理应用的算法,包括分类,去噪和信号分离,从要表示的给定信号集学习字典。我们可以期望这样的字典表示来自同一来源的先前看不见的信号的误差将与给定示例的那些相同吗?我们假设信号是从固定分布产生的,并从统计学习理论的角度研究这些问题。我们根据系数选择的两种约束条件,对学习词典的质量进行泛化界定,如表示的预期L2误差所测量的那样。使用字典。对于l1正则化系数选择的情况,我们提供O的阶数的泛化边界,其中n是维度,p是字典中元素的数量,λ是系数向量的l1范数的界限,m是数字样本,补充了现有结果。对于将新信号表示为最多k个字典元素的组合的情况,我们在关于字典的正交性(低巴贝尔函数)的假设下提供阶数O的界限。我们进一步表明,这种假设适用于大多数高维词典中的强大概率论。我们的结果还包括收敛为1 = m的边界,这是以前不知道的问题。我们使用具有弱平滑度要求的内核在一般设置中提供类似的结果。
课程简介: A large set of signals can sometimes be described sparsely using a dictionary, that is, every element can be represented as a linear combination of few elements from the dictionary. Algorithms for various signal processing applications, including classification, denoising and signal separation, learn a dictionary from a given set of signals to be represented. Can we expect that the error in representing by such a dictionary a previously unseen signal from the same source will be of similar magnitude as those for the given examples? We assume signals are generated from a fixed distribution, and study these questions from a statistical learning theory perspective. We develop generalization bounds on the quality of the learned dictionary for two types of constraints on the coefficient selection, as measured by the expected L2 error in representation when the dictionary is used. For the case of l1 regularized coefficient selection we provide a generalization bound of the order of O, where n is the dimension, p is the number of elements in the dictionary, λ is a bound on the l1 norm of the coefficient vector and m is the number of samples, which complements existing results. For the case of representing a new signal as a combination of at most k dictionary elements, we provide a bound of the order O under an assumption on the closeness to orthogonality of the dictionary (low Babel function). We further show that this assumption holds for most dictionaries in high dimensions in a strong probabilistic sense. Our results also include bounds that converge as 1=m, not previously known for this problem. We provide similar results in a general setting using kernels with weak smoothness requirements.
关 键 词: 字典; 元素
课程来源: 视频讲座网
最后编审: 2021-02-10:nkq
阅读次数: 119