0


模型建筑应该反映语言结构吗?

Should Model Architecture Reflect Linguistic Structure?
课程网址: http://videolectures.net/iclr2016_dyer_model_architecture/  
主讲教师: Chris Dyer
开课单位: 卡内基梅隆大学
开课时间: 2016-05-27
课程语种: 英语
中文简介:
有限字母表上的序列递归神经网络是一种非常有效的自然语言模型。RNN现在获得的语言建模结果大大改善了长期以来最先进的基线,以及各种条件语言建模任务,如机器翻译、图片说明生成和对话生成。尽管这些令人印象深刻的结果,这样的模型是一个先天不合适的语言模型。批评的一点是,语言使用者一直在创造和理解新词,挑战了有限词汇假设。第二,单词之间的关系是根据潜在的嵌套结构而不是按顺序的表面顺序来计算的(Chomsky,1957;Everaert、Huybregts、Chomsky、Berwick和Bolhuis,2015)。 在这篇演讲中,我将讨论两个模型,这两个模型探讨了一个假设,即更多(先验)合适的语言模型将导致在现实世界中的语言处理任务中取得更好的表现。第一种方法将子词单元(字节、字符或语素)组合成词汇表示,从而能够更自然地解释和生成新的单词形式。第二种,我们称之为递归神经网络语法(RNNGs),是一种新的句子生成模型,它显式地为单词和短语之间的嵌套层次关系建模。RNNGs通过一个递归的语法过程来操作,这让人联想到概率上下文无关语法生成,但是使用rnn来参数化决策,该过程以整个(自上而下,从左到右)语法派生历史为条件,大大放宽了上下文无关的独立性假设。实验结果表明,与不利用语言结构的模型相比,RNNGs在生成语言方面取得了更好的效果。
课程简介: Sequential recurrent neural networks (RNNs) over finite alphabets are remarkably effective models of natural language. RNNs now obtain language modeling results that substantially improve over long-standing state-of-the-art baselines, as well as in various conditional language modeling tasks such as machine translation, image caption generation, and dialogue generation. Despite these impressive results, such models are a priori inappropriate models of language. One point of criticism is that language users create and understand new words all the time, challenging the finite vocabulary assumption. A second is that relationships among words are computed in terms of latent nested structures rather than sequential surface order (Chomsky, 1957; Everaert, Huybregts, Chomsky, Berwick, and Bolhuis, 2015). In this talk I discuss two models that explore the hypothesis that more (a priori) appropriate models of language will lead to better performance on real-world language processing tasks. The first composes sub word units (bytes, characters, or morphemes) into lexical representations, enabling more naturalistic interpretation and generation of novel word forms. The second, which we call recurrent neural network grammars (RNNGs), is a new generative model of sentences that explicitly models nested, hierarchical relationships among words and phrases. RNNGs operate via a recursive syntactic process reminiscent of probabilistic context-free grammar generation, but decisions are parameterized using RNNs that condition on the entire (top-down, left-to-right) syntactic derivation history, greatly relaxing context-free independence assumptions. Experimental results show that RNNGs obtain better results in generating language than models that don’t exploit linguistic structures.
关 键 词: 模型建筑; 语言结构
课程来源: 视频讲座网
数据采集: 2020-11-29:yxd
最后编审: 2020-11-29:yxd
阅读次数: 22