tailieunhanh - Báo cáo khoa học: "Improving Word Representations via Global Context and Multiple Word Prototypes"

Unsupervised word representations are very useful in NLP tasks both as inputs to learning algorithms and as extra word features in NLP systems. However, most of these models are built with only local context and one representation per word. This is problematic because words are often polysemous and global context can also provide useful information for learning word meanings. We present a new neural network architecture which 1) learns word embeddings that better capture the semantics of words by incorporating both local and global document context, and 2) accounts for homonymy and polysemy by learning multiple embeddings per word | Improving Word Representations via Global Context and Multiple Word Prototypes Eric H. Huang Richard Socher Christopher D. Manning Andrew Y. Ng Computer Science Department Stanford University Stanford CA 94305 USA ehhuang manning ang @ richard@ Abstract Unsupervised word representations are very useful in NLP tasks both as inputs to learning algorithms and as extra word features in NLP systems. However most of these models are built with only local context and one representation per word. This is problematic because words are often polysemous and global context can also provide useful information for learning word meanings. We present a new neural network architecture which 1 learns word embeddings that better capture the semantics of words by incorporating both local and global document context and 2 accounts for homonymy and polysemy by learning multiple embeddings per word. We introduce a new dataset with human judgments on pairs of words in sentential context and evaluate our model on it showing that our model outperforms competitive baselines and other neural language models. 1 1 Introduction Vector-space models VSM represent word meanings with vectors that capture semantic and syntactic information of words. These representations can be used to induce similarity measures by computing distances between the vectors leading to many useful applications such as information retrieval Manning et al. 2008 document classification Sebas-tiani 2002 and question answering Tellex et al. 2003 . The dataset and word vectors can be downloaded at http ehhuang . 873 Despite their usefulness most VSMs share a common problem that each word is only represented with one vector which clearly fails to capture homonymy and polysemy. Reisinger and Mooney 2010b introduced a multi-prototype VSM where word sense discrimination is first applied by clustering contexts and then prototypes are built using the contexts of the sense-labeled words. However

TỪ KHÓA LIÊN QUAN
crossorigin="anonymous">
Đã phát hiện trình chặn quảng cáo AdBlock
Trang web này phụ thuộc vào doanh thu từ số lần hiển thị quảng cáo để tồn tại. Vui lòng tắt trình chặn quảng cáo của bạn hoặc tạm dừng tính năng chặn quảng cáo cho trang web này.