TY - GEN
T1 - Term Definitions Help Hypernymy Detection
AU - Yin, Wenpeng
AU - Roth, Dan
N1 - Funding Information:
We thank all the reviewers for providing insightful comments and critiques. This research is supported in part by DARPA under agreement number FA8750-13-2-0008, and by a gift from Google.
Publisher Copyright:
© 2018 Association for Computational Linguistics.
PY - 2018
Y1 - 2018
N2 - Existing methods of hypernymy detection mainly rely on statistics over a big corpus, either mining some co-occurring patterns like “animals such as cats” or embedding words of interest into context-aware vectors. These approaches are therefore limited by the availability of a large enough corpus that can cover all terms of interest and provide sufficient contextual information to represent their meaning. In this work, we propose a new paradigm, HYPERDEF, for hypernymy detection – expressing word meaning by encoding word definitions, along with context driven representation. This has two main benefits: (i) Definitional sentences express (sense-specific) corpus-independent meanings of words, hence definition-driven approaches enable strong generalization – once trained, the model is expected to work well in open-domain testbeds; (ii) Global context from a large corpus and definitions provide complementary information for words.
AB - Existing methods of hypernymy detection mainly rely on statistics over a big corpus, either mining some co-occurring patterns like “animals such as cats” or embedding words of interest into context-aware vectors. These approaches are therefore limited by the availability of a large enough corpus that can cover all terms of interest and provide sufficient contextual information to represent their meaning. In this work, we propose a new paradigm, HYPERDEF, for hypernymy detection – expressing word meaning by encoding word definitions, along with context driven representation. This has two main benefits: (i) Definitional sentences express (sense-specific) corpus-independent meanings of words, hence definition-driven approaches enable strong generalization – once trained, the model is expected to work well in open-domain testbeds; (ii) Global context from a large corpus and definitions provide complementary information for words.
UR - http://www.scopus.com/inward/record.url?scp=85084295148&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85084295148&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85084295148
T3 - NAACL HLT 2018 - Lexical and Computational Semantics, SEM 2018, Proceedings of the 7th Conference
SP - 203
EP - 213
BT - NAACL HLT 2018 - Lexical and Computational Semantics, SEM 2018, Proceedings of the 7th Conference
A2 - Nissim, Malvina
A2 - Berant, Jonathan
A2 - Lenci, Alessandro
PB - Association for Computational Linguistics (ACL)
T2 - 7th Joint Conference on Lexical and Computational Semantics, SEM 2018, co-located with NAACL HLT 2018
Y2 - 5 June 2018 through 6 June 2018
ER -