Learning new facts from knowledge bases with neural tensor networks and semantic word vectors

D Chen, R Socher, CD Manning, AY Ng - arXiv preprint arXiv:1301.3618, 2013 - arxiv.org
arXiv preprint arXiv:1301.3618, 2013arxiv.org
Knowledge bases provide applications with the benefit of easily accessible, systematic
relational knowledge but often suffer in practice from their incompleteness and lack of
knowledge of new entities and relations. Much work has focused on building or extending
them by finding patterns in large unannotated text corpora. In contrast, here we mainly aim to
complete a knowledge base by predicting additional true relationships between entities,
based on generalizations that can be discerned in the given knowledgebase. We introduce …
Knowledge bases provide applications with the benefit of easily accessible, systematic relational knowledge but often suffer in practice from their incompleteness and lack of knowledge of new entities and relations. Much work has focused on building or extending them by finding patterns in large unannotated text corpora. In contrast, here we mainly aim to complete a knowledge base by predicting additional true relationships between entities, based on generalizations that can be discerned in the given knowledgebase. We introduce a neural tensor network (NTN) model which predicts new relationship entries that can be added to the database. This model can be improved by initializing entity representations with word vectors learned in an unsupervised fashion from text, and when doing this, existing relations can even be queried for entities that were not present in the database. Our model generalizes and outperforms existing models for this problem, and can classify unseen relationships in WordNet with an accuracy of 75.8%.
arxiv.org
Showing the best result for this search. See all results