|国家预印本平台
首页|Endowing Language Models with Multimodal Knowledge Graph Representations

Endowing Language Models with Multimodal Knowledge Graph Representations

Endowing Language Models with Multimodal Knowledge Graph Representations

来源:Arxiv_logoArxiv
英文摘要

We propose a method to make natural language understanding models more parameter efficient by storing knowledge in an external knowledge graph (KG) and retrieving from this KG using a dense index. Given (possibly multilingual) downstream task data, e.g., sentences in German, we retrieve entities from the KG and use their multimodal representations to improve downstream task performance. We use the recently released VisualSem KG as our external knowledge repository, which covers a subset of Wikipedia and WordNet entities, and compare a mix of tuple-based and graph-based algorithms to learn entity and relation representations that are grounded on the KG multimodal information. We demonstrate the usefulness of the learned entity representations on two downstream tasks, and show improved performance on the multilingual named entity recognition task by $0.3\%$--$0.7\%$ F1, while we achieve up to $2.5\%$ improvement in accuracy on the visual sense disambiguation task. All our code and data are available in: \url{https://github.com/iacercalixto/visualsem-kg}.

Kyunghyun Cho、Yibo Liu、Houda Alberts、Iacer Calixto、Ningyuan Huang、Yash R. Deshpande、Clara Vania

计算技术、计算机技术

Kyunghyun Cho,Yibo Liu,Houda Alberts,Iacer Calixto,Ningyuan Huang,Yash R. Deshpande,Clara Vania.Endowing Language Models with Multimodal Knowledge Graph Representations[EB/OL].(2022-06-27)[2025-08-02].https://arxiv.org/abs/2206.13163.点此复制

评论