Prompting as Probing: Using Language Models for Knowledge Base Construction

Research output: Chapter in Book / Report / Conference proceedingConference contributionAcademicpeer-review

Abstract

Language Models (LMs) have proven to be useful in various downstream applications, such as sum-marisation, translation, question answering and text classification. LMs are becoming increasingly important tools in Artificial Intelligence, because of the vast quantity of information they can store. In this work, we present ProP (Prompting as Probing), which utilizes GPT-3, a large Language Model originally proposed by OpenAI in 2020, to perform the task of Knowledge Base Construction (KBC). ProP implements a multi-step approach that combines a variety of prompting techniques to achieve this. Our results show that manual prompt curation is essential, that the LM must be encouraged to give answer sets of variable lengths, in particular including empty answer sets, that true/false questions are a useful device to increase precision on suggestions generated by the LM, that the size of the LM is a crucial factor, and that a dictionary of entity aliases improves the LM score. Our evaluation study indicates that these proposed techniques can substantially enhance the quality of the final predictions: ProP won track 2 of the LM-KBC competition, outperforming the baseline by 36.4 percentage points. Our implementation is available on https://github.com/HEmile/iswc-challenge.

Original languageEnglish
Title of host publicationLM-KBC 2022 Knowledge Base Construction from Pre-trained Language Models 2022
Subtitle of host publicationProceedings of the Semantic Web Challenge on Knowledge Base Construction from Pre-trained Language Models 2022 co-located with the 21st International Semantic Web Conference (ISWC2022) Virtual Event, Hanghzou, China, October 2022
EditorsSneha Singhania, Tuan-Phong Nguyen, Simon Razniewski
PublisherCEUR-WS.org
Pages11-34
Number of pages24
Publication statusPublished - 2022
Event2022 Semantic Web Challenge on Knowledge Base Construction from Pre-Trained Language Models, LM-KBC 2022 - Virtual, Hanghzou, China
Duration: 1 Oct 2022 → …

Publication series

NameCEUR Workshop Proceedings
PublisherCEUR Workshop Proceedings
Volume3274
ISSN (Print)1613-0073

Conference

Conference2022 Semantic Web Challenge on Knowledge Base Construction from Pre-Trained Language Models, LM-KBC 2022
Country/TerritoryChina
CityVirtual, Hanghzou
Period1/10/22 → …

Bibliographical note

Funding Information:
We thank Frank van Harmelen for his insightful comments. This research was funded by the Vrije Universiteit Amsterdam and the Netherlands Organisation for Scientific Research (NWO) via the Spinoza grant (SPI 63-260) awarded to Piek Vossen, the Hybrid Intelligence Centre via the Zwaartekracht grant (024.004.022), Elsevier’s Discovery Lab, and Huawei’s DReaMS Lab.

Publisher Copyright:
© 2022 Copyright for this paper by its authors.

Fingerprint

Dive into the research topics of 'Prompting as Probing: Using Language Models for Knowledge Base Construction'. Together they form a unique fingerprint.

Cite this