Show simple item record

dc.contributor.authorDong, H
dc.contributor.authorChen, J
dc.contributor.authorHe, Y
dc.contributor.authorGao, Y
dc.contributor.authorHorrocks, I
dc.date.accessioned2024-03-04T15:29:02Z
dc.date.issued2024-05-19
dc.date.updated2024-03-04T14:32:36Z
dc.description.abstractWe investigate the task of inserting new concepts extracted from texts into an ontology using language models. We explore an approach with three steps: edge search which is to find a set of candidate locations to insert (i.e., subsumptions between concepts), edge formation and enrichment which leverages the ontological structure to produce and enhance the edge candidates, and edge selection which eventually locates the edge to be placed into. In all steps, we propose to leverage neural methods, where we apply embedding-based methods and contrastive learning with Pre-trained Language Models (PLMs) such as BERT for edge search, and adapt a BERT fine-tuning-based multi-label Edge-Cross-encoder, and Large Language Models (LLMs) such as GPT series, FLAN-T5, and Llama 2, for edge selection. We evaluate the methods on recent datasets created using the SNOMED CT ontology and the MedMentions entity linking benchmark. The best settings in our framework use fine-tuned PLM for search and a multi-label Cross-encoder for selection. Zero-shot prompting of LLMs is still not adequate for the task, and we propose explainable instruction tuning of LLMs for improved performance. Our study shows the advantages of PLMs and highlights the encouraging performance of LLMs that motivates future studies.en_GB
dc.description.sponsorshipEngineering and Physical Sciences Research Council (EPSRC)en_GB
dc.description.sponsorshipSamsung Research UK (SRUK)en_GB
dc.identifier.citationIn: The Semantic Web: 21st International Conference, ESWC 2024, Hersonissos, Crete, Greece, 26 - 30 May 2024. Proceedings, Part I, edited by Albert Meroño Peñuela, Anastasia Dimou, Raphaël Troncy, Olaf Hartig, Maribel Acosta, Mehwish Alam, Heiko Paulheim, and Pasquale Lisena, pp. 79 - 99. Lecture Notes in Computer Science, vol. 14664en_GB
dc.identifier.doi10.1007/978-3-031-60626-7_5
dc.identifier.grantnumberEP/V050869/1en_GB
dc.identifier.grantnumberEP/S032347/1en_GB
dc.identifier.grantnumberEP/S019111/1en_GB
dc.identifier.urihttp://hdl.handle.net/10871/135471
dc.identifierORCID: 0000-0001-6828-6891 (Dong, Hang)
dc.language.isoenen_GB
dc.publisherSpringeren_GB
dc.rights.embargoreasonUnder embargo until 19 May 2025 in compliance with publisher policyen_GB
dc.rights© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG.
dc.subjectOntology Enrichmenten_GB
dc.subjectConcept Placementen_GB
dc.subjectPre-trained Language Modelsen_GB
dc.subjectLarge Language Modelsen_GB
dc.subjectSNOMED CTen_GB
dc.titleA Language Model based Framework for New Concept Placement in Ontologiesen_GB
dc.typeConference paperen_GB
dc.date.available2024-03-04T15:29:02Z
dc.descriptionThis is the author accepted manuscript. The final version is available from Springer via the DOI in this recorden_GB
dc.rights.urihttp://www.rioxx.net/licenses/all-rights-reserveden_GB
rioxxterms.versionAMen_GB
rioxxterms.licenseref.startdate2024-03-04
rioxxterms.typeConference Paper/Proceeding/Abstracten_GB
refterms.dateFCD2024-03-04T14:32:39Z
refterms.versionFCDAM
refterms.panelBen_GB
pubs.name-of-conferenceExtended Semantic Web Conference


Files in this item

This item appears in the following Collection(s)

Show simple item record