Skip to content

Commit

Permalink
Paper Revision{2023.findings-acl.275}, closes #4190.
Browse files Browse the repository at this point in the history
  • Loading branch information
anthology-assist committed Jan 13, 2025
1 parent 7021a1b commit af120c1
Showing 1 changed file with 3 additions and 1 deletion.
4 changes: 3 additions & 1 deletion data/xml/2023.findings.xml
Original file line number Diff line number Diff line change
Expand Up @@ -6358,9 +6358,11 @@
<author><first>Pascale</first><last>Fung</last><affiliation>Hong Kong University of Science and Technology</affiliation></author>
<pages>4504-4522</pages>
<abstract>Dialogue systems can leverage large pre-trained language models and knowledge to generate fluent and informative responses. However, these models are still prone to produce hallucinated responses not supported by the input source, which greatly hinders their application. The heterogeneity between external knowledge and dialogue context challenges representation learning and source integration, which further contributes to unfaithfulness. To handle this challenge and generate more faithful responses, this paper presents RHO (ρ) utilizing the representations of linked entities and relation predicates from a knowledge graph (KG). We propose (1) local knowledge grounding to combine textual embeddings with the corresponding KG embeddings; and (2) global knowledge grounding to equip RHO with multi-hop reasoning abilities via the attention mechanism. In addition, we devise a response re-ranking technique based on walks over KG sub-graphs for better conversational reasoning. Experimental results on OpenDialKG (Moon et al., 2019) show that our approach significantly outperforms state-of-the-art methods on both automatic and human evaluation by a large margin, especially in hallucination reduction (17.54% in FeQA (Durmus et al., 2020)).</abstract>
<url hash="30f72dcd">2023.findings-acl.275</url>
<url hash="8f36fcb1">2023.findings-acl.275</url>
<bibkey>ji-etal-2023-rho</bibkey>
<doi>10.18653/v1/2023.findings-acl.275</doi>
<revision id="1" href="2023.findings-acl.275v1" hash="30f72dcd"/>
<revision id="2" href="2023.findings-acl.275v2" hash="8f36fcb1" date="2025-01-12">Minor updates.</revision>
</paper>
<paper id="276">
<title>Transformer Language Models Handle Word Frequency in Prediction Head</title>
Expand Down

0 comments on commit af120c1

Please sign in to comment.