Mitigating spatial hallucination in large language models for path planning via prompt engineering

被引:0
|
作者
Zhang, Hongjie [1 ]
Deng, Hourui [1 ]
Ou, Jie [2 ]
Feng, Chaosheng [1 ]
机构
[1] Sichuan Normal Univ, Coll Comp Sci, Chengdu 610101, Peoples R China
[2] Univ Elect Sci & Technol China, Sch Informat & Software Engn, Chengdu 611731, Peoples R China
来源
SCIENTIFIC REPORTS | 2025年 / 15卷 / 01期
关键词
D O I
10.1038/s41598-025-93601-5
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Spatial reasoning in Large Language Models (LLMs) serves as a foundation for embodied intelligence. However, even in simple maze environments, LLMs often struggle to plan correct paths due to hallucination issues. To address this, we propose S2ERS, an LLM-based technique that integrates entity and relation extraction with the on-policy reinforcement learning algorithm Sarsa for optimal path planning. We introduce three key improvements: (1) To tackle the hallucination of spatial, we extract a graph structure of entities and relations from the text-based maze description, aiding LLMs in accurately comprehending spatial relationships. (2) To prevent LLMs from getting trapped in dead ends due to context inconsistency hallucination by long-term reasoning, we insert the state-action value function Q into the prompts, guiding the LLM's path planning. (3) To reduce the token consumption of LLMs, we utilize multi-step reasoning, dynamically inserting local Q-tables into the prompt to assist the LLM in outputting multiple steps of actions at once. Our comprehensive experimental evaluation, conducted using closed-source LLMs ChatGPT 3.5, ERNIE-Bot 4.0 and open-source LLM ChatGLM-6B, demonstrates that S2ERS significantly mitigates the spatial hallucination issues in LLMs, and improves the success rate and optimal rate by approximately 29% and 19%, respectively, in comparison to the SOTA CoT methods.
引用
收藏
页数:13
相关论文
共 50 条
  • [1] Towards Mitigating Hallucination in Large Language Models via Self-Reflection
    Ji, Ziwei
    Yu, Tiezheng
    Xu, Yan
    Lee, Nayeon
    Ishii, Etsuko
    Fung, Pascale
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 1827 - 1843
  • [2] Mitigating Factual Inconsistency and Hallucination in Large Language Models
    Muneeswaran, I
    Shankar, Advaith
    Varun, V.
    Gopalakrishnan, Saisubramaniam
    Vaddina, Vishal
    PROCEEDINGS OF THE 17TH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING, WSDM 2024, 2024, : 1169 - 1170
  • [3] Integrating chemistry knowledge in large language models via prompt engineering
    Liu, Hongxuan
    Yin, Haoyu
    Luo, Zhiyao
    Wang, Xiaonan
    SYNTHETIC AND SYSTEMS BIOTECHNOLOGY, 2025, 10 (01) : 23 - 38
  • [4] WORDFLOW: Social Prompt Engineering for Large Language Models
    Wang, Zijie J.
    Chakravarthy, Aishwarya
    Munechika, David
    Chau, Duen Horng
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 3: SYSTEM DEMONSTRATIONS, 2024, : 42 - 50
  • [5] Improving large language models for clinical named entity recognition via prompt engineering
    Hu, Yan
    Chen, Qingyu
    Du, Jingcheng
    Peng, Xueqing
    Keloth, Vipina Kuttichi
    Zuo, Xu
    Zhou, Yujia
    Li, Zehan
    Jiang, Xiaoqian
    Lu, Zhiyong
    Roberts, Kirk
    Xu, Hua
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2024, 31 (09)
  • [6] Mitigating Value Hallucination in Dyna-Style Planning via Multistep Predecessor Models
    Aminmansour F.
    Jafferjee T.
    Imani E.
    Talvitie E.J.
    Bowling M.
    White M.
    Journal of Artificial Intelligence Research, 2024, 80 : 441 - 473
  • [7] Mitigating Value Hallucination in Dyna-Style Planning via Multistep Predecessor Models
    Aminmansour, Farzane
    Jafferjee, Taher
    Imani, Ehsan
    Talvitie, Erin J.
    Bowling, Michael
    White, Martha
    JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2024, 80 : 441 - 473
  • [8] Prompt Engineering: Guiding the Way to Effective Large Language Models
    Aljanabi M.
    Yaseen M.G.
    Ali A.H.
    Mohammed M.A.
    Iraqi Journal for Computer Science and Mathematics, 2023, 4 (04): : 151 - 155
  • [9] Mitigating Hallucination in Visual-Language Models via Re-balancing Contrastive Decoding
    Liang, Xiaoyu
    Yu, Jiayuan
    Mu, Lianrui
    Zhuang, Jiedong
    Hu, Jiaqi
    Yang, Yuchen
    Ye, Jiangnan
    Lu, Lu
    Chen, Jian
    Hu, Haoji
    PATTERN RECOGNITION AND COMPUTER VISION, PT V, PRCV 2024, 2025, 15035 : 482 - 496
  • [10] Empowerment of Large Language Models in Psychological Counseling through Prompt Engineering
    Huang, Shanshan
    Fu, Fuxiang
    Yang, Ke
    Zhang, Ke
    Yang, Fan
    2024 IEEE 4TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING AND ARTIFICIAL INTELLIGENCE, SEAI 2024, 2024, : 220 - 225