Skip to content

Latest commit

 

History

History
160 lines (147 loc) · 69.7 KB

Truthfulness&Misinformation.md

File metadata and controls

160 lines (147 loc) · 69.7 KB

Truthfulness&Misinformation

Different from the main README🕵️

  • Within this subtopic, we will be updating with the latest articles. This will help researchers in this area to quickly understand recent trends.
  • In addition to providing the most recent updates, we will also add keywords to each subtopic to help you find content of interest more quickly.
  • Within each subtopic, we will also update with profiles of scholars we admire and endorse in the field. Their work is often of high quality and forward-looking!"

📑Papers

Date Institute Publication Paper Keywords
21.09 University of Oxford ACL2022 TruthfulQA: Measuring How Models Mimic Human Falsehoods Benchmark&Truthfulness
23.07 Microsoft Research Asia, Hong Kong University of Science and Technology, University of Science and Technology of China, Tsinghua University, Sony AI ResearchSquare Defending ChatGPT against Jailbreak Attack via Self-Reminder Jailbreak Attack&Self-Reminder&AI Security
23.10 University of Zurich arxiv Lost in Translation -- Multilingual Misinformation and its Evolution Misinformation&Multilingual
23.10 New York University&Javier Rando arxiv Personas as a Way to Model Truthfulness in Language Models Truthfulness&Truthful Persona
23.11 Dialpad Canada Inc arxiv Are Large Language Models Reliable Judges? A Study on the Factuality Evaluation Capabilities of LLMs Factuality Assessment
23.11 The University of Manchester arxiv Emotion Detection for Misinformation: A Review Survey&Misinformation&Emotions
23.11 University of Virginia arxiv Can Language Models Be Tricked by Language Illusions? Easier with Syntax, Harder with Semantics Language Illusions
23.11 University of Illinois Urbana-Champaign arxiv Learn to Refuse: Making Large Language Models More Controllable and Reliable through Knowledge Scope Limitation and Refusal Mechanism Hallucinations&Refusal Mechanism
23.11 University of Washington Bothell arxiv Creating Trustworthy LLMs: Dealing with Hallucinations in Healthcare AI Healthcare&Trustworthiness&Hallucinations
23.11 Intuit AI Research EMNLP2023 SAC3: Reliable Hallucination Detection in Black-Box Language Models via Semantic-aware Cross-check Consistency Hallucination Detection&Trustworthiness
23.11 Shanghai Jiao Tong University arxiv Support or Refute: Analyzing the Stance of Evidence to Detect Out-of-Context Mis- and Disinformation Misinformation&Disinformation&Out-of-Context
23.11 Hamad Bin Khalifa University arxiv ArAIEval Shared Task: Persuasion Techniques and Disinformation Detection in Arabic Text Disinformation&Arabic Text
23.11 UNC-Chapel Hill arxiv Holistic Analysis of Hallucination in GPT-4V(ision): Bias and Interference Challenges Hallucination&Benchmark&Multimodal
23.11 Cornell University arxiv Adapting Fake News Detection to the Era of Large Language Models Fake news detection&Generated News&Misinformation
23.11 Harbin Institute of Technology arxiv A Survey on Hallucination in Large Language Models: Principles, Taxonomy, Challenges, and Open Questions Hallucination&Factual Consistency&Trustworthiness
23.11 Korea University, KAIST AI,LG AI Research arXiv VOLCANO: Mitigating Multimodal Hallucination through Self-Feedback Guided Revision Multimodal Models&Hallucination&Self-Feedback
23.11 Beijing Jiaotong University, Alibaba Group, Peng Cheng Lab arXiv AMBER: An LLM-free Multi-dimensional Benchmark for MLLMs Hallucination Evaluation Multi-modal Large Language Models&Hallucination&Benchmark
23.11 LMU Munich; Munich Center of Machine Learning; Google Research arXiv Hallucination Augmented Recitations for Language Models Hallucination&Counterfactual Datasets
23.11 Stanford University, UNC Chapel Hill arxiv Fine-tuning Language Models for Factuality Factuality&Reference-Free Truthfulness&Direct Preference Optimization
23.11 Corporate Data and Analytics Office (CDAO) arxiv Hallucination-minimized Data-to-answer Framework for Financial Decision-makers Financial Decision Making&Hallucination Minimization
23.11 Arizona State University arxiv Can Knowledge Graphs Reduce Hallucinations in LLMs? : A Survey Knowledge Graphs&Hallucinations&Survey
23.11 Kempelen Institute of Intelligent Technologies; Brno University of Technology arxiv Disinformation Capabilities of Large Language Models Disinformation Generation&Safety Filters&Automated Evaluation
23.11 UNC-Chapel Hill, University of Washington arxiv EVER: Mitigating Hallucination in Large Language Models through Real-Time Verification and Rectification Hallucination&Real-Time Verification&Rectification
23.11 Peking University, WeChat AI, Tencent Inc. arXiv RECALL: A Benchmark for LLMs Robustness against External Counterfactual Knowledge External Counterfactual Knowledge&Benchmarking&Robustness
23.11 PolyAI Limited arXiv Dial BEINFO for Faithfulness: Improving Factuality of Information-Seeking Dialogue via Behavioural Fine-Tuning Factuality&Behavioural Fine-Tuning&Hallucination
23.11 The Hong Kong University of Science and Technology, University of Illinois Urbana-Champaign arxiv R-Tuning: Teaching Large Language Models to Refuse Unknown Questions Hallucination&Refusal-Aware Instruction Tuning&Knowledge Gap
23.11 University of Southern California, University of Pennsylvania, University of California Davis arxiv Deceiving Semantic Shortcuts on Reasoning Chains: How Far Can Models Go without Hallucination? Hallucinations&Semantic Associations&Benchmark
23.11 The Ohio State University, University of California Davis arxiv How Trustworthy are Open-Source LLMs? An Assessment under Malicious Demonstrations Shows their Vulnerabilities Trustworthiness&Malicious Demonstrations&Adversarial Attacks
23.11 University of Sheffield arXiv Lighter yet More Faithful: Investigating Hallucinations in Pruned Large Language Models for Abstractive Summarization Hallucinations&&Language Model Reliability
23.11 Institute of Information Engineering Chinese Academy of Sciences, University of Chinese Academy of Sciences arxiv Can Large Language Models Understand Content and Propagation for Misinformation Detection: An Empirical Study Misinformation Detection
23.11 Shanghai Jiaotong University, Amazon AWS AI, Westlake University, IGSNRR Chinese Academy of Sciences, China arXiv Enhancing Uncertainty-Based Hallucination Detection with Stronger Focus Hallucination Detection&Uncertainty-Based Methods&Factuality Checking
23.11 Institute of Software Chinese Academy of Sciences, University of Chinese Academy of Sciences arXiv Mitigating Large Language Model Hallucinations via Autonomous Knowledge Graph-based Retrofitting Hallucinations&Knowledge Graphs&Retrofitting
23.11 Applied Research Quantiphi arxiv Minimizing Factual Inconsistency and Hallucination in Large Language Models Factual Inconsistency&Hallucination
23.11 Microsoft Research, Georgia Tech arxiv Calibrated Language Models Must Hallucinate Hallucination&Calibration&Statistical Analysis
23.11 School of Information Renmin University of China arxiv UHGEval: Benchmarking the Hallucination of Chinese Large Language Models via Unconstrained Generation Hallucination&Evaluation Benchmark
23.11 DAMO Academy Alibaba Group, Nanyang Technological University, Hupan Lab arxiv Mitigating Object Hallucinations in Large Vision-Language Models through Visual Contrastive Decoding Vision-Language Models&Object Hallucinations
23.11 Shanghai AI Laboratory arxiv Beyond Hallucinations: Enhancing LVLMs through Hallucination-Aware Direct Preference Optimization Multimodal Language Models&Hallucination Problem&Direct Preference Optimization
23.12 Singapore Management University, Beijing Forestry University, University of Electronic Science and Technology of China MMM 2024 Mitigating Fine-Grained Hallucination by Fine-Tuning Large Vision-Language Models with Caption Rewrites Vision-language Models&Hallucination&Fine-grained Evaluation
23.12 Mila, McGill University EMNLP2023(findings) Evaluating Dependencies in Fact Editing for Language Models: Specificity and Implication Awareness Knowledge Bases&Dataset&Evaluation Protocol
23.12 MIT CSAIL arxiv Cognitive Dissonance: Why Do Language Model Outputs Disagree with Internal Representations of Truthfulness? Truthfulness&Internal Representations
23.12 University of Illinois Chicago, Bosch Research North America & Bosch Center for Artificial Intelligence (BCAI), UNC Chapel-Hill arxiv DELUCIONQA: Detecting Hallucinations in Domain-specific Question Answering Hallucination Detection&Domain-specific QA&Retrieval-augmented LLMs
23.12 The University of Hong Kong, Beihang University AAAI2024 Improving Factual Error Correction by Learning to Inject Factual Errors Factual Error Correction
23.12 Allen Institute for AI arxiv BARDA: A Belief and Reasoning Dataset that Separates Factual Accuracy and Reasoning Ability Dataset&Factual Accuracy&Reasoning Ability
23.12 Tsinghua University, Shanghai Jiao Tong University, Stanford University, Nanyang Technological University arxiv The Earth is Flat because...: Investigating LLMs’ Belief towards Misinformation via Persuasive Conversation Misinformation&Persuasive Conversation&Factual Questions
23.12 University of California Davis arXiv A Revisit of Fake News Dataset with Augmented Fact-checking by ChatGPT Fake News&Fact-checking
23.12 Amazon Web Services arxiv On Early Detection of Hallucinations in Factual Question Answering Hallucinations&Factual Question Answering
23.12 University of California Santa Cruz arxiv Don’t Believe Everything You Read: Enhancing Summarization Interpretability through Automatic Identification of Hallucinations in Large Language Models Hallucinations&Faithfulness&Token-level
23.12 Department of Radiology, The University of Tokyo Hospital arxiv Theory of Hallucinations based on Equivariance Hallucinations&Equivariance
23.12 Georgia Institute of Technology arXiv REDUCING LLM HALLUCINATIONS USING EPISTEMIC NEURAL NETWORKS Hallucinations&Uncertainty Estimation&TruthfulQA
23.12 Institute of Artificial Intelligence, School of Computer Science and Technology, Soochow University, Tencent AI Lab arXiv Alleviating Hallucinations of Large Language Models through Induced Hallucinations Hallucinations&Induce-then-Contrast Decoding&Factuality
23.12 SKLOIS Institute of Information Engineering Chinese Academy of Sciences, School of Cyber Security University of Chinese Academy of Sciences arXiv LLM Factoscope: Uncovering LLMs’ Factual Discernment through Inner States Analysis Factual Detection&Inner States
24.01 The Chinese University of Hong Kong, Tencent AI Lab arxiv The Earth is Flat? Unveiling Factual Errors in Large Language Models Factual Errors&Knowledge Graph&Answer Assessment
24.01 NewsBreak, University of Illinois Urbana-Champaign arxiv RAGTruth: A Hallucination Corpus for Developing Trustworthy Retrieval-Augmented Language Models Retrieval-Augmented Generation&Hallucination Detection&Dataset
24.01 University of California Berkeley, Université de Montréal, McGill University, Mila arxiv Uncertainty Resolution in Misinformation Detection Misinformation&Uncertainty Resolution
24.01 Yale University, Stanford University arxiv Large Legal Fictions: Profiling Legal Hallucinations in Large Language Models Legal Hallucinations
24.01 Islamic University of Technology, AI Institute University of South Carolina, Stanford University, Amazon AI arxiv A Comprehensive Survey of Hallucination Mitigation Techniques in Large Language Models ß Hallucination Mitigation
24.01 Renmin University of China, Renmin University of China, DIRO, Université de Montréal arxiv The Dawn After the Dark: An Empirical Study on Factuality Hallucination in Large Language Models Hallucination&Detection and Mitigation&Empirical Study
24.01 IIT Hyderabad India, Parmonic USA, University of Glasgow UK, LDRP Institute of Technology and Research India arxiv Fighting Fire with Fire: Adversarial Prompting to Generate a Misinformation Detection Dataset Misinformation Detection&LLM-generated Synthetic Data
24.01 University College London arxiv Hallucination Benchmark in Medical Visual Question Answering Medical Visual Question Answering&Hallucination Benchmark
24.01 Soochow University arxiv LightHouse: A Survey of AGI Hallucination AGI Hallucination
24.01 University of Washington, Carnegie Mellon University, Allen Institute for AI arxiv Fine-grained Hallucination Detection and Editing for Language Models Hallucination Detection&FAVA
24.01 Dartmouth College, Université de Montréal, McGill University,Mila arxiv Comparing GPT-4 and Open-Source Language Models in Misinformation Mitigation GPT-4&Misinformation Detection
24.01 Utrecht University arxiv The Pitfalls of Defining Hallucination Hallucination
24.01 Samsung AI Center arxiv Hallucination Detection and Hallucination Mitigation: An Investigation Hallucination Detection&Hallucination Mitigation
24.01 McGill University, Mila, Université de Montréal arxiv Combining Confidence Elicitation and Sample-based Methods for Uncertainty Quantification in Misinformation Mitigation Misinformation Mitigation&Uncertainty Quantification&Sample-based Consistency
24.01 LY Corporation arxiv On the Audio Hallucinations in Large Audio-Video Language Models Audio Hallucinations&Audio-visual Learning&Audio-video language Models
24.01 Sun Yat-sen University Tencent AI Lab arXiv Mitigating Hallucinations of Large Language Models via Knowledge Consistent Alignment Hallucination Mitigation&Knowledge Consistent Alignment
24.01 National University of Singapore arxiv Hallucination is Inevitable: An Innate Limitation of Large Language Models Hallucination&Real World LLMs
24.01 X2Robot&International Digital Economy Academy arXiv Learning to Trust Your Feelings: Leveraging Self-awareness in LLMs for Hallucination Mitigation Hallucination Mitigation&Knowledge Probing&Reinforcement Learning
24.01 University of Texas at Austin, Northeastern University arxiv Diverse but Divisive: LLMs Can Exaggerate Gender Differences in Opinion Related to Harms of Misinformation Misinformation Detection&Socio-Technical Systems
24.01 National University of Defense Technology, National University of Singapore arxiv SWEA: Changing Factual Knowledge in Large Language Models via Subject Word Embedding Altering Factual Knowledge Editing&Word Embeddings
24.02 University of Washington, University of California Berkeley, The Hong Kong University of Science and Technology, Carnegie Mellon University arxiv Don’t Hallucinate Abstain: Identifying LLM Knowledge Gaps via Multi-LLM Collaboration Knowledge Gaps&Multi-LLM Collaboration
24.02 IT Innovation and Research Center, Huawei Technologies arxiv A Survey on Hallucination in Large Vision-Language Models Large Vision-Language Models&Hallucination&Mitigation Strategies
24.02 Tianjin University, National University of Singapore, A*STAR arxiv SKIP \N: A SIMPLE METHOD TO REDUCE HALLUCINATION IN LARGE VISION-LANGUAGE MODELS Semantic Shift Bias&Hallucination Mitigation&Vision-Language Models
24.02 University of Marburg, University of Mannheim EACL Findings 2024 The Queen of England is not England’s Queen: On the Lack of Factual Coherency in PLMs Factual Coherency&Knowledge Bases
24.02 MBZUAI, Monash University, LibrAI, Sofia University arxiv Factuality of Large Language Models in the Year 2024 Factuality&Evaluation&Multimodal LLMs
24.02 Institute of Information Engineering, Chinese Academy of Sciences, University of Chinese Academy of Sciences arxiv Are Large Language Models Table-based Fact-Checkers? Table-based Fact Verification&In-context Learning
24.02 Zhejiang University, Ant Group arxiv Unified Hallucination Detection for Multimodal Large Language Models Multimodal Large Language Models&Hallucination Detection&Benchmark
24.02 Alibaba Cloud, Zhejiang University ICLR2024 INSIDE: LLMS’ INTERNAL STATES RETAIN THE POWER OF HALLUCINATION DETECTION Hallucination Detection&EigenScore
24.02 The Hong Kong University of Science and Technology, University of Illinois at Urbana-Champaign, The Hong Kong Polytechnic University arxiv The Instinctive Bias: Spurious Images lead to Hallucination in MLLMs Multimodal Large Language Models&Hallucination
24.02 Institute of Automation Chinese Academy of Sciences, University of Chinese Academy of Sciences arxiv Can Large Language Models Detect Rumors on Social Media? Rumor Detection&Social Media
24.02 CAS Key Laboratory of AI Safety, School of Computer Science and Technology University of Chinese Academy of Science, International Digital Economy Academy IDEA Research arxiv A Survey on Large Language Model Hallucination via a Creativity Perspective Creativity&Hallucination
24.02 University College London, Speechmatics, MATS, Anthropic, FAR AI arxiv Debating with More Persuasive LLMs Leads to More Truthful Answers Debate&Truthfulness
24.02 University of Illinois Urbana-Champaign, DAMO Academy Alibaba Group, Northwestern University arxiv Towards Faithful Explainable Fact-Checking via Multi-Agent Debate Fact-checking&Explainability
24.02 Rice Universitym, Texas A&M University, Wake Forest University, New Jersey Institute of Technology, Meta Platforms Inc. arxiv Large Language Models As Faithful Explainers Explainability&Fidelity&Optimization
24.02 The Hong Kong University of Science and Technology arxiv Do LLMs Know about Hallucination? An Empirical Investigation of LLM’s Hidden States Hallucination&Hidden States&Model Interpretation
24.02 UC Santa Cruz, ByteDance Research, Northwestern University arxiv MEASURING AND REDUCING LLM HALLUCINATION WITHOUT GOLD-STANDARD ANSWERS VIA EXPERTISE-WEIGHTING Large Language Models (LLMs)&Hallucination&Factualness Evaluations&FEWL
24.02 Paul G. Allen School of Computer Science & Engineering, University of Washington arxiv Comparing Hallucination Detection Metrics for Multilingual Generation Hallucination Detection&Multilingual Generation&Lexical Metrics&Natural Language Inference (NLI)
24.02 Institute of Computing Technology, Chinese Academy of Sciences; University of Chinese Academy of Sciences arxiv Retrieve Only When It Needs: Adaptive Retrieval Augmentation for Hallucination Mitigation in Large Language Models Large Language Models (LLMs)&Hallucination Mitigation&Retrieval Augmentation&Rowen
24.02 Institute of Automation, Chinese Academy of Sciences, University of Chinese Academy of Sciences, Nanjing University arxiv Logical Closed Loop: Uncovering Object Hallucinations in Large Vision-Language Models Object Hallucination&Vision-Language Models (LVLMs)
24.02 Institute of Mathematics and Statistics University of São Paulo, Artificial Intelligence Specialist in the Banking Sector arxiv Hallucinations or Attention Misdirection? The Path to Strategic Value Extraction in Business Using Large Language Models Hallucinations&Generative Artificial Intelligence
24.02 Stevens Institute of Technology, Peraton Labs arxiv Can Large Language Models Detect Misinformation in Scientific News Reporting? Scientific Reporting&Misinformation&Explainability
24.02 Middle East Technical University arxiv HypoTermQA: Hypothetical Terms Dataset for Benchmarking Hallucination Tendency of LLMs Hallucination&Benchmarking Dataset
24.02 National University of Singapore arxiv Seeing is Believing: Mitigating Hallucination in Large Vision-Language Models via CLIP-Guided Decoding Vision-Language Models&Hallucination&CLIP-Guided Decoding
24.02 University of California Los Angeles, Cisco Research arxiv Characterizing Truthfulness in Large Language Model Generations with Local Intrinsic Dimension Truthfulness&Local Intrinsic Dimension
24.02 Institute of Automation Chinese Academy of Sciences, School of Artificial Intelligence University of Chinese Academy of Sciences, Hunan Normal University arxiv Whispers that Shake Foundations: Analyzing and Mitigating False Premise Hallucinations in Large Language Models False Premise Hallucinations&Attention Mechanism
24.02 Shanghai Artificial Intelligence Laboratory, Renmin University of China, University of Chinese Academy of Sciences, Shanghai Jiao Tong University, The University of Sydney arxiv Towards Tracing Trustworthiness Dynamics: Revisiting Pre-training Period of Large Language Models Trustworthiness Dynamics&Pre-training
24.03 École polytechnique fédérale de Lausanne, Carnegie Mellon University, University of Maryland College Park arxiv "Flex Tape Can’t Fix That": Bias and Misinformation in Edited Language Models Model Editing&Demographic Bias&Misinformation
24.03 East China Normal University arxiv DiaHalu: A Dialogue-level Hallucination Evaluation Benchmark for Large Language Models Dialogue-level Hallucination&Benchmarking&Human-machine Interaction
24.03 Peking University arxiv Evaluating and Mitigating Number Hallucinations in Large Vision-Language Models: A Consistency Perspective Number Hallucination&Vision-Language Models&Consistency Training
24.03 City University of Hong Kong, National University of Singapore, Shanghai Jiao Tong University, Stanford University, Penn State University, Hong Kong University of Science and Technology arxiv IN-CONTEXT SHARPNESS AS ALERTS: AN INNER REPRESENTATION PERSPECTIVE FOR HALLUCINATION MITIGATION Hallucination&Inner Representation&Entropy
24.03 Microsoft arxiv In Search of Truth: An Interrogation Approach to Hallucination Detection Hallucination Detection&Interrogation Technique&Balanced Accuracy
24.03 Mohamed bin Zayed University of Artificial Intelligence arxiv Multimodal Large Language Models to Support Real-World Fact-Checking Multimodal Large Language Models&Fact-Checking&Misinformation
24.03 KAIST, Microsoft Research Asia arxiv ERBENCH: AN ENTITY-RELATIONSHIP BASED AUTOMATICALLY VERIFIABLE HALLUCINATION BENCHMARK FOR LARGE LANGUAGE MODELS Hallucination&Entity-Relationship Model&Benchmarking
24.03 University of Alberta, Platform and Content Group, Tencent arxiv SIFiD: Reassess Summary Factual Inconsistency Detection with LLM Factual Consistency&Summarization
24.03 Key Laboratory of Intelligent Information Processing, Institute of Computing Technology, Chinese Academy of Sciences; University of Chinese Academy of Sciences arxiv Truth-Aware Context Selection: Mitigating the Hallucinations of Large Language Models Being Misled by Untruthful Contexts Truth Detection&Context Selection
24.03 UC Berkeley, Google DeepMind arxiv Unfamiliar Finetuning Examples Control How Language Models Hallucinate Large Language Models&Finetuning&Hallucination Control
24.03 University of Alberta, Platform and Content Group, Tencent arxiv SIFiD: Reassess Summary Factual Inconsistency Detection with LLM Factual Consistency&Summarization
24.03 Key Laboratory of Intelligent Information Processing, Institute of Computing Technology, Chinese Academy of Sciences; University of Chinese Academy of Sciences arxiv Truth-Aware Context Selection: Mitigating the Hallucinations of Large Language Models Being Misled by Untruthful Contexts Truth Detection&Context Selection
24.03 UC Berkeley, Google DeepMind arxiv Unfamiliar Finetuning Examples Control How Language Models Hallucinate Large Language Models&Finetuning&Hallucination Control
24.03 Google Research, UC San Diego COLING 2024 Detecting Hallucination and Coverage Errors in Retrieval Augmented Generation for Controversial Topics Conversational Systems&Evaluation Methodologies
24.03 University of Maryland, University of Antwerp, New York University arxiv Evaluating LLMs for Gender Disparities in Notable Persons Bias&Fairness&Hallucinations
24.03 University of Duisburg-Essen arxiv The Human Factor in Detecting Errors of Large Language Models: A Systematic Literature Review and Future Research Directions Hallucination
24.03 Wuhan University, Beihang University, The University of Sydney, Nanyang Technological University COLING 2024 Take Care of Your Prompt Bias! Investigating and Mitigating Prompt Bias in Factual Knowledge Extraction Factual Knowledge Extraction&Prompt Bias
24.03 Carnegie Mellon University arxiv Enhancing LLM Factual Accuracy with RAG to Counter Hallucinations: A Case Study on Domain-Specific Queries in Private Knowledge-Bases Retrieval Augmented Generation (RAG)&Private Knowledge-Bases&Hallucinations
24.03 Integrated Vision and Language Lab KAIST South Korea arxiv What if...?: Counterfactual Inception to Mitigate Hallucination Effects in Large Multimodal Models Large Multimodal Models&Hallucination
24.03 UCAS arxiv MMIDR: Teaching Large Language Model to Interpret Multimodal Misinformation via Knowledge Distillation Multimodal Misinformation Detection&Knowledge Distillation
24.03 Seoul National University, Sogang University arxiv Exploiting Semantic Reconstruction to Mitigate Hallucinations in Vision-Language Models Semantic Reconstruction&Vision-Language Models&Hallucination Mitigation
24.03 University of Illinois Urbana-Champaign arxiv Hallucination Detection in Foundation Models for Decision-Making: A Flexible Definition and Review of the State of the Art Hallucination Detection&Foundation Models&Decision-Making
24.03 Shanghai Jiao Tong University arxiv Rejection Improves Reliability: Training LLMs to Refuse Unknown Questions Using RL from Knowledge Feedback Knowledge Feedback&Reliable Reward Model&Refusal Mechanism
24.03 Universität Hamburg, The University of Sydney arxiv Mitigating Hallucinations in Large Vision-Language Models with Instruction Contrastive Decoding Instruction Contrastive Decoding&Large Vision-Language Models&Hallucination Mitigation
24.03 AI Institute University of South Carolina, Indian Institute of Technology Kharagpur, Islamic University of Technology, Stanford University, Amazon AI arxiv “Sorry Come Again?” Prompting – Enhancing Comprehension and Diminishing Hallucination with [PAUSE] -injected Optimal Paraphrasing Prompt Engineering&Hallucination Mitigation&[PAUSE] Injection
24.04 Beihang University, School of Computer Science and Engineering, School of Software, Shandong University arxiv Exploring and Evaluating Hallucinations in LLM-Powered Code Generation Code Generation&Hallucination
24.03 Department of Electronic Engineering, Tsinghua University, Pattern Recognition Center, WeChat AI, Tencent Inc, China NAACL 2024 On Large Language Models’ Hallucination with Regard to Known Facts Hallucination&Inference Dynamics
24.04 University of California, Berkeley NAACL 2024 ALOHa: A New Measure for Hallucination in Captioning Models Adversarial Attack&AI-Text Detection
24.04 Technical University of Munich, University of Stavanger, University of Alberta arxiv PoLLMgraph: Unraveling Hallucinations in Large Language Models via State Transition Dynamics Hallucination Detection&State Transition Dynamics&Large Language Models
24.04 University of Edinburgh, University College London, Peking University, Together AI arxiv The Hallucinations Leaderboard – An Open Effort to Measure Hallucinations in Large Language Models Hallucination Detection&Benchmarking
24.04 IIIT Hyderabad, Purdue University, Northwestern University, Indiana University Indianapolis arxiv Halu-NLP at SemEval-2024 Task 6: MetaCheckGPT - A Multi-task Hallucination Detection Using LLM Uncertainty and Meta-models Hallucination Detection&LLM Uncertainty&Meta-models
24.04 ServiceNow NAACL 2024 Reducing hallucination in structured outputs via Retrieval-Augmented Generation Retrieval-Augmented Generation&Structured Outputs&Generative AI
24.04 Technion – Israel Institute of Technology, Google Research arxiv Constructing Benchmarks and Interventions for Combating Hallucinations in LLMs Hallucinations&Benchmarks
24.04 The University of Texas at Austin, Salesforce AI Research arxiv MiniCheck: Efficient Fact-Checking of LLMs on Grounding Documents Fact-Checking&Efficiency
24.04 Meta, Technical University of Munich arxiv Uncertainty-Based Abstention in LLMs Improves Safety and Reduces Hallucinations Safety&Hallucinations&Uncertainty

💻Presentations & Talks

📖Tutorials & Workshops

Date Type Title URL
23.10 Tutorials Awesome-LLM-Safety link

📰News & Articles

🧑‍🏫Scholars