Code for ACL 2024 paper "TruthX: Alleviating Hallucinations by Editing Large Language Models in Truthful Space"
safety
llama
representation
language-model
mistral
explainable-ai
hallucination
baichuan
hallucinations
gpt-4
truthfulness
llm
llms
chatgpt
chatglm
llm-inference
llama2
llama3
-
Updated
Mar 26, 2024 - Python