Consistency-Constrained Large Language Models for Reliable Legal Reasoning
Main Article Content
Abstract
Large language models (LLMs) exhibit strong general reasoning capabilities but remain vulnerable to instability and hallucinations, particularly under small variations in user prompts. This work introduces a consistency-constrained framework that integrates counterfactual alignment signals into both training and decoding, reducing divergence between predictions on semantically equivalent inputs. Experiments across multiple reasoning benchmarks show that the proposed model significantly improves stability and lowers hallucination rates without sacrificing task accuracy. These findings demonstrate that consistency can be explicitly shaped as an intrinsic property of LLMs, enabling more predictable and reliable reasoning behavior.
Article Details

This work is licensed under a Creative Commons Attribution 4.0 International License.
Mind forge Academia also operates under the Creative Commons Licence CC-BY 4.0. This allows for copy and redistribute the material in any medium or format for any purpose, even commercially. The premise is that you must provide appropriate citation information.