Papers
Topics
Authors
Recent
Search
2000 character limit reached

Human–LLM Negotiation Studies

Updated 9 April 2026
  • Human–LLM Negotiation Studies is a research domain that examines automated agents’ interaction with humans in bargaining using formal economic and behavioral models.
  • It employs asymmetric, incomplete-information games and the MERIT framework to quantify negotiation performance through metrics like deal rate and utility.
  • The field integrates cognitive modeling, personality analysis, and governance protocols to enhance LLM reasoning, strategic alignment, and safe negotiation practices.

Human–LLM Negotiation Studies

The study of negotiation between humans and LLMs examines how automated agents emulate, interact with, and diverge from human strategic, linguistic, and behavioral patterns in bargaining and dispute resolution. This research domain integrates formal tools from economics, behavioral game theory, computational linguistics, decision sciences, and AI alignment. Central challenges include quantifying negotiation skill, identifying systematic discrepancies with human behavior, and developing benchmarks and training protocols that steer LLMs toward human-aligned outcomes in both cooperative and adversarial settings.

1. Formal Models and Evaluation Benchmarks

The formalization of negotiation tasks in this literature typically centers on asymmetric, incomplete-information games, in which each party possesses private parameters such as valuation, cost, or budget. Core frameworks are instantiated in buyer–seller price haggling, multi-issue bargaining, and controlled environments like the ultimatum game (Xia et al., 2024, Yadav et al., 30 May 2025, Oh et al., 11 Feb 2026). For example, in the canonical bargaining setting, the buyer’s utility is UBuyer=BDU_\mathrm{Buyer} = B - D (budget minus deal price), and the seller’s utility is USeller=DCU_\mathrm{Seller} = D - C (deal price minus true cost) (Xia et al., 2024). These formulations support extensive quantitative assessment over large product datasets (e.g., AmazonHistoryPrice, 930 items, 2009–2023), and enable the distinction between mutual-interest (MI) and conflicting-interest (CI) games depending on private parameter overlap.

Evaluation of agent performance is multidimensional. Traditional metrics include deal rate, profit (absolute or normalized), and agreement efficiency. The MERIT framework (Oh et al., 11 Feb 2026) introduces a suite of economically grounded, human-aligned metrics:

  • Agent Utility (CS): CS=(PwtpPdeal)/(PwtpPcost)CS = (P_{\mathrm{wtp}} - P_{\mathrm{deal}})/(P_{\mathrm{wtp}} - P_{\mathrm{cost}}).
  • Negotiation Power (NP): NP=(PinitialPdeal)/(PinitialPcost)NP = (P_{\mathrm{initial}} - P_{\mathrm{deal}})/(P_{\mathrm{initial}} - P_{\mathrm{cost}}).
  • Acquisition Ratio (AR): AR=cos(vacquired,vdesired)AR = \cos(\vec{v}_{\mathrm{acquired}}, \vec{v}_{\mathrm{desired}}), embedding product semantics.

These are combined as Merit=αCS+βNP+γARMerit = \alpha CS + \beta NP + \gamma AR with weights learned from human preference data (α1.0139,β0.8812,γ1.1049\alpha \approx 1.0139, \beta \approx 0.8812, \gamma \approx 1.1049) (Oh et al., 11 Feb 2026). Specialized benchmarks such as AgoraBench span diverse settings—monopoly, deception, installment payments, and reputation—expanding beyond single-issue splits (Oh et al., 11 Feb 2026).

2. Cognitive and Strategic Dimensions

Human–LLM negotiation research foregrounds the necessity of advanced cognitive modeling, notably Theory-of-Mind (ToM) and strategic depth. Baseline LLMs often exhibit brittle opponent modeling, failing to project reservation prices, anchor dynamically, or adapt to context/power (Shah et al., 15 Dec 2025, Oh et al., 11 Feb 2026). Empirical studies in settings such as the ultimatum game demonstrate that explicit ToM prompting—zero-order (self introspection), first-order (opponent belief modeling), and combined levels—substantially improves alignment to human normative behavior, fairness indices, and acceptance/rejection thresholds (Yadav et al., 30 May 2025). First-order ToM, in particular, most reduces deviation from human-typical proposals, while combined ToM best calibrates responder behavior.

Complementary quantitative models analyze concession dynamics via parameterized trajectories (e.g., hyperbolic tangent fitting: y(x)=d+btanh(axc)y(x) = d + b \tanh(ax - c)), enabling the calculation of metrics such as burstiness (τ)(\tau) and concession rigidity index (CRI) (Shah et al., 15 Dec 2025). These reveal that while humans exhibit context-sensitive, bursty concessions (τ0.39\tau \approx 0.39USeller=DCU_\mathrm{Seller} = D - C0, USeller=DCU_\mathrm{Seller} = D - C1–USeller=DCU_\mathrm{Seller} = D - C2), LLMs systematically default to edge-anchoring, exhibit unresponsive CRI, and lack power-adaptive pacing. Model scaling alone (across GPT-4.1, GPT-4-o, LLaMA derivatives) does not improve these deficits (Shah et al., 15 Dec 2025).

3. Personality, Social Dynamics, and Behavioral Alignment

Multiple studies incorporate the Big Five personality model—Openness, Conscientiousness, Extraversion, Agreeableness, Neuroticism—either as controlled LLM prompts or as analytical covariates (Huang et al., 2024, Cohen et al., 19 Jun 2025, Kwon et al., 19 Sep 2025). LLM-based negotiation agents exhibit trait-linked effects: high Agreeableness and Extraversion consistently boost believability, goal achievement, and knowledge acquisition, while Neuroticism negatively impacts outcomes (Cohen et al., 19 Jun 2025). Simulation work quantifies personality-impact using causal ATE and regression over metrics such as intrinsic utility, joint Nash utility, concession rates, and success rate (Huang et al., 2024). Strategy regression evidences that “accommodate” and “concede” tactics yield higher joint utility, while “assertive/aggressive” behaviors are penalized.

Behavioral alignment is analyzed along three axes:

  • Linguistic style (e.g., LIWC gap, linguistic entrainment),
  • Emotional expression (anger trajectory/magnitude),
  • Strategic behavior (IRP strategy distribution: Interests, Rights, Power) (Kwon et al., 19 Sep 2025).

GPT-4.1 achieves close alignment to human linguistic and affective dynamics, while Claude-3.7 excels in strategic IRP mimicry. Alignment gaps persist in emotional variance and strategic adaptability.

4. Tactics, Reasoning Deficits, and Human Strategy

Fine-grained annotation of negotiation tactics (Ethos–Logos–Pathos taxonomy) via LLM-judge frameworks (AC1 > 0.8 reliability) identifies key predictors of short- and long-term success in high-complexity multi-party games like Diplomacy (Li et al., 20 Dec 2025). Success correlates most strongly with explicit move discussion, rapport-building, reasoning, information sharing, and apologies. LLM-generated dialogues underuse socio-emotional tactics and demonstrate stylistic divergence (measured via vector distances in tactic frequency), though supervised fine-tuning on human-successful data partially closes the gap.

Human participants, in adversarial or non-collaborative negotiation with LLMs, exploit “reasoning hacks” such as repeated defect claims, context manipulation, and prompt injection (“ignore all instructions”), often obtaining highly favorable or irrational deals. These methods exploit LLMs’ reasoning deficits—non-monotonic concessions, blind trust, arithmetic errors—underscoring the critical importance of AI-negotiation literacy for users and robust consistency checks for models (Schneider et al., 2023).

5. Architectures, Governance, and B2B Protocols

The deployment of LLM-based negotiation systems in enterprise or mission-critical contexts necessitates governance, information control, and safety protocols. The GAIA framework (Zhao et al., 9 Nov 2025) formalizes delegation with bounded authorization over labeled state transitions (START, SCREEN, NEGOTIATE, ESCALATE, etc.), rigorous information gating enabled by Task-Completeness Index (TCI), dual feedback integration (human micro-corrections and AI Critic modules), and escalation mechanisms for commitments or boundary violations. Four safety invariants—no unauthorized commitment, information-gate monotonicity, safety preflight, and liveness—guarantee both operational and audit integrity. Protocol-level and human judgment metrics quantify efficiency, safety, satisfaction, and normalized utility.

A plausible implication is that such architectures, when instantiated with robust validation and escalation workflows, could bridge the gap between controlled agent-to-agent benchmark research and safe, auditable LLM-human negotiation in B2B and regulated environments.

6. Mitigation, Enhancement, and Open Challenges

Multiple studies converge on the result that model scaling or raw parameter increase does not resolve fundamental strategic and ToM limitations (Shah et al., 15 Dec 2025, Xia et al., 2024). Instead, architectural and protocol innovations are required:

  • Reward Model Enhancement: MERIT feedback, integrating cardinal (CS), ordinal (AR), and skill-based (NP) dimensions via private reward signaling, substantively improves both deal rate and human-preference alignment in complex scenarios (Merit increases from 0.9–1.4 to 1.45–1.84) (Oh et al., 11 Feb 2026).
  • Modularity: OG-Narrator (Xia et al., 2024) demonstrates the efficacy of decoupling deterministic strategic planning (Offer Generator) from natural language realization (LLM Narrator), raising buyer deal rates from ~27% to ~89% and profit by an order of magnitude.
  • Prompting and Fine-Tuning: Explicit ToM prompting, ICL-MF schema, and fine-tuning on human-preferred negotiations boost ToM-related reasoning and opponent-awareness; in MERIT, the inclusion of “Thoughts – Talk – Action” increases opponent modeling frequency 25.6% vs. 2.1% in baseline (Oh et al., 11 Feb 2026, Yadav et al., 30 May 2025).
  • Metric-Guided Training: Use of burstiness USeller=DCU_\mathrm{Seller} = D - C3 and CRI as auxiliary losses, opponent-modeling curricula, and reinforcement from human feedback is proposed as a route toward humanizable negotiation models (Shah et al., 15 Dec 2025).

Open problems include bilateral (not buyer-centric) reward modeling, adaptation to cross-cultural and regulatory domains, tool-augmented negotiation (e.g., live price search), and robust defense against adversarial human strategies in LLM-mediated bargaining (Schneider et al., 2023, Oh et al., 11 Feb 2026, Zhao et al., 9 Nov 2025).


Key Papers Referenced:

  • "MERIT Feedback Elicits Better Bargaining in LLM Negotiators" (Oh et al., 11 Feb 2026)
  • "Effects of Theory of Mind and Prosocial Beliefs on Steering Human-Aligned Behaviors of LLMs in Ultimatum Games" (Yadav et al., 30 May 2025)
  • "Measuring Bargaining Abilities of LLMs: A Benchmark and A Buyer-Enhancement Method" (Xia et al., 2024)
  • "LLM Rationalis? Measuring Bargaining Capabilities of AI Negotiators" (Shah et al., 15 Dec 2025)
  • "How Personality Traits Influence Negotiation Outcomes? A Simulation based on LLMs" (Huang et al., 2024)
  • "Evaluating Behavioral Alignment in Conflict Dialogue: A Multi-Dimensional Comparison of LLM Agents and Humans" (Kwon et al., 19 Sep 2025)
  • "GAIA: A General Agency Interaction Architecture for LLM-Human B2B Negotiation & Screening" (Zhao et al., 9 Nov 2025)
  • "Exploring Big Five Personality and AI Capability Effects in LLM-Simulated Negotiation Dialogues" (Cohen et al., 19 Jun 2025)
  • "Negotiating with LLMs: Prompt Hacks, Skill Gaps, and Reasoning Deficits" (Schneider et al., 2023)
  • "Measuring Fine-Grained Negotiation Tactics of Humans and LLMs in Diplomacy" (Li et al., 20 Dec 2025)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Human–LLM Negotiation Studies.