Personalized Decision Modeling: Utility Optimization or Textualized‑Symbolic Reasoning

1 Johns Hopkins University
2 University of Virginia
NeurIPS 2025 (Spotlight)
* This work was completed while Hongru Du was at Johns Hopkins University.
Correspondence to: Hongru Du and Hao (Frank) Yang.

Abstract

We study personalized decision modeling in settings where individual choices deviate from population-optimal predictions. We introduce ATHENA—an Adaptive Textual‑symbolic Human‑centric Reasoning framework—built in two stages: (1) a group‑level symbolic utility discovery procedure (LLM‑augmented symbolic search) learns compact, interpretable utility forms; and (2) an individual‑level semantic adaptation step refines a personal template with TextGrad to capture each person’s preferences and constraints. On Swissmetro (travel mode choice) and COVID‑19 vaccine uptake tasks, ATHENA consistently outperforms utility‑based, ML, and LLM baselines, improving F1 by ≥ 6.5% over the strongest alternatives while maintaining calibrated probabilistic predictions. Together, these stages bridge classic Random Utility Maximization with textual semantic context, yielding interpretable structure and personalized reasoning.

Method at a Glance

fg,ktk=1Kϕ(g,C,S,Bt1){f^t_{g,k}}_{k=1}^K \sim \phi(\cdot \mid g, C, S, B^{t-1})

Group-level Symbolic Utility Discovery samples candidate symbolic utility functions fg,ktk=1K{f^t_{g,k}}_{k=1}^K from the LLM-conditioned distribution ϕ(g,C,S,Bt1)\phi(\cdot \mid g, C, S, B^{t-1}), guided by the concept library CC, symbolic library SS, and prior feedback Bt1B^{t-1}.

Pit+1PitηPiL(Pit,Di)P^{t+1}_i \leftarrow P^t_i - \eta\,\nabla_{P_i} \, \mathcal{L}(P^t_i, D_i)

Individual-level Semantic Adaptation refines each person’s PiP_i through iterative updates. This process personalizes the template based on individual data DiD_i and semantic gradients, capturing heterogeneous preferences and contextual constraints.


Two-stage ATHENA pipeline combining symbolic discovery and semantic adaptation

Figure 1: Overview of ATHENA framework. The Group-level symbolic utility discovery stage uses an LLM-driven symbolic optimization to find compact utilities fgf_g^*. The Individual-level semantic adaptation stage refines personalized templates Pi\mathcal{P}_i^* via TextGrad to model individual decision rules.

Two‑Stage Pipeline

  1. Group‑Level Symbolic Utility Discovery LLMs sample candidate symbolic utility functions (Eq. 3) from a structured symbolic–semantic space, guided by a concept library CC and feedback Bt1B^{t-1}. Through iterative evaluation and mutation/crossover, the model converges to the optimal symbolic utilityfgf_g^* that best captures group‑level decision regularities.

  2. Individual‑Level Semantic Adaptation Using the discovered fgf_g^* as a strong prior, each individual’s textual semantic templatePiP_i is optimized by TextGrad (Eq. 7) to incorporate heterogeneous personal preferences and constraints. The adaptation continues until the template converges to PiP_i^*.


y^iϕ(Pi,Xifg(Xi;θg))\hat{y}_i \sim \phi\big(P_i^*, X_i \,\big|\, f_g^*(X_i; \theta_g^*)\big)

Personalized Decision Inference integrates the learned symbolic utility fgf_g^* with the adapted semantic template PiP_i^* to generate individualized predictions y^i\hat{y}_i that reflect both group‑level reasoning and personal context.


Optimization flow diagram showing symbolic feedback and TextGrad updates

Figure 2: ATHENA pipeline applied to travel–mode choice. Using Swissmetro as an example, the Initialization encodes constraints and symbolic features. In Group-level optimization, LLMs sample and prune utility formulas {fg}\{f_g^*\}. In Individual adaptation, each fgf_g^* guides a personalized prompt Pi\mathcal{P}_i^* to capture heterogeneity.

Algorithm 1 · ATHENA Optimization Flow

  1. Require Demographic group gg, dataset Dg\mathcal{D}_g, domain concept C\mathcal{C}, symbolic building block S\mathcal{S}

  2. InitializeB0None\mathcal{B}_0 \leftarrow \texttt{None}
  3. // Stage 1: Group-Level Symbolic Utility Discovery
  4. fort=1 to Tt = 1 \text{ to } Tdo

    Sample symbolic utility functions {fg,kt}k=1Kϕ(g,C,S,Bt1)\{f_{g,k}^{t}\}_{k=1}^K \sim \phi(\cdot \mid g, \mathcal{C}, \mathcal{S}, \mathcal{B}^{t-1})

    Update Bt{fg,+t,fg,t}\mathcal{B}^t \leftarrow \{f_{g,+}^t, f_{g,-}^t\} using Eq. (3)(3)

    Select best function fgargminfFgLg(f,Dg)f_g^* \leftarrow \arg\min_{f \in \mathcal{F}_g} \mathcal{L}_g(f, \mathcal{D}_g)

    ifstopping condition in Eq.(4)(4)then break
  5. // Stage 2: Individual-Level Semantic Adaptation
  6. for eachindividualigi \in g

    Initialize semantic template Pi0ϕ(fg,i,C)\mathcal{P}_i^{0} \sim \phi( \cdot \mid f_g^*, i, \mathcal{C})

    fort=1 to Tt = 1 \text{ to } T'do

    Update Pit+1PitηLi(Pit,Di)\mathcal{P}_i^{t+1} \leftarrow \mathcal{P}_i^{t} - \eta \nabla \mathcal{L}_i(\mathcal{P}_i^{t}, \mathcal{D}_i) using Eq. (7)(7)

  7. return {Pi}ig\{\mathcal{P}_i^{*}\}_{i \in g}, predict decisions using Eq. (8)(8)

Results

Overall Performance

Performance comparison of LLM-based, classical choice, and machine learning methods on the three-class Swissmetro and three-class COVID-19 Vaccine choice tasks.

MethodLLM ModelSwissmetroVaccine
Acc. ↑F1. ↑CE ↓AUC ↑Acc. ↑F1. ↑CE ↓AUC ↑
LLM-
Based
Zeroshotgemini-2.0-flash0.59200.29400.92570.65610.58000.50920.83280.7607
GPT-4o-mini0.63000.27572.72580.36570.54330.53870.85620.7395
Zeroshot-CoTgemini-2.0-flash0.58800.34780.94150.63310.58000.50730.84360.7526
GPT-4o-mini0.64200.29600.89570.62370.55000.53530.85400.7465
Fewshotgemini-2.0-flash0.75800.70278.72440.79560.56670.574012.03240.7053
GPT-4o-mini0.68150.49457.00290.73950.50670.50976.61100.6891
TextGradgemini-2.0-flash0.55680.29801.20110.54000.42410.40145.78130.6363
GPT-4o-mini0.65000.36200.90790.53640.50840.49624.54120.6709
ATHENAgemini-2.0-flash0.76790.72220.90410.83870.67970.59680.76100.8370
GPT-4o-mini0.81340.76551.08630.88250.73450.71610.75510.8704
Utility
Theory
MNL/0.61010.38870.83530.70740.41500.19551.05100.4301
CLogit/0.57140.24240.89160.59760.41500.19551.05100.5000
Latent Class MNL/0.61010.39670.81750.71820.19500.10881.09860.5000
Machine
Learning
Logistic Regression/0.56200.55700.93100.74600.65000.66900.76300.8330
Random Forest/0.71000.70500.73800.88100.63000.64700.72900.8420
XGBoost/0.70800.70500.70400.88100.63000.64801.14200.8150
BERT/0.72460.49940.70370.88110.63500.65410.74090.8168
TabNet/0.63750.40600.78870.88100.66500.66840.89680.8147
MLP/0.64750.63860.76260.83500.60680.60620.93200.8205

BibTeX

@inproceedings{zhao2025athena,
title = {Personalized Decision Modeling: Utility Optimization or Textualized-Symbolic Reasoning},
author = {Yibo Zhao, Yang Zhao, Hongru Du, Hao Frank Yang},
booktitle = {The Thirty-Ninth Annual Conference on Neural Information Processing Systems (NeurIPS)},
year = {2025}
}