Tuning-Free Personalized Alignment via Trial-Error-Explain In-Context Learning


1        2 

Challenge: Personalizing Language Models for Text Generation in Realistic Settings

Language models are aligned to the collective voice of many, resulting in generic outputs that do not align with specific users' styles. Yet end users often have specific needs that call for personalized text generation, such as writing an email. Therefore, adapting LLMs for personalized text generation has drawn a growing interest, but prior work on personalized text generation either rely on large amounts of personal or parameter updates that result in cumbersome per-user model parameters. Can we instead personalize language models for text generation tasks without any parameter updates and with only a few examples?

Trial-Error-Explain In-Context Learning (TICL)

TICL Overview

To address this challenge, we propose Trial-Error-Explain In-Context Learning (TICL), an inference-only approach that front-loads inference compute to create a user-specific in-context learning prompt that does not require extra generation steps at test time. The key idea of TICL is to iteratively expand a few-shot in-context learning prompt via a trial-error-explain process, adding model-generated negative samples and explanations that provide fine-grained guidance towards a specific user's style.

TICL Teaser

We find that a providing model with its own "failures" as negative samples and corresponding explanations enables the model to learn stylistic context more effectively and overcome the bias towards structural and formal phrases observed in their zero-shot and few-shot outputs. TICL achieves favorable win rates on pairwise comparisons with LLM-as-a-judge up to 91.5% against the previous state-of-the-art and outperforms competitive tuning-free baselines for personalized alignment tasks of writing emails, essays and news articles.

BibTeX

@misc{cho2025tuningfreepersonalizedalignmenttrialerrorexplain,
        title={Tuning-Free Personalized Alignment via Trial-Error-Explain In-Context Learning}, 
        author={Hyundong Cho and Karishma Sharma and Nicolaas Jedema and Leonardo F. R. Ribeiro and Alessandro Moschitti and Ravi Krishnan and Jonathan May},
        year={2025},
        eprint={2502.08972},
        archivePrefix={arXiv},
        primaryClass={cs.CL},
        url={https://arxiv.org/abs/2502.08972}, 
  }