Sequential Decision-Making for Inline Text Autocomplete

By Rohan Chitnis, Shentao Yang, and Alborz Geramifard

Reinforcement Learning Journal, vol. 2, 2024, pp. 946–960.

Presented at the Reinforcement Learning Conference (RLC), Amherst Massachusetts, August 9–12, 2024.


Download:

Abstract:

Autocomplete suggestions are fundamental to modern text entry systems, with applications in domains such as messaging and email composition. Typically, autocomplete suggestions are generated from a language model with a confidence threshold. However, this threshold does not directly take into account the cognitive burden imposed on the user by surfacing suggestions, such as the effort to switch contexts from typing to reading the suggestion, and the time to decide whether to accept the suggestion. In this paper, we study the problem of improving inline autocomplete suggestions in text entry systems via a sequential decision-making formulation, and use reinforcement learning (RL) to learn suggestion policies through repeated interactions with a target user over time. This formulation allows us to factor cognitive burden into the objective of training an autocomplete model, through a reward function based on text entry speed. We acquired theoretical and experimental evidence that, under certain objectives, the sequential decision-making formulation of the autocomplete problem provides a better suggestion policy than myopic single-step reasoning. However, aligning these objectives with real users requires further exploration. In particular, we hypothesize that the objectives under which sequential decision-making can improve autocomplete systems are not tailored solely to text entry speed, but more broadly to metrics such as user satisfaction and convenience.


Citation Information:

Rohan Chitnis, Shentao Yang, and Alborz Geramifard. "Sequential Decision-Making for Inline Text Autocomplete." Reinforcement Learning Journal, vol. 2, 2024, pp. 946–960.

BibTeX:

@article{chitnis2024sequential,
    title={Sequential Decision-Making for Inline Text Autocomplete},
    author={Chitnis, Rohan and Yang, Shentao and Geramifard, Alborz},
    journal={Reinforcement Learning Journal},
    volume={2},
    pages={946--960},
    year={2024}
}