@2014shafto formalize rational pedagogical reasoning—that is, reasoning in a setting in which one agent chooses information to transmit to another agent in order to teach a concept—using a recursive Bayesian model that captures two problems:

  • The teacher needs to choose helpful examples that serve as data for the learner;
  • The learner needs to infer the correct concept from the data, under the assumption that the teacher is choosing helpful examples.

The teacher and learner problems form a mutually dependent system of equations given by:

where is data chosen by the teacher and is the correct hypothesis.

The parameter captures a “soft maximization” of the posterior probability, or the degree to which the teacher tends towards maximizing the posterior instead of choosing the most representative samples. As , the teacher chooses uniformly among data consistent with the hypothesis, with being the case of random sampling (i.e., not helpful) and representing selection in direct proportion to the posterior probability they give to the target hypothesis.

Under this model, teachers compute the utility of selecting an example using theory-of-mind. @2022chen extend this model to capture uncertainty during adaptive and interactive pedagogical setting.