Just a minor fact that doesn't seem to be stated in many places - maybe it's too obvious to state:
In the absence of knowledge of what a probability distribution is, people sometimes like to pick the maximum entropy distribution consistent with the knowledge they do have.
More generally you may have some prior idea of what the distribution is. In that case you might pick the distribution with the maximum entropy *relative* to your prior guess consistent with any new knowledge.
A special case is when you draw from a joint distribution on X and Y and observe X. The posterior distribution of Y is given by the conditional distribution P(Y|X). This is precisely the maximum entropy distribution relative to the original distribution with the constraint that Y is known.
So conditional probabilities are in fact maximum (relative) entropy distributions.
GNU social JP is a social network, courtesy of GNU social JP管理人. It runs on GNU social, version 2.0.2-dev, available under the GNU Affero General Public License.
All GNU social JP content and data are available under the Creative Commons Attribution 3.0 license.