How does entropy depend on location and scale?Entropy of Cauchy (Lorentz) DistributionDeriving Negentropy. Getting stuck[Revised]Proving the expected bolddensity of being the Nth order statistics is decreasing in sample sizeRelationship between least-squares regression and information theoryAsymptotic relation for a class of probability distribution functionsLet $X_(1)leq X_(2)$ be the order statistics. Evaluate $operatornameVar(X_(j))$, $operatornameCov(X_(1),X_(2))$Why we multiply the density instead of taking the integral?Is continuous mutual information the correct analogue of the discrete version?What does the integral of a function times a function of a random variable represent, conceptually?If $Xsim operatornamelognormal$ then $Y:=(X-dmid xgeq d)$ has approximately a Generalized Pareto distributionSequence of shifted exponential distributions has uniform conditionals?
Which high-degree derivatives play an essential role?
What was the ASCII end of medium (EM) character intended to be used for?
Polynomial and roots problems
Can I deep fry food in butter instead of vegetable oil?
German idiomatic equivalents of 能骗就骗 (if you can cheat, then cheat)
Why should I allow multiple IP addresses on a website for a single session?
Simplify the code
Are there advantages in writing by hand over typing out a story?
Cannot overlay, because ListPlot does not draw same X range despite the same PlotRange
Can I hire several veteran soldiers to accompany me?
How can I change my buffer system for protein purification?
How to model a Coral or Sponge Structure?
What verb goes with "coup"?
What was the point of separating stdout and stderr?
Available snapshots for main net?
Does "boire un jus" tend to mean "coffee" or "juice of fruit"?
Are the plates of a battery really charged?
2019 2-letters 33-length list
What is the meaning of ゴト in the context of 鮎
What is the meaning of "it" in "as luck would have it"?
Avoiding repetition when using the "snprintf idiom" to write text
Installed software from source, how to say yum not to install it from package?
Why are symbols not written in words?
How far can gerrymandering go?
How does entropy depend on location and scale?
Entropy of Cauchy (Lorentz) DistributionDeriving Negentropy. Getting stuck[Revised]Proving the expected bolddensity of being the Nth order statistics is decreasing in sample sizeRelationship between least-squares regression and information theoryAsymptotic relation for a class of probability distribution functionsLet $X_(1)leq X_(2)$ be the order statistics. Evaluate $operatornameVar(X_(j))$, $operatornameCov(X_(1),X_(2))$Why we multiply the density instead of taking the integral?Is continuous mutual information the correct analogue of the discrete version?What does the integral of a function times a function of a random variable represent, conceptually?If $Xsim operatornamelognormal$ then $Y:=(X-dmid xgeq d)$ has approximately a Generalized Pareto distributionSequence of shifted exponential distributions has uniform conditionals?
$begingroup$
The entropy of a continuous distribution with density function $f$ is defined to be the negative of the expectation of $log(f),$ and therefore equals
$$H_f = -int_-infty^infty log(f(x)) f(x)mathrmdx.$$
We also say that any random variable $X$ whose distribution has density $f$ has entropy $H_f.$ (This integral is well-defined even when $f$ has zeros, because $log(f(x))f(x)$ can be taken to equal zero at such values.)
When $X$ and $Y$ are random variables for which $Y = X+mu$ ($mu$ is a constant), $Y$ is said to be a version of $X$ shifted by $mu.$ Similarly, when $Y = Xsigma$ ($sigma$ is a positive constant), $Y$ is said to be a version of $X$ scaled by $sigma.$ Combining a scale with a shift gives $Y=Xsigma + mu.$
These relations occur frequently. For instance, changing the units of measurement of $X$ shifts and scales it.
How is the entropy of $Y = Xsigma + mu$ related to that of $X?$
distributions data-transformation entropy
$endgroup$
add a comment |
$begingroup$
The entropy of a continuous distribution with density function $f$ is defined to be the negative of the expectation of $log(f),$ and therefore equals
$$H_f = -int_-infty^infty log(f(x)) f(x)mathrmdx.$$
We also say that any random variable $X$ whose distribution has density $f$ has entropy $H_f.$ (This integral is well-defined even when $f$ has zeros, because $log(f(x))f(x)$ can be taken to equal zero at such values.)
When $X$ and $Y$ are random variables for which $Y = X+mu$ ($mu$ is a constant), $Y$ is said to be a version of $X$ shifted by $mu.$ Similarly, when $Y = Xsigma$ ($sigma$ is a positive constant), $Y$ is said to be a version of $X$ scaled by $sigma.$ Combining a scale with a shift gives $Y=Xsigma + mu.$
These relations occur frequently. For instance, changing the units of measurement of $X$ shifts and scales it.
How is the entropy of $Y = Xsigma + mu$ related to that of $X?$
distributions data-transformation entropy
$endgroup$
add a comment |
$begingroup$
The entropy of a continuous distribution with density function $f$ is defined to be the negative of the expectation of $log(f),$ and therefore equals
$$H_f = -int_-infty^infty log(f(x)) f(x)mathrmdx.$$
We also say that any random variable $X$ whose distribution has density $f$ has entropy $H_f.$ (This integral is well-defined even when $f$ has zeros, because $log(f(x))f(x)$ can be taken to equal zero at such values.)
When $X$ and $Y$ are random variables for which $Y = X+mu$ ($mu$ is a constant), $Y$ is said to be a version of $X$ shifted by $mu.$ Similarly, when $Y = Xsigma$ ($sigma$ is a positive constant), $Y$ is said to be a version of $X$ scaled by $sigma.$ Combining a scale with a shift gives $Y=Xsigma + mu.$
These relations occur frequently. For instance, changing the units of measurement of $X$ shifts and scales it.
How is the entropy of $Y = Xsigma + mu$ related to that of $X?$
distributions data-transformation entropy
$endgroup$
The entropy of a continuous distribution with density function $f$ is defined to be the negative of the expectation of $log(f),$ and therefore equals
$$H_f = -int_-infty^infty log(f(x)) f(x)mathrmdx.$$
We also say that any random variable $X$ whose distribution has density $f$ has entropy $H_f.$ (This integral is well-defined even when $f$ has zeros, because $log(f(x))f(x)$ can be taken to equal zero at such values.)
When $X$ and $Y$ are random variables for which $Y = X+mu$ ($mu$ is a constant), $Y$ is said to be a version of $X$ shifted by $mu.$ Similarly, when $Y = Xsigma$ ($sigma$ is a positive constant), $Y$ is said to be a version of $X$ scaled by $sigma.$ Combining a scale with a shift gives $Y=Xsigma + mu.$
These relations occur frequently. For instance, changing the units of measurement of $X$ shifts and scales it.
How is the entropy of $Y = Xsigma + mu$ related to that of $X?$
distributions data-transformation entropy
distributions data-transformation entropy
asked 9 hours ago
whuber♦whuber
212k34 gold badges465 silver badges847 bronze badges
212k34 gold badges465 silver badges847 bronze badges
add a comment |
add a comment |
1 Answer
1
active
oldest
votes
$begingroup$
Since the probability element of $X$ is $f(x)mathrmdx,$ the change of variable $y = xsigma + mu$ is equivalent to $x = (y-mu)/sigma,$ whence from
$$f(x)mathrmdx = fleft(fracy-musigmaright)mathrmdleft(fracy-musigmaright) = frac1sigma fleft(fracy-musigmaright) mathrmdy$$
it follows that the density of $Y$ is
$$f_Y(y) = frac1sigmafleft(fracy-musigmaright).$$
Consequently the entropy of $Y$ is
$$H(Y) = -int_-infty^infty logleft(frac1sigmafleft(fracy-musigmaright)right) frac1sigmafleft(fracy-musigmaright) mathrmdy$$
which, upon changing the variable back to $x = (y-mu)/sigma,$ produces
$$eqalign
H(Y) &= -int_-infty^infty logleft(frac1sigmafleft(xright)right) fleft(xright) mathrmdx \
&= -int_-infty^infty left(logleft(frac1sigmaright) + logleft(fleft(xright)right)right) fleft(xright) mathrmdx \
&= logleft(sigmaright) int_-infty^infty f(x) mathrmdx -int_-infty^infty logleft(fleft(xright)right) fleft(xright) mathrmdx \
&= log(sigma) + H_f.
$$
These calculations used basic properties of the logarithm, the linearity of integration, and that fact that $f(x)mathrmdx$ integrates to unity (the Law of Total Probability).
The conclusion is
The entropy of $Y = Xsigma + mu$ is the entropy of $X$ plus $log(sigma).$
In words, shifting a random variable does not change its entropy (we may think of the entropy as depending on the values of the probability density, but not on where those values occur), while scaling a variable (which, for $sigma ge 1$ "stretches" or "smears" it out) increases its entropy by $log(sigma).$ This supports the intuition that high-entropy distributions are "more spread out" than low-entropy distributions.
As a consequence of this result, we are free to choose convenient values of $mu$ and $sigma$ when computing the entropy of any distribution. For example, the entropy of a Normal$(mu,sigma)$ distribution can be found by setting $mu=0$ and $sigma=1.$ The logarithm of the density in this case is
$$log(f(x)) = -frac12log(2pi) - x^2/2,$$
whence
$$H = -E[-frac12log(2pi) - X^2/2] = frac12log(2pi) + frac12.$$
Consequently the entropy of a Normal$(mu,sigma)$ distribution is obtained simply by adding $logsigma$ to this result, giving
$$H = frac12log(2pi) + frac12 + log(sigma) = frac12log(2pi,e,sigma^2)$$
as reported by Wikipedia.
$endgroup$
add a comment |
Your Answer
StackExchange.ready(function()
var channelOptions =
tags: "".split(" "),
id: "65"
;
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function()
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled)
StackExchange.using("snippets", function()
createEditor();
);
else
createEditor();
);
function createEditor()
StackExchange.prepareEditor(
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: false,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: null,
bindNavPrevention: true,
postfix: "",
imageUploader:
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
,
onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
);
);
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f415435%2fhow-does-entropy-depend-on-location-and-scale%23new-answer', 'question_page');
);
Post as a guest
Required, but never shown
1 Answer
1
active
oldest
votes
1 Answer
1
active
oldest
votes
active
oldest
votes
active
oldest
votes
$begingroup$
Since the probability element of $X$ is $f(x)mathrmdx,$ the change of variable $y = xsigma + mu$ is equivalent to $x = (y-mu)/sigma,$ whence from
$$f(x)mathrmdx = fleft(fracy-musigmaright)mathrmdleft(fracy-musigmaright) = frac1sigma fleft(fracy-musigmaright) mathrmdy$$
it follows that the density of $Y$ is
$$f_Y(y) = frac1sigmafleft(fracy-musigmaright).$$
Consequently the entropy of $Y$ is
$$H(Y) = -int_-infty^infty logleft(frac1sigmafleft(fracy-musigmaright)right) frac1sigmafleft(fracy-musigmaright) mathrmdy$$
which, upon changing the variable back to $x = (y-mu)/sigma,$ produces
$$eqalign
H(Y) &= -int_-infty^infty logleft(frac1sigmafleft(xright)right) fleft(xright) mathrmdx \
&= -int_-infty^infty left(logleft(frac1sigmaright) + logleft(fleft(xright)right)right) fleft(xright) mathrmdx \
&= logleft(sigmaright) int_-infty^infty f(x) mathrmdx -int_-infty^infty logleft(fleft(xright)right) fleft(xright) mathrmdx \
&= log(sigma) + H_f.
$$
These calculations used basic properties of the logarithm, the linearity of integration, and that fact that $f(x)mathrmdx$ integrates to unity (the Law of Total Probability).
The conclusion is
The entropy of $Y = Xsigma + mu$ is the entropy of $X$ plus $log(sigma).$
In words, shifting a random variable does not change its entropy (we may think of the entropy as depending on the values of the probability density, but not on where those values occur), while scaling a variable (which, for $sigma ge 1$ "stretches" or "smears" it out) increases its entropy by $log(sigma).$ This supports the intuition that high-entropy distributions are "more spread out" than low-entropy distributions.
As a consequence of this result, we are free to choose convenient values of $mu$ and $sigma$ when computing the entropy of any distribution. For example, the entropy of a Normal$(mu,sigma)$ distribution can be found by setting $mu=0$ and $sigma=1.$ The logarithm of the density in this case is
$$log(f(x)) = -frac12log(2pi) - x^2/2,$$
whence
$$H = -E[-frac12log(2pi) - X^2/2] = frac12log(2pi) + frac12.$$
Consequently the entropy of a Normal$(mu,sigma)$ distribution is obtained simply by adding $logsigma$ to this result, giving
$$H = frac12log(2pi) + frac12 + log(sigma) = frac12log(2pi,e,sigma^2)$$
as reported by Wikipedia.
$endgroup$
add a comment |
$begingroup$
Since the probability element of $X$ is $f(x)mathrmdx,$ the change of variable $y = xsigma + mu$ is equivalent to $x = (y-mu)/sigma,$ whence from
$$f(x)mathrmdx = fleft(fracy-musigmaright)mathrmdleft(fracy-musigmaright) = frac1sigma fleft(fracy-musigmaright) mathrmdy$$
it follows that the density of $Y$ is
$$f_Y(y) = frac1sigmafleft(fracy-musigmaright).$$
Consequently the entropy of $Y$ is
$$H(Y) = -int_-infty^infty logleft(frac1sigmafleft(fracy-musigmaright)right) frac1sigmafleft(fracy-musigmaright) mathrmdy$$
which, upon changing the variable back to $x = (y-mu)/sigma,$ produces
$$eqalign
H(Y) &= -int_-infty^infty logleft(frac1sigmafleft(xright)right) fleft(xright) mathrmdx \
&= -int_-infty^infty left(logleft(frac1sigmaright) + logleft(fleft(xright)right)right) fleft(xright) mathrmdx \
&= logleft(sigmaright) int_-infty^infty f(x) mathrmdx -int_-infty^infty logleft(fleft(xright)right) fleft(xright) mathrmdx \
&= log(sigma) + H_f.
$$
These calculations used basic properties of the logarithm, the linearity of integration, and that fact that $f(x)mathrmdx$ integrates to unity (the Law of Total Probability).
The conclusion is
The entropy of $Y = Xsigma + mu$ is the entropy of $X$ plus $log(sigma).$
In words, shifting a random variable does not change its entropy (we may think of the entropy as depending on the values of the probability density, but not on where those values occur), while scaling a variable (which, for $sigma ge 1$ "stretches" or "smears" it out) increases its entropy by $log(sigma).$ This supports the intuition that high-entropy distributions are "more spread out" than low-entropy distributions.
As a consequence of this result, we are free to choose convenient values of $mu$ and $sigma$ when computing the entropy of any distribution. For example, the entropy of a Normal$(mu,sigma)$ distribution can be found by setting $mu=0$ and $sigma=1.$ The logarithm of the density in this case is
$$log(f(x)) = -frac12log(2pi) - x^2/2,$$
whence
$$H = -E[-frac12log(2pi) - X^2/2] = frac12log(2pi) + frac12.$$
Consequently the entropy of a Normal$(mu,sigma)$ distribution is obtained simply by adding $logsigma$ to this result, giving
$$H = frac12log(2pi) + frac12 + log(sigma) = frac12log(2pi,e,sigma^2)$$
as reported by Wikipedia.
$endgroup$
add a comment |
$begingroup$
Since the probability element of $X$ is $f(x)mathrmdx,$ the change of variable $y = xsigma + mu$ is equivalent to $x = (y-mu)/sigma,$ whence from
$$f(x)mathrmdx = fleft(fracy-musigmaright)mathrmdleft(fracy-musigmaright) = frac1sigma fleft(fracy-musigmaright) mathrmdy$$
it follows that the density of $Y$ is
$$f_Y(y) = frac1sigmafleft(fracy-musigmaright).$$
Consequently the entropy of $Y$ is
$$H(Y) = -int_-infty^infty logleft(frac1sigmafleft(fracy-musigmaright)right) frac1sigmafleft(fracy-musigmaright) mathrmdy$$
which, upon changing the variable back to $x = (y-mu)/sigma,$ produces
$$eqalign
H(Y) &= -int_-infty^infty logleft(frac1sigmafleft(xright)right) fleft(xright) mathrmdx \
&= -int_-infty^infty left(logleft(frac1sigmaright) + logleft(fleft(xright)right)right) fleft(xright) mathrmdx \
&= logleft(sigmaright) int_-infty^infty f(x) mathrmdx -int_-infty^infty logleft(fleft(xright)right) fleft(xright) mathrmdx \
&= log(sigma) + H_f.
$$
These calculations used basic properties of the logarithm, the linearity of integration, and that fact that $f(x)mathrmdx$ integrates to unity (the Law of Total Probability).
The conclusion is
The entropy of $Y = Xsigma + mu$ is the entropy of $X$ plus $log(sigma).$
In words, shifting a random variable does not change its entropy (we may think of the entropy as depending on the values of the probability density, but not on where those values occur), while scaling a variable (which, for $sigma ge 1$ "stretches" or "smears" it out) increases its entropy by $log(sigma).$ This supports the intuition that high-entropy distributions are "more spread out" than low-entropy distributions.
As a consequence of this result, we are free to choose convenient values of $mu$ and $sigma$ when computing the entropy of any distribution. For example, the entropy of a Normal$(mu,sigma)$ distribution can be found by setting $mu=0$ and $sigma=1.$ The logarithm of the density in this case is
$$log(f(x)) = -frac12log(2pi) - x^2/2,$$
whence
$$H = -E[-frac12log(2pi) - X^2/2] = frac12log(2pi) + frac12.$$
Consequently the entropy of a Normal$(mu,sigma)$ distribution is obtained simply by adding $logsigma$ to this result, giving
$$H = frac12log(2pi) + frac12 + log(sigma) = frac12log(2pi,e,sigma^2)$$
as reported by Wikipedia.
$endgroup$
Since the probability element of $X$ is $f(x)mathrmdx,$ the change of variable $y = xsigma + mu$ is equivalent to $x = (y-mu)/sigma,$ whence from
$$f(x)mathrmdx = fleft(fracy-musigmaright)mathrmdleft(fracy-musigmaright) = frac1sigma fleft(fracy-musigmaright) mathrmdy$$
it follows that the density of $Y$ is
$$f_Y(y) = frac1sigmafleft(fracy-musigmaright).$$
Consequently the entropy of $Y$ is
$$H(Y) = -int_-infty^infty logleft(frac1sigmafleft(fracy-musigmaright)right) frac1sigmafleft(fracy-musigmaright) mathrmdy$$
which, upon changing the variable back to $x = (y-mu)/sigma,$ produces
$$eqalign
H(Y) &= -int_-infty^infty logleft(frac1sigmafleft(xright)right) fleft(xright) mathrmdx \
&= -int_-infty^infty left(logleft(frac1sigmaright) + logleft(fleft(xright)right)right) fleft(xright) mathrmdx \
&= logleft(sigmaright) int_-infty^infty f(x) mathrmdx -int_-infty^infty logleft(fleft(xright)right) fleft(xright) mathrmdx \
&= log(sigma) + H_f.
$$
These calculations used basic properties of the logarithm, the linearity of integration, and that fact that $f(x)mathrmdx$ integrates to unity (the Law of Total Probability).
The conclusion is
The entropy of $Y = Xsigma + mu$ is the entropy of $X$ plus $log(sigma).$
In words, shifting a random variable does not change its entropy (we may think of the entropy as depending on the values of the probability density, but not on where those values occur), while scaling a variable (which, for $sigma ge 1$ "stretches" or "smears" it out) increases its entropy by $log(sigma).$ This supports the intuition that high-entropy distributions are "more spread out" than low-entropy distributions.
As a consequence of this result, we are free to choose convenient values of $mu$ and $sigma$ when computing the entropy of any distribution. For example, the entropy of a Normal$(mu,sigma)$ distribution can be found by setting $mu=0$ and $sigma=1.$ The logarithm of the density in this case is
$$log(f(x)) = -frac12log(2pi) - x^2/2,$$
whence
$$H = -E[-frac12log(2pi) - X^2/2] = frac12log(2pi) + frac12.$$
Consequently the entropy of a Normal$(mu,sigma)$ distribution is obtained simply by adding $logsigma$ to this result, giving
$$H = frac12log(2pi) + frac12 + log(sigma) = frac12log(2pi,e,sigma^2)$$
as reported by Wikipedia.
answered 9 hours ago
whuber♦whuber
212k34 gold badges465 silver badges847 bronze badges
212k34 gold badges465 silver badges847 bronze badges
add a comment |
add a comment |
Thanks for contributing an answer to Cross Validated!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f415435%2fhow-does-entropy-depend-on-location-and-scale%23new-answer', 'question_page');
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown