by reqo on 6/21/24, 10:54 PM with 61 comments
by eru on 6/26/24, 3:25 AM
Hmm, that sentence at the beginning is already wrong. Random variables can measure anything, not just numbers. Heads or Tails of a coin, or colours of cars etc.
It's fine to restrict yourself to numeric random variables only. But if you are writing a rant telling other people to be more careful in their analysis, you better dot your i's and cross your t's yourself.
by auraai on 6/26/24, 5:49 AM
https://en.wikipedia.org/wiki/Girsanov_theorem https://en.wikipedia.org/wiki/Risk-neutral_measure
by btown on 6/26/24, 7:00 AM
The fun part comes when you can ask this computation graph: "what parameters for a random variable early on in the chain would be the ones that optimize some function of variables later in the chain?" And, handwaving a ton of nuance here, when those parameters are weights in a neural network, the function is a loss function on the training data, and the optimization is done by automatic differentiation (e.g. https://pytorch.org/tutorials/beginner/introyt/autogradyt_tu...), you have modern AI.
If you're interested in the theoretical underpinnings here, Bishop's PRML is perhaps the classic starting point: https://www.microsoft.com/en-us/research/uploads/prod/2006/0...
by panic on 6/26/24, 6:36 AM
In that article, squaring a number in interval arithmetic is different from multiplying two independent numbers with the same interval. Here, squaring a random variable is different from multiplying two independent random variables with the same distribution.
by condwanaland on 6/26/24, 2:19 AM
by jhrmnn on 6/26/24, 4:18 AM
by kazinator on 6/26/24, 6:46 AM
by dwqwdqd on 6/26/24, 2:10 AM
X = 1 with probability 0.5, 0 with probability 0.5 Y = 0 when X = 1, 1 when X = 0 (for the \omega for which X(\omega) = 1, Y(\omega) = 0).
They're both bernoulli distributions with p=0.5 (i.e. they follow the same distribution) and P(X=Y) = 0
by baking on 6/26/24, 1:31 AM
by Davidzheng on 6/26/24, 2:47 AM
by tpoacher on 6/27/24, 8:18 PM
A better convention would make clear the distinction between the set of possible outcomes, the act of obtaining a (range of) samples from that set, and the probability that those events match a value range of interest. p(x=X) is not enough to capture all that information. let alone p(x) vs p(X).
by clircle on 6/26/24, 2:09 AM
by dinobones on 6/26/24, 5:14 AM
I’m so glad I have ChatGPT now, I always ask for applied examples and ask it to explain things intuitively. I would’ve been a 4.0 student if I would’ve had ChatGPT as my personal tutor when I was in school.
by glitchc on 6/26/24, 4:20 AM