What does phi mean in this context?











up vote
1
down vote

favorite












In an article on Selection Bias in A/B Testing, AirBnB proposes a solution to their own estimated bias. This solution is to subtract from the aggregated effect a bias estimate, captured by this equation:
$$
hat{beta} = sum_{i=1}^n W_i phi left( frac{W_i b_i - X_i}{W_i} right)
$$

where




$X_1, ldots, X_n$ are random variables defined on a same probability space, and each $X_i$ follows a distribution with finite mean $a_i$ and finite variance $sigma_i^2$ (the distributions are not necessarily identical.) We regard $a_i$ as the unknown true effect and usually estimate it by the unbiased estimate $X_i$;



$b_i$ is the cut-off from the reference distribution for significance level $alpha_i$, usually set at $0.05$; and



$W_i$ is the estimated standard deviation of $X_i$, to define the bias estimate




In this context, what does $phi$ stand for? My current understanding (by separating the two terms inside the parenthesis into
$$
b_i - X_i / W_i
$$

is that they're calculating the individual bias estimates as the difference between the cutoff and "how many" standard deviations fit into the estimate. Then they're adding these up, but I don't know what $W_iphi$ is doing inside the sum.










share|cite|improve this question
























  • The phi coefficient is a correlation between two variables.
    – Phil H
    Oct 2 at 0:38










  • So, the equation states that Wi is correlated to the expression in parenthesis, but it does not affect the actual sum?
    – Khashir
    Oct 3 at 4:53















up vote
1
down vote

favorite












In an article on Selection Bias in A/B Testing, AirBnB proposes a solution to their own estimated bias. This solution is to subtract from the aggregated effect a bias estimate, captured by this equation:
$$
hat{beta} = sum_{i=1}^n W_i phi left( frac{W_i b_i - X_i}{W_i} right)
$$

where




$X_1, ldots, X_n$ are random variables defined on a same probability space, and each $X_i$ follows a distribution with finite mean $a_i$ and finite variance $sigma_i^2$ (the distributions are not necessarily identical.) We regard $a_i$ as the unknown true effect and usually estimate it by the unbiased estimate $X_i$;



$b_i$ is the cut-off from the reference distribution for significance level $alpha_i$, usually set at $0.05$; and



$W_i$ is the estimated standard deviation of $X_i$, to define the bias estimate




In this context, what does $phi$ stand for? My current understanding (by separating the two terms inside the parenthesis into
$$
b_i - X_i / W_i
$$

is that they're calculating the individual bias estimates as the difference between the cutoff and "how many" standard deviations fit into the estimate. Then they're adding these up, but I don't know what $W_iphi$ is doing inside the sum.










share|cite|improve this question
























  • The phi coefficient is a correlation between two variables.
    – Phil H
    Oct 2 at 0:38










  • So, the equation states that Wi is correlated to the expression in parenthesis, but it does not affect the actual sum?
    – Khashir
    Oct 3 at 4:53













up vote
1
down vote

favorite









up vote
1
down vote

favorite











In an article on Selection Bias in A/B Testing, AirBnB proposes a solution to their own estimated bias. This solution is to subtract from the aggregated effect a bias estimate, captured by this equation:
$$
hat{beta} = sum_{i=1}^n W_i phi left( frac{W_i b_i - X_i}{W_i} right)
$$

where




$X_1, ldots, X_n$ are random variables defined on a same probability space, and each $X_i$ follows a distribution with finite mean $a_i$ and finite variance $sigma_i^2$ (the distributions are not necessarily identical.) We regard $a_i$ as the unknown true effect and usually estimate it by the unbiased estimate $X_i$;



$b_i$ is the cut-off from the reference distribution for significance level $alpha_i$, usually set at $0.05$; and



$W_i$ is the estimated standard deviation of $X_i$, to define the bias estimate




In this context, what does $phi$ stand for? My current understanding (by separating the two terms inside the parenthesis into
$$
b_i - X_i / W_i
$$

is that they're calculating the individual bias estimates as the difference between the cutoff and "how many" standard deviations fit into the estimate. Then they're adding these up, but I don't know what $W_iphi$ is doing inside the sum.










share|cite|improve this question















In an article on Selection Bias in A/B Testing, AirBnB proposes a solution to their own estimated bias. This solution is to subtract from the aggregated effect a bias estimate, captured by this equation:
$$
hat{beta} = sum_{i=1}^n W_i phi left( frac{W_i b_i - X_i}{W_i} right)
$$

where




$X_1, ldots, X_n$ are random variables defined on a same probability space, and each $X_i$ follows a distribution with finite mean $a_i$ and finite variance $sigma_i^2$ (the distributions are not necessarily identical.) We regard $a_i$ as the unknown true effect and usually estimate it by the unbiased estimate $X_i$;



$b_i$ is the cut-off from the reference distribution for significance level $alpha_i$, usually set at $0.05$; and



$W_i$ is the estimated standard deviation of $X_i$, to define the bias estimate




In this context, what does $phi$ stand for? My current understanding (by separating the two terms inside the parenthesis into
$$
b_i - X_i / W_i
$$

is that they're calculating the individual bias estimates as the difference between the cutoff and "how many" standard deviations fit into the estimate. Then they're adding these up, but I don't know what $W_iphi$ is doing inside the sum.







probability-theory probability-distributions






share|cite|improve this question















share|cite|improve this question













share|cite|improve this question




share|cite|improve this question








edited Oct 2 at 21:33









mwt

939416




939416










asked Oct 1 at 23:49









Khashir

1062




1062












  • The phi coefficient is a correlation between two variables.
    – Phil H
    Oct 2 at 0:38










  • So, the equation states that Wi is correlated to the expression in parenthesis, but it does not affect the actual sum?
    – Khashir
    Oct 3 at 4:53


















  • The phi coefficient is a correlation between two variables.
    – Phil H
    Oct 2 at 0:38










  • So, the equation states that Wi is correlated to the expression in parenthesis, but it does not affect the actual sum?
    – Khashir
    Oct 3 at 4:53
















The phi coefficient is a correlation between two variables.
– Phil H
Oct 2 at 0:38




The phi coefficient is a correlation between two variables.
– Phil H
Oct 2 at 0:38












So, the equation states that Wi is correlated to the expression in parenthesis, but it does not affect the actual sum?
– Khashir
Oct 3 at 4:53




So, the equation states that Wi is correlated to the expression in parenthesis, but it does not affect the actual sum?
– Khashir
Oct 3 at 4:53










1 Answer
1






active

oldest

votes

















up vote
0
down vote













According to their paper, $phi$ refers to the density function of standard normal distribution.



Let me include some details:



They use the following result:




Suppose $Z$ follows standard normal distribution, then



$$f_{Z|Zge t}(s) = frac{phi(s)}{1-Phi(t)}$$
and hence
$$E[Z|Z ge t]=frac{int_t^infty sphi(s), ds}{1-Phi(t)}=frac{-int_t^infty phi'(s), ds}{1-Phi(t)}=frac{phi(t)}{1-Phi(t)}$$




Assuming $X_i$ follows normal distribution $N(a_i, sigma_i^2)$.
begin{align}
&beta = E[S_A-T_A] \&=sum_{i=1}^nE[I((X_i-a_i)>(b_isigma_i-a_i))(X_i-a_i)]\
&= sum_{i=1}^n sigma_i Eleft[Ileft(frac{X_i-a_i}{sigma_i}>frac{b_isigma_i-a_i}{sigma_i}right)frac{X_i-a_i}{sigma_i}right]\
&=sum_{i=1}^n sigma_iPleft(frac{X_i-a_i}{sigma_i}>frac{b_isigma_i-a_i}{sigma_i}right)Eleft[frac{X_i-a_i}{sigma_i}left|frac{X_i-a_i}{sigma}>frac{b_isigma_i-a_i}{sigma}right.right]\
&=sum_{i=1}^n sigma_i Pleft(frac{X_i-a_i}{sigma_i}>frac{b_isigma_i-a_i}{sigma_i}right)frac{phileft( frac{b_isigma_i-a_i}{sigma_i}right)}{Pleft(frac{X_i-a_i}{sigma_i}>frac{b_isigma_i-a_i}{sigma_i}right)} \
&=sum_{i=1}^n sigma_i phileft( frac{b_isigma_i-a_i}{sigma_i}right)\
end{align}



They then replace some parameters via estimation.






share|cite|improve this answer























    Your Answer





    StackExchange.ifUsing("editor", function () {
    return StackExchange.using("mathjaxEditing", function () {
    StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
    StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
    });
    });
    }, "mathjax-editing");

    StackExchange.ready(function() {
    var channelOptions = {
    tags: "".split(" "),
    id: "69"
    };
    initTagRenderer("".split(" "), "".split(" "), channelOptions);

    StackExchange.using("externalEditor", function() {
    // Have to fire editor after snippets, if snippets enabled
    if (StackExchange.settings.snippets.snippetsEnabled) {
    StackExchange.using("snippets", function() {
    createEditor();
    });
    }
    else {
    createEditor();
    }
    });

    function createEditor() {
    StackExchange.prepareEditor({
    heartbeatType: 'answer',
    convertImagesToLinks: true,
    noModals: true,
    showLowRepImageUploadWarning: true,
    reputationToPostImages: 10,
    bindNavPrevention: true,
    postfix: "",
    imageUploader: {
    brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
    contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
    allowUrls: true
    },
    noCode: true, onDemand: true,
    discardSelector: ".discard-answer"
    ,immediatelyShowMarkdownHelp:true
    });


    }
    });














    draft saved

    draft discarded


















    StackExchange.ready(
    function () {
    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f2938671%2fwhat-does-phi-mean-in-this-context%23new-answer', 'question_page');
    }
    );

    Post as a guest















    Required, but never shown

























    1 Answer
    1






    active

    oldest

    votes








    1 Answer
    1






    active

    oldest

    votes









    active

    oldest

    votes






    active

    oldest

    votes








    up vote
    0
    down vote













    According to their paper, $phi$ refers to the density function of standard normal distribution.



    Let me include some details:



    They use the following result:




    Suppose $Z$ follows standard normal distribution, then



    $$f_{Z|Zge t}(s) = frac{phi(s)}{1-Phi(t)}$$
    and hence
    $$E[Z|Z ge t]=frac{int_t^infty sphi(s), ds}{1-Phi(t)}=frac{-int_t^infty phi'(s), ds}{1-Phi(t)}=frac{phi(t)}{1-Phi(t)}$$




    Assuming $X_i$ follows normal distribution $N(a_i, sigma_i^2)$.
    begin{align}
    &beta = E[S_A-T_A] \&=sum_{i=1}^nE[I((X_i-a_i)>(b_isigma_i-a_i))(X_i-a_i)]\
    &= sum_{i=1}^n sigma_i Eleft[Ileft(frac{X_i-a_i}{sigma_i}>frac{b_isigma_i-a_i}{sigma_i}right)frac{X_i-a_i}{sigma_i}right]\
    &=sum_{i=1}^n sigma_iPleft(frac{X_i-a_i}{sigma_i}>frac{b_isigma_i-a_i}{sigma_i}right)Eleft[frac{X_i-a_i}{sigma_i}left|frac{X_i-a_i}{sigma}>frac{b_isigma_i-a_i}{sigma}right.right]\
    &=sum_{i=1}^n sigma_i Pleft(frac{X_i-a_i}{sigma_i}>frac{b_isigma_i-a_i}{sigma_i}right)frac{phileft( frac{b_isigma_i-a_i}{sigma_i}right)}{Pleft(frac{X_i-a_i}{sigma_i}>frac{b_isigma_i-a_i}{sigma_i}right)} \
    &=sum_{i=1}^n sigma_i phileft( frac{b_isigma_i-a_i}{sigma_i}right)\
    end{align}



    They then replace some parameters via estimation.






    share|cite|improve this answer



























      up vote
      0
      down vote













      According to their paper, $phi$ refers to the density function of standard normal distribution.



      Let me include some details:



      They use the following result:




      Suppose $Z$ follows standard normal distribution, then



      $$f_{Z|Zge t}(s) = frac{phi(s)}{1-Phi(t)}$$
      and hence
      $$E[Z|Z ge t]=frac{int_t^infty sphi(s), ds}{1-Phi(t)}=frac{-int_t^infty phi'(s), ds}{1-Phi(t)}=frac{phi(t)}{1-Phi(t)}$$




      Assuming $X_i$ follows normal distribution $N(a_i, sigma_i^2)$.
      begin{align}
      &beta = E[S_A-T_A] \&=sum_{i=1}^nE[I((X_i-a_i)>(b_isigma_i-a_i))(X_i-a_i)]\
      &= sum_{i=1}^n sigma_i Eleft[Ileft(frac{X_i-a_i}{sigma_i}>frac{b_isigma_i-a_i}{sigma_i}right)frac{X_i-a_i}{sigma_i}right]\
      &=sum_{i=1}^n sigma_iPleft(frac{X_i-a_i}{sigma_i}>frac{b_isigma_i-a_i}{sigma_i}right)Eleft[frac{X_i-a_i}{sigma_i}left|frac{X_i-a_i}{sigma}>frac{b_isigma_i-a_i}{sigma}right.right]\
      &=sum_{i=1}^n sigma_i Pleft(frac{X_i-a_i}{sigma_i}>frac{b_isigma_i-a_i}{sigma_i}right)frac{phileft( frac{b_isigma_i-a_i}{sigma_i}right)}{Pleft(frac{X_i-a_i}{sigma_i}>frac{b_isigma_i-a_i}{sigma_i}right)} \
      &=sum_{i=1}^n sigma_i phileft( frac{b_isigma_i-a_i}{sigma_i}right)\
      end{align}



      They then replace some parameters via estimation.






      share|cite|improve this answer

























        up vote
        0
        down vote










        up vote
        0
        down vote









        According to their paper, $phi$ refers to the density function of standard normal distribution.



        Let me include some details:



        They use the following result:




        Suppose $Z$ follows standard normal distribution, then



        $$f_{Z|Zge t}(s) = frac{phi(s)}{1-Phi(t)}$$
        and hence
        $$E[Z|Z ge t]=frac{int_t^infty sphi(s), ds}{1-Phi(t)}=frac{-int_t^infty phi'(s), ds}{1-Phi(t)}=frac{phi(t)}{1-Phi(t)}$$




        Assuming $X_i$ follows normal distribution $N(a_i, sigma_i^2)$.
        begin{align}
        &beta = E[S_A-T_A] \&=sum_{i=1}^nE[I((X_i-a_i)>(b_isigma_i-a_i))(X_i-a_i)]\
        &= sum_{i=1}^n sigma_i Eleft[Ileft(frac{X_i-a_i}{sigma_i}>frac{b_isigma_i-a_i}{sigma_i}right)frac{X_i-a_i}{sigma_i}right]\
        &=sum_{i=1}^n sigma_iPleft(frac{X_i-a_i}{sigma_i}>frac{b_isigma_i-a_i}{sigma_i}right)Eleft[frac{X_i-a_i}{sigma_i}left|frac{X_i-a_i}{sigma}>frac{b_isigma_i-a_i}{sigma}right.right]\
        &=sum_{i=1}^n sigma_i Pleft(frac{X_i-a_i}{sigma_i}>frac{b_isigma_i-a_i}{sigma_i}right)frac{phileft( frac{b_isigma_i-a_i}{sigma_i}right)}{Pleft(frac{X_i-a_i}{sigma_i}>frac{b_isigma_i-a_i}{sigma_i}right)} \
        &=sum_{i=1}^n sigma_i phileft( frac{b_isigma_i-a_i}{sigma_i}right)\
        end{align}



        They then replace some parameters via estimation.






        share|cite|improve this answer














        According to their paper, $phi$ refers to the density function of standard normal distribution.



        Let me include some details:



        They use the following result:




        Suppose $Z$ follows standard normal distribution, then



        $$f_{Z|Zge t}(s) = frac{phi(s)}{1-Phi(t)}$$
        and hence
        $$E[Z|Z ge t]=frac{int_t^infty sphi(s), ds}{1-Phi(t)}=frac{-int_t^infty phi'(s), ds}{1-Phi(t)}=frac{phi(t)}{1-Phi(t)}$$




        Assuming $X_i$ follows normal distribution $N(a_i, sigma_i^2)$.
        begin{align}
        &beta = E[S_A-T_A] \&=sum_{i=1}^nE[I((X_i-a_i)>(b_isigma_i-a_i))(X_i-a_i)]\
        &= sum_{i=1}^n sigma_i Eleft[Ileft(frac{X_i-a_i}{sigma_i}>frac{b_isigma_i-a_i}{sigma_i}right)frac{X_i-a_i}{sigma_i}right]\
        &=sum_{i=1}^n sigma_iPleft(frac{X_i-a_i}{sigma_i}>frac{b_isigma_i-a_i}{sigma_i}right)Eleft[frac{X_i-a_i}{sigma_i}left|frac{X_i-a_i}{sigma}>frac{b_isigma_i-a_i}{sigma}right.right]\
        &=sum_{i=1}^n sigma_i Pleft(frac{X_i-a_i}{sigma_i}>frac{b_isigma_i-a_i}{sigma_i}right)frac{phileft( frac{b_isigma_i-a_i}{sigma_i}right)}{Pleft(frac{X_i-a_i}{sigma_i}>frac{b_isigma_i-a_i}{sigma_i}right)} \
        &=sum_{i=1}^n sigma_i phileft( frac{b_isigma_i-a_i}{sigma_i}right)\
        end{align}



        They then replace some parameters via estimation.







        share|cite|improve this answer














        share|cite|improve this answer



        share|cite|improve this answer








        edited 2 days ago

























        answered 2 days ago









        Siong Thye Goh

        95.4k1462116




        95.4k1462116






























            draft saved

            draft discarded




















































            Thanks for contributing an answer to Mathematics Stack Exchange!


            • Please be sure to answer the question. Provide details and share your research!

            But avoid



            • Asking for help, clarification, or responding to other answers.

            • Making statements based on opinion; back them up with references or personal experience.


            Use MathJax to format equations. MathJax reference.


            To learn more, see our tips on writing great answers.





            Some of your past answers have not been well-received, and you're in danger of being blocked from answering.


            Please pay close attention to the following guidance:


            • Please be sure to answer the question. Provide details and share your research!

            But avoid



            • Asking for help, clarification, or responding to other answers.

            • Making statements based on opinion; back them up with references or personal experience.


            To learn more, see our tips on writing great answers.




            draft saved


            draft discarded














            StackExchange.ready(
            function () {
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f2938671%2fwhat-does-phi-mean-in-this-context%23new-answer', 'question_page');
            }
            );

            Post as a guest















            Required, but never shown





















































            Required, but never shown














            Required, but never shown












            Required, but never shown







            Required, but never shown

































            Required, but never shown














            Required, but never shown












            Required, but never shown







            Required, but never shown







            Popular posts from this blog

            Bressuire

            Cabo Verde

            Gyllenstierna