Is it valid to iterate over every permutation of a regression specification and compute an “average significance?”

The name of the pictureThe name of the pictureThe name of the pictureClash Royale CLAN TAG#URR8PPP












2












$begingroup$


I had an idea, and was wondering if it's ever done and, if so, how to do it in an appropriate manner.



Let's say I run an OLS model and the results come back significant. There is discussion as to whether the positive results hold up given a different set of control variables. However, it's theoretically unclear which control variables are actually relevant, and we don't just want to throw in the kitchen sink.



Let's say it's computationally feasible to run every single model (with every possible permutation of control variables), and we collect the coefficient and standard error of the coefficient of interest in each. We can report the proportion of models where this coefficient is significant, but is there some kind of meta-test that can be done showing that the test statistic is significant in a significant number of alternative specifications? Is this ever done?










share|cite|improve this question









$endgroup$











  • $begingroup$
    I don't have the time to write up a full explanation right now, but you may want to look into Leamer bounds
    $endgroup$
    – duckmayr
    Jan 19 at 3:18










  • $begingroup$
    Leamer bounds was exactly what I was looking for. Would love you to go into more depth if you ever get time
    $endgroup$
    – Parseltongue
    Jan 22 at 0:12















2












$begingroup$


I had an idea, and was wondering if it's ever done and, if so, how to do it in an appropriate manner.



Let's say I run an OLS model and the results come back significant. There is discussion as to whether the positive results hold up given a different set of control variables. However, it's theoretically unclear which control variables are actually relevant, and we don't just want to throw in the kitchen sink.



Let's say it's computationally feasible to run every single model (with every possible permutation of control variables), and we collect the coefficient and standard error of the coefficient of interest in each. We can report the proportion of models where this coefficient is significant, but is there some kind of meta-test that can be done showing that the test statistic is significant in a significant number of alternative specifications? Is this ever done?










share|cite|improve this question









$endgroup$











  • $begingroup$
    I don't have the time to write up a full explanation right now, but you may want to look into Leamer bounds
    $endgroup$
    – duckmayr
    Jan 19 at 3:18










  • $begingroup$
    Leamer bounds was exactly what I was looking for. Would love you to go into more depth if you ever get time
    $endgroup$
    – Parseltongue
    Jan 22 at 0:12













2












2








2


1



$begingroup$


I had an idea, and was wondering if it's ever done and, if so, how to do it in an appropriate manner.



Let's say I run an OLS model and the results come back significant. There is discussion as to whether the positive results hold up given a different set of control variables. However, it's theoretically unclear which control variables are actually relevant, and we don't just want to throw in the kitchen sink.



Let's say it's computationally feasible to run every single model (with every possible permutation of control variables), and we collect the coefficient and standard error of the coefficient of interest in each. We can report the proportion of models where this coefficient is significant, but is there some kind of meta-test that can be done showing that the test statistic is significant in a significant number of alternative specifications? Is this ever done?










share|cite|improve this question









$endgroup$




I had an idea, and was wondering if it's ever done and, if so, how to do it in an appropriate manner.



Let's say I run an OLS model and the results come back significant. There is discussion as to whether the positive results hold up given a different set of control variables. However, it's theoretically unclear which control variables are actually relevant, and we don't just want to throw in the kitchen sink.



Let's say it's computationally feasible to run every single model (with every possible permutation of control variables), and we collect the coefficient and standard error of the coefficient of interest in each. We can report the proportion of models where this coefficient is significant, but is there some kind of meta-test that can be done showing that the test statistic is significant in a significant number of alternative specifications? Is this ever done?







regression least-squares






share|cite|improve this question













share|cite|improve this question











share|cite|improve this question




share|cite|improve this question










asked Jan 18 at 22:24









ParseltongueParseltongue

299114




299114











  • $begingroup$
    I don't have the time to write up a full explanation right now, but you may want to look into Leamer bounds
    $endgroup$
    – duckmayr
    Jan 19 at 3:18










  • $begingroup$
    Leamer bounds was exactly what I was looking for. Would love you to go into more depth if you ever get time
    $endgroup$
    – Parseltongue
    Jan 22 at 0:12
















  • $begingroup$
    I don't have the time to write up a full explanation right now, but you may want to look into Leamer bounds
    $endgroup$
    – duckmayr
    Jan 19 at 3:18










  • $begingroup$
    Leamer bounds was exactly what I was looking for. Would love you to go into more depth if you ever get time
    $endgroup$
    – Parseltongue
    Jan 22 at 0:12















$begingroup$
I don't have the time to write up a full explanation right now, but you may want to look into Leamer bounds
$endgroup$
– duckmayr
Jan 19 at 3:18




$begingroup$
I don't have the time to write up a full explanation right now, but you may want to look into Leamer bounds
$endgroup$
– duckmayr
Jan 19 at 3:18












$begingroup$
Leamer bounds was exactly what I was looking for. Would love you to go into more depth if you ever get time
$endgroup$
– Parseltongue
Jan 22 at 0:12




$begingroup$
Leamer bounds was exactly what I was looking for. Would love you to go into more depth if you ever get time
$endgroup$
– Parseltongue
Jan 22 at 0:12










1 Answer
1






active

oldest

votes


















4












$begingroup$

One similar practice is called model averaging. You fit many models, make many predictions, and average the results, weighting each by the posterior probability that the model is correct. There's a nice introduction here.



https://www2.stat.duke.edu/courses/Spring05/sta244/Handouts/press.pdf



I would advise you not to do this with coefficients, because it's not how BMA was meant to be used. The issue is that the same coefficient can have a different interpretation and target parameter in different models. Because of this, averaging two coefficients often doesn't make conceptual sense. Katharine Bannar explains via example: in a model for brain weight based on body size and gestation time, gestation time will have little association with brain weight once body size is already accounted for. If body size is left out of the model, the gestation coefficient would increase dramatically. There is more explanation here.



https://esajournals.onlinelibrary.wiley.com/doi/full/10.1002/eap.1419






share|cite|improve this answer









$endgroup$












    Your Answer





    StackExchange.ifUsing("editor", function ()
    return StackExchange.using("mathjaxEditing", function ()
    StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix)
    StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
    );
    );
    , "mathjax-editing");

    StackExchange.ready(function()
    var channelOptions =
    tags: "".split(" "),
    id: "65"
    ;
    initTagRenderer("".split(" "), "".split(" "), channelOptions);

    StackExchange.using("externalEditor", function()
    // Have to fire editor after snippets, if snippets enabled
    if (StackExchange.settings.snippets.snippetsEnabled)
    StackExchange.using("snippets", function()
    createEditor();
    );

    else
    createEditor();

    );

    function createEditor()
    StackExchange.prepareEditor(
    heartbeatType: 'answer',
    autoActivateHeartbeat: false,
    convertImagesToLinks: false,
    noModals: true,
    showLowRepImageUploadWarning: true,
    reputationToPostImages: null,
    bindNavPrevention: true,
    postfix: "",
    imageUploader:
    brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
    contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
    allowUrls: true
    ,
    onDemand: true,
    discardSelector: ".discard-answer"
    ,immediatelyShowMarkdownHelp:true
    );



    );













    draft saved

    draft discarded


















    StackExchange.ready(
    function ()
    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f387998%2fis-it-valid-to-iterate-over-every-permutation-of-a-regression-specification-and%23new-answer', 'question_page');

    );

    Post as a guest















    Required, but never shown

























    1 Answer
    1






    active

    oldest

    votes








    1 Answer
    1






    active

    oldest

    votes









    active

    oldest

    votes






    active

    oldest

    votes









    4












    $begingroup$

    One similar practice is called model averaging. You fit many models, make many predictions, and average the results, weighting each by the posterior probability that the model is correct. There's a nice introduction here.



    https://www2.stat.duke.edu/courses/Spring05/sta244/Handouts/press.pdf



    I would advise you not to do this with coefficients, because it's not how BMA was meant to be used. The issue is that the same coefficient can have a different interpretation and target parameter in different models. Because of this, averaging two coefficients often doesn't make conceptual sense. Katharine Bannar explains via example: in a model for brain weight based on body size and gestation time, gestation time will have little association with brain weight once body size is already accounted for. If body size is left out of the model, the gestation coefficient would increase dramatically. There is more explanation here.



    https://esajournals.onlinelibrary.wiley.com/doi/full/10.1002/eap.1419






    share|cite|improve this answer









    $endgroup$

















      4












      $begingroup$

      One similar practice is called model averaging. You fit many models, make many predictions, and average the results, weighting each by the posterior probability that the model is correct. There's a nice introduction here.



      https://www2.stat.duke.edu/courses/Spring05/sta244/Handouts/press.pdf



      I would advise you not to do this with coefficients, because it's not how BMA was meant to be used. The issue is that the same coefficient can have a different interpretation and target parameter in different models. Because of this, averaging two coefficients often doesn't make conceptual sense. Katharine Bannar explains via example: in a model for brain weight based on body size and gestation time, gestation time will have little association with brain weight once body size is already accounted for. If body size is left out of the model, the gestation coefficient would increase dramatically. There is more explanation here.



      https://esajournals.onlinelibrary.wiley.com/doi/full/10.1002/eap.1419






      share|cite|improve this answer









      $endgroup$















        4












        4








        4





        $begingroup$

        One similar practice is called model averaging. You fit many models, make many predictions, and average the results, weighting each by the posterior probability that the model is correct. There's a nice introduction here.



        https://www2.stat.duke.edu/courses/Spring05/sta244/Handouts/press.pdf



        I would advise you not to do this with coefficients, because it's not how BMA was meant to be used. The issue is that the same coefficient can have a different interpretation and target parameter in different models. Because of this, averaging two coefficients often doesn't make conceptual sense. Katharine Bannar explains via example: in a model for brain weight based on body size and gestation time, gestation time will have little association with brain weight once body size is already accounted for. If body size is left out of the model, the gestation coefficient would increase dramatically. There is more explanation here.



        https://esajournals.onlinelibrary.wiley.com/doi/full/10.1002/eap.1419






        share|cite|improve this answer









        $endgroup$



        One similar practice is called model averaging. You fit many models, make many predictions, and average the results, weighting each by the posterior probability that the model is correct. There's a nice introduction here.



        https://www2.stat.duke.edu/courses/Spring05/sta244/Handouts/press.pdf



        I would advise you not to do this with coefficients, because it's not how BMA was meant to be used. The issue is that the same coefficient can have a different interpretation and target parameter in different models. Because of this, averaging two coefficients often doesn't make conceptual sense. Katharine Bannar explains via example: in a model for brain weight based on body size and gestation time, gestation time will have little association with brain weight once body size is already accounted for. If body size is left out of the model, the gestation coefficient would increase dramatically. There is more explanation here.



        https://esajournals.onlinelibrary.wiley.com/doi/full/10.1002/eap.1419







        share|cite|improve this answer












        share|cite|improve this answer



        share|cite|improve this answer










        answered Jan 18 at 22:44









        eric_kernfelderic_kernfeld

        3,0231827




        3,0231827



























            draft saved

            draft discarded
















































            Thanks for contributing an answer to Cross Validated!


            • Please be sure to answer the question. Provide details and share your research!

            But avoid


            • Asking for help, clarification, or responding to other answers.

            • Making statements based on opinion; back them up with references or personal experience.

            Use MathJax to format equations. MathJax reference.


            To learn more, see our tips on writing great answers.




            draft saved


            draft discarded














            StackExchange.ready(
            function ()
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f387998%2fis-it-valid-to-iterate-over-every-permutation-of-a-regression-specification-and%23new-answer', 'question_page');

            );

            Post as a guest















            Required, but never shown





















































            Required, but never shown














            Required, but never shown












            Required, but never shown







            Required, but never shown

































            Required, but never shown














            Required, but never shown












            Required, but never shown







            Required, but never shown






            Popular posts from this blog

            How to check contact read email or not when send email to Individual?

            Bahrain

            Postfix configuration issue with fips on centos 7; mailgun relay