Often when I read ML papers the authors compare their results against a benchmark (e.g. using RMSE, accuracy, …) and say “our results improved with our new method by X%”. Nobody makes a significance test if the new method Y outperforms benchmark Z. Is there a reason why? Especially when you break your results down e.g. to the anaylsis of certain classes in object classification this seems important for me. Or do I overlook something?

  • Recent_Ad4998@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    One thing I have found is that if you have a large dataset, the standard error can become so small that any difference in average performance will be significant. Obviously not always the case, depending on the size of the variance etc but I imagine it might be why it’s often acceptable not to include them.

    • iswedlvera@alien.topB
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      This is the reason. People do significance tests when you want to draw conclusions with 20 samples on an entire population. If you have thousands of samples there won’t be much point.

        • iswedlvera@alien.topB
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 year ago

          I see what you mean. Yeah it shouldn’t be by default I don’t do statistical significance tests.