• new_name_who_dis_@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    10 months ago

    Well it depends on what you are building. If you are actually doing ML research, i.e. you want to publish papers, people are doing evaluation and you won’t get published without it. There’s a bunch of tricks that have been used to evaluate generative models that you can find in these papers. I remember in grad school our TA made us read a paper and then in the discussion he said that he thought the method they proposed was not good at all, he wanted us to read it to learn about their evaluation metric which he deemed “very clever”.

  • vikigenius@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    10 months ago

    It’s kind of weird that they use HFRL as the initialism instead of the much more common RLHF.

  • martianunlimited@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    10 months ago

    The typical measure for most ML conferences is the Fréchet inception distance (FID) but I have seen a number of generative AI papers, and what those values actually mean practically can be extremely obtuse, I appreciate papers that reports both the FID as a metric and also produce some representative examples of the output. (in the suplementary material if space is an issue)

  • Holiday-Union-6750@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    10 months ago

    Vibes with what I’ve seen in my job and the industry in general. Sadly, the greatest fun is only for huge corporations. Worth reading, definitely!