The new global study, in partnership with The Upwork Research Institute, interviewed 2,500 global C-suite executives, full-time employees and freelancers. Results show that the optimistic expectations about AI’s impact are not aligning with the reality faced by many employees. The study identifies a disconnect between the high expectations of managers and the actual experiences of employees using AI.

Despite 96% of C-suite executives expecting AI to boost productivity, the study reveals that, 77% of employees using AI say it has added to their workload and created challenges in achieving the expected productivity gains. Not only is AI increasing the workloads of full-time employees, it’s hampering productivity and contributing to employee burnout.

  • Womble@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    edit-2
    5 months ago

    Forgive me for not trusting an ariticle that says that AI will use a petawatt within the next two years. Either the person who wrote it doesnt understand the difference between energy and power or they are very sloppy.

    Chat GPT took 50GWh to train source

    Americans burn 355 million gallons of gasoline a day source and at 33.5 Kwh/gal source that comes out to 12,000GWh per day burnt in gasoline.

    Water usage is more balanced, depending on where the data centres are it can either be a significant problem or not at all. The water doesnt vanish it just goes back into the air, but that can be problematic if it is a significant draw on local freshwater sources. e.g. using river water just before it flows into the sea, 0 issue, using a ground aquifer in a desert, big problem.

    • Flying Squid@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      5 months ago

      Training is already over. This has nothing to do with training, so that is irrelevant. This is about how much power is needed as it is used more and more. I think you know that.

      Also, I’m not sure why you think just because cars emit a lot of CO2, it doesn’t mean that other sources that emit a lot of CO2, but less than cars, are a good thing.

      The water doesnt vanish it just goes back into the air,

      Cool, tell that to all the people who rely on glaciers for their fresh water. That only includes a huge percentage of people in India and China.

      But really, what you’re telling me is that studies and scientists are wrong and you’re right. Cool. Good luck convincing people of that.

      • Womble@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        5 months ago

        This New Yorker article estimates GPT usage at 0.5GWhr a day, which comes out to 0.0041% of the energy burnt just in vehicle gasoline per day in the USA (and this is for worldwide usage for chatGPT).

        I’m not asking you to trust me at all, I’ve listed my sources, if you disagree with any of them or multiplying three numbers together that’s fine.

        Cool, tell that to all the people who rely on glaciers for their fresh water. That only includes a huge percentage of people in India and China.

        Yes, if you read my last reply I answered that directly. Water usage can be a big issue, or it can be a non-issue, its locale dependent.

        • Flying Squid@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          edit-2
          5 months ago

          What New Yorker article? You didn’t link to one. I, however, linked to Yale University which has a slightly better track record on science than The New Yorker.

          And, again, you are arguing that emitting less CO2 is a good thing. It is not.

          And if water can be a big issue, why is AI a good thing when it uses it up? You can say “people shouldn’t build data centers in those locations,” but they are. And the world doesn’t run on “shouldn’t.”

          Edit: Now you linked to it. It’s paywalled, which means I can’t read it and I doubt you did either.

          • Womble@lemmy.world
            link
            fedilink
            English
            arrow-up
            0
            ·
            edit-2
            5 months ago

            Apologies, I didn’t post the link, it’s edited now.

            If you want to take issue with all energy usage that’s fine, its a position to take. But it’s quite a fringe one given that harnessing energy is what gives us the quality of life we have. Thankfully electricity is one of the easiest forms of energy to decarbonise and is already happening rapidly with solar and wind power, we need to transition more of our energy usage to it in order to reduce fossil fuel usage. My main point is that this railing against AI energy usage is akin to the whole plastic straw ban, mostly performative and distracting from the places where truely vast amounts of fossil fuels are burnt that need to be tackled urgently.

            You can say “people shouldn’t build data centres in those locations,” but they are. And the world doesn’t run on “shouldn’t.”

            I’m 100% behind forcing data centres to use sustainable water sources or other methods of cooling. But that is a far cry from AI energy consumption being a major threat, the vast majority of data centre usage isn’t AI anyway, it’s serving websites like the one we are talking on right now.

            • rekorse@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              ·
              5 months ago

              Why can’t we analyze AI on its own merits? We dont base our decisions on whether an idea is more or less polluting than automobiles. We can look at what we are getting for what’s being put into it.

              The big tech companies could scrap their AI tech today and it wouldnt change most peoples lives.