https://www.youtube.com/watch?v=QEbI6v2oPvQ

I had a lot of trouble setting up ROCm and Automatic1111. I tried first with Docker, then natively and failed many times. Then I found this video. It has a good overview for the setup and a couple of critical bits that really helped me. Those were the reinstallation of compatible version of PyTorch and how to test if ROCm and pytorch are working. I still had a few of those Python problems that crop up when updating A1111, but a quick search in A1111 bug reports gave work arounds for those. And a strange HIP hardware error came at startup, but a simple reboot solved that.

Also he says he couldn’t make it work with ROCm 5.7, but for me now 2 months later, ROCm 5.7 with 7900 XTX and Ubuntu 22.04 worked.

And coming from a Windows DirectML setup, the speed is heavenly.

    • fakeluke@alien.topB
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      Looking good!
      How do you come out of the python dependency hell in fedora?
      When i to set up the python3.10 env and try to install the requirements, torchsde always complain about a specific numpy, python or pip versions.

    • 1stnoob@alien.topB
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      u still need to install their crappy drivers that take your system down with them on a daily basis :>

  • orucreiss@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    i am looking for a guide that is based on latest version of rocm&pytorch etc. available? Is it running good?

    • liberal_alien@alien.topOPB
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      So far I’m pretty happy with it. I can do 1024 resolution gens at 2.7 it/s. If I try to put more of them in a batch, then I might run out of memory, but compared to Windows and DirectML this is quite a bit faster and has better memory management.

      I also tried some animatediff for the first time on this, but only managed to render a 256 resolution gif. Even 512 resolution caused a crash.

      I also managed to get ComfyUI setup to serve Krita as a Stable Diffusion backend, but I only just got it to work and don’t have the first clue about how to use it properly yet. I used this plugin: https://github.com/Acly/krita-ai-diffusion.

  • ArtsM@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    I did this setup for native use on Fedora 39 workstation about a week and a half ago, the amount of dicking about with python versions and venvs to get a compatible python+pytorch+rocm version together was a nightmare, 3 setups that pytorch site said were “supported” before it finally worked with rocm5.7. It was my first experience with setting it up natively, have ran a docker version in the past and will probably stick to that in the future.

  • CasimirsBlake@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    It should not be this involved. It is still a cluster of a process. But I hope some folks can get this to work.

    • DarkeoX@alien.topB
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      It could be way easier with proper Docker images. That’s what I tend to do for all these projects.

      ROCM team had the good idea to release Ubuntu image with the whole SDK & runtime pre-installed. But that’s simply not enough to conquer the market and gain trust. Ideally, they’d release images bundled with some of the most popular FLOSS ML tools ready to use and the latest stable ROCm version.