@dosse91 - eviltoast
  • 4 Posts
  • 101 Comments
Joined 1 year ago
cake
Cake day: June 9th, 2023

help-circle

  • Arch Linux. Everyone said it was hard to use, unstable, etc. but my experience with it has been the exact opposite.

    Yes, the install process is needlessly complicated (although it got a lot simpler now that we have archinstall), but the OS itself is rock solid and rarely has any issues that require more than a reboot or a package reinstall to solve. The AUR is a godsend too if you don’t want or don’t know how to compile stuff from source.


  • The first time I heard about programming being obsolete was when I was taught UML in university. That was over almost 15 years ago and it didn’t happen, if anything programmers now also had to know UML, which isn’t all that bad but it definitely didn’t replace anything, it’s just useful for designing and documenting projects.

    I also heard from colleagues that in the 80s and 90s people said that SQL was supposed to be used by users directly, making (some) programming obsolete.

    Now AI bullshit claims to be making programming obsolete. I won’t hold my breath.

















  • I may be partially responsible for this lazy ass implementation.

    3 months ago I was playing around with stable diffusion a lot and because I sleep in the same room where my PC is, I used to lower the TDP of the GPU during the night to 150w to keep it quiet. One day while SD was running, I lowered the TDP in LACT and pressed Apply but instead of getting quieter, the fans ramped up and I was shocked seeing that the card was in fact pulling 420w instead of its rated 293w (6900xt).

    I tracked down the issue to the driver incorrectly applying the power limit, basically if you set a TDP that’s too low for the current power state, the driver would disable the power limit entirely until the card entered a lower power state, after which, your new TDP would be correctly applied.

    Running a modern GPU without power limits is bad and potentially dangerous for everything involved: the GPU, the VRMs, even the power supply cables may melt as we’ve seen with nVidia cards. So I reported the issue immediately to the AMDGPU developers (my issue is linked in the article).

    They quickly came up with a fix, which I tested, which wouldn’t allow you to set a TDP lower than the lowest valid TDP for the highest power state. This gets the job done but it’s a kludge more than a fix, ideally the driver should realize that the new TDP is too low for the current power state and switch to a lower power state, and I don’t know why AMD implemented such a shitty solution in their official kernel driver.