• 0 Posts
  • 31 Comments
Joined 2 years ago
cake
Cake day: April 23rd, 2023

help-circle



  • Onihikage@beehaw.orgtoLinux@lemmy.mlLinux suggestion
    link
    fedilink
    English
    arrow-up
    0
    ·
    3 months ago

    I can highly recommend Bazzite for your needs. It has a KDE version which is clearly your favorite Desktop Environment (DE), it’s extremely safe/stable due to being an Atomic distro (you can always boot into the previous image if a system update broke something), has incredible documentation, supports almost any traditional app through Distrobox (VPN requires rpm-ostree for now), has a scripted easy install of Waydroid for native android emulation, and has a few tweaks preconfigured to ensure the desktop gaming experience is a little more seamless out of the box than a stock distro. It really seems to tick all the boxes for what you’re looking for.

    If you want more focus on development and less on gaming, the Universal Blue team also makes Aurora for more developer-focused workloads, but Steam not being included in the image does introduce some usability regressions - Steam running via Flatpak or Distrobox is just plain less capable than a native install, though work is ongoing to make native installs Just Work even on Atomic systems.






  • Webtoon is still shitty in other ways. When they adapt a property, they want it their way, regardless of the author’s original vision. I’ve seen several stories that originated on Royal Road get Webtoon adaptations, and the adaptations always seem to change or leave out important parts of the story, making characters look stupid or just completely replacing entire sets of characters, forcing the story to diverge substantially when inevitably something they got rid of turns out to have been critically important to where the author was taking things. They turn great stories into middling slop every single time.



  • Unfortunately I can’t even test Llama 3.1 in Alpaca because it refuses to download, showing some error message with the important bits cut off.

    That said, the Alpaca download interface seems much more robust, allowing me to select a model and then select any version of it for download, not just apparently picking whatever version it thinks I should use. That’s an improvement for sure. On GPT4All I basically have to download the model manually if I want one that’s not the default, and when I do that there’s a decent chance it doesn’t run on GPU.

    However, GPT4All allows me to plainly see how I can edit the system prompt and many other parameters the model is run with, and even configure multiple sets of parameters for the same model. That allows me to effectively pre-configure a model in much more creative ways, such as programming it to be a specific character with a specific background and mindset. I can get the Mistral model from earlier to act like anything from a very curt and emotionally neutral virtual intelligence named Jarvis to a grumpy fantasy monster whose behavior is transcribed by a narrator. GPT4All can even present an API endpoint to localhost for other programs to use.

    Alpaca seems to have some degree of model customization, but I can’t tell how well it compares, probably because I’m not familiar with using ollama and I don’t feel like tinkering with it since it doesn’t want to use my GPU. The one thing I can see that’s better in it is the use of multiple models at the same time; right now GPT4All will unload one model before it loads another.


  • I have a fairly substantial 16gb AMD GPU, and when I load in Llama 3.1 8B Instruct 128k (Q4_0), it gives me about 12 tokens per second. That’s reasonably fast enough for me, but only 50% faster than CPU (which I test by loading mlabonne’s abliterated Q4_K_M version, which runs on CPU in GPT4All, though I have no idea if that’s actually meant to be comparable in performance).

    Then I load in Nous Hermes 2 Mistral 7B DPO (also Q4_0) and it blazes through at 50+ tokens per second. So I don’t really know what’s going on there. Seems like performance varies a lot from model to model, but I don’t know enough to speculate why. I can’t even try Gemma2 models, GPT4All just crashes with them. I should probably test Alpaca to see if these perform any different there…




  • The ELI5 for Fedora’s atomic desktops is that if Windows had an Atomic Desktop version, Program Files and most of the Windows folder would be read only, and each program you installed yourself would go into its own folder in your user directory. That’s the basic idea. It’s harder to screw up an Atomic system as long as you stick to containerized app formats like flatpak/appimage whenever possible. It makes it easier for everyone to diagnose problems, and easier for users to roll back if an update has problems. Even if you were to install it right now, you could use one simple command to “roll back” to any image from the last three months.

    The benefit of Bazzite is you have all of the above, plus a lot of gaming-related stuff preinstalled which, if you were to install them yourself in a normal Fedora environment, you’d likely have to spend a lot of time just learning how they’re supposed to be configured, how they interact, which versions have problems, and how to troubleshoot problems when an update to one app breaks a prerequisite for something else; eventually you end up in config hell instead of actually using your computer. With Bazzite, the image maintainers are the ones in config hell - they work out the kinks, app versioning, communicate with upstream to fix issues, all that, so your system should be in the most functional state that a Linux system can be, so you only have to think about using your apps.

    tl;dr

    • Atomic Desktops are more resilient to randomly breaking from updates or user error, and are easier to revert to a prior state if problems do arise
    • Bazzite is a custom Atomic image with lots of gaming stuff preinstalled and preconfigured to work properly out of the box
    • If you’re a gamer and wanting to try out Linux, Bazzite is going to be the least painful way to get your feet wet.
    • Immutable distros are excellent for daily driving. I daily drive one myself!




  • I’d get premium if they weren’t so insistent on bundling in bullshit I don’t want or care about to justify the high price. I put up with enough of that from cable TV. I’ll pay when there’s an ad-free tier that doesn’t do anything else and is a reasonable price for “the service that’s free with ads, but without ads”. If there was a per-device premium tier that I could throw on my Roku, and all my family members could have premium when they stream from there, I’d pay for that. I’d pay for family tier if it didn’t have the dumb single-household rule which screws over truckers and those who travel for a living.

    Google has options they could take to convince consumers to pay to not see ads, but there’s no creativity left there, no effort to court the market or adapt the service and prices to what potential customers need and are willing to pay. And it’s because they believe they are the market, and want to keep it that way.