• 0 Posts
  • 145 Comments
Joined 2 years ago
cake
Cake day: June 25th, 2023

help-circle






  • Yes, because back when I was learning almost 20 years ago I was able to google terms and read stuff for myself and it was also requirement for posting on forums, yet I was still getting a lot of help from the community. Times has changed it seems, so did the culture. Should I always assume ignorance and lack of interest? And now before I saw your comment I responded more comprehensively anyway, because why not, I’m not mad or anything. Should I take more time to write the response the first time around? Uh maybe idk


  • Desktop environments or window managers that support Wayland (one of the two displaying systems for Linux, newer one with aim to replace the obsolete one) and already implemented color management protocol in their compositors (programs that compose the image that is being displayed).

    In essence, everything that has recent version of Plasma 6 or current version of Hyperland is able to do HDR. Soon there will be new version of GNOME that does that too.

    Sooo… not Linux Mint, not Debian stable, not Ubuntu LTS.





  • Hi there, having two dualsense and one ps4 controller, using them for ages on Linux and they mostly run great, but your issues doesn’t sound completely new either.

    It’s very important on how you installed Steam and whether it’s native package or Flatpak. For Flatpak you might need special udev rules to allow the controller inside sandbox, usually can be installed using steam-devices package.

    As others said, enable Playstation Controller support in Steam’s controller settings page.

    Check if Steam overlay is functioning. In-game, press Shift+Tab and you should see the overlay and then you should be able to get to controller settings. Try out both with Steam Input enabled and disabled - by default I guess it depends on the game, but mostly enabling it will make it work for games that have issues picking up ds natively.

    Test your controllers using something like jstest-gtk. Perhaps there is something else connected that acts as player 1 controller.








  • Now that you can get latest software from Flathub, there’s really nothing wrong with Debian “stable” except for more recent hardware support that requires newer kernel at the very least (recent userspace drivers will also come from Flatpak if the software like Steam is also a Flatpak). That is, if the stable repo has all you need and there’s no reason to supplement it with external packages.

    There are however perfectly valid reasons for going with rolling to get recent improvements, which I for one care about. For example, now that PipeWire is pretty mature, Debian 13 will ship good version and it will serve well for the next 2-3 years, but some 2 years ago it was really important to get the latest and greatest to have good experience - and even early it was better than PulseAudio would ever be, just still improving rapidly, not ready for full freeze. Other example - KDE Plasma improved significantly from version 6.0 onwards introducing long awaited functionality like fractional scaling, HDR, but also improved stability and general polish. It will only be introduced in Debian 13, one full year after it was introduced.

    Lastly, there’s nothing wrong with rolling and it isn’t really “unstable”. Using Arch full time for the last 12 years, I only had like 2-3 situations when update actually broke something and it wasn’t my misconfiguration or a skill issue. Even then it could easily be avoided by using linux-lts kernel. In fact my Debian/Ubuntu installs were much less stable as there was always something missing that I needed (in era before Flatpaks or AppImages especially) relying on 3rd party apt repos, causing breakages and conflicts. I would usually upgrade Debian to testing or unstable anyway, so rolling, but one that’s actually open for breakage.


  • Generally the industry shifted in a direction where it heavily relies on containers for running cloud applications. This solves many problems with traditional server systems where you’d be sticking to certain distro, so certain dependencies are in fixed versions, which brings some limitations. Container is an environment to run process in an isolated way so that it had its own root filesystem, its own view on what resources are available, sort of like it was separate machine, but it’s still running on the same machine natively using the same kernel as the host. You can then have multiple of such containers, all serving its narrow purpose and they all come with the complete fs and whatever distro release they are tested with. Nowadays cloud computing is all about containers and they come from images that are built in OCI format using Dockerfile syntax. After building an image, it is typically pushed into registry where it can be pulled from over network to be utilized across different nodes, which makes it pretty easy to scale and propagate changes in cloud environments.

    Now what that means to Bazzite/Universal Blue is that it uses similar tech to deploy the system, though the target here is your local machine. Of course some of the characteristics aren’t relevant in this scenario, but it solves some of the same problem - build predictable and reproducible environment that can be thoroughly tested before publishing. The general idea is similar to how devops build cloud apps: there is CI pipeline that runs the build using giant Dockerfile (or Containerfile, same thing) inside of which they include everything that the system needs (running traditional package manager and act as it was normal Linux distro during the build), which then results as image that’s being pushed to registry. Bazzite users then install updates by pulling new version of the image and ‘rebasing’ to it. It is called rebasing here, because rpm-ostree lets users add additional layers with more packages on top of that.

    EDIT: here’s the Containerfile I’ve been talking about: https://github.com/ublue-os/bazzite/blob/main/Containerfile Might give you some idea on how this works.