Probably because there’s also permission to use the X11 socket.
Probably because there’s also permission to use the X11 socket.
I think you’d have to modify the edid, since you’re setting a custom refresh rate, not a hidden one.
I’ve use wxEDID to force enable VRR before.
Well, aren’t you glad they’re removing go-git
then!
My understanding is that most of that all lives in mesa, and the kernel driver basically just abstracts the hardware.
Does it also restore the content of unsaved files of the application?
That’s up to the application.
If not, I’ll prefer
systemctl hibernate
. I wonder, what this new feature is for.
I believe this is for storing the position of specific windows, for multi-window applications (e.g. GIMP’s multi-window mode). So hibernation is very unrelated.
There’s The Serial Port, It’s not really ‘home networks’, but he finds and sets up very early (~80-90s) ISP gear and explains how it works and the history of it. Similar to how Ben Eater uses an ‘old’ 6502 to explain stuff.
I’ve had the same experience, you’re much better off RDPing into the VM. But I’d like to know if anyone has a better solution that doesn’t require an extra GPU.
On Asus motherboards you can enable ‘Memory Context Restore’, and it’ll remember the training. Unfortunately it seems rapid changes in the weather make my system unstable with it on.
cant move services as every other service sucks
What are your requirements?
I use Tidal and I know High/Max quality works in the web UI, just needs widevine support.
if they use AMD that’s better on linux, they don’t need to know what a GPU driver is.
Same goes for Intel, unless they need to use OneAPI.
That looks to be Volcanic Islands, which has good support with amdgpu
and no support by radeon
, according to Wikipedia.
I’m not sure what you meant by “set up radron kernel driver”, but you could maybe try blacklisting it.
I have no idea how CoW interacts with NTFS
With btrfs you can disable COW for specific files, that might give you a little performance boost.
I believe if your swap partition is on an encrypted LVM, you can still hibernate with kernel lockdown enabled.
Maybe, but also I think I was looking at the raw ‘data bits’, not ‘binary’ data. It’s actually almost exactly 4GiB, even when dropping down to minimum error correction (1.7 GiB otherwise).
(1454942×2953)÷1024÷1024÷1024≈4.00
Edit: So if alphanumeric mode could store lowercase letters, base64 would’ve stored more.
For those wondering, when using the biggest QR code with the maximum error correction (10,208 bytes), 1,454,942 QR codes is slightly less than 14GiB, which should be more than enough for a Windows ISO.
My math: (1454942×10208)÷1024÷1024÷1024≈13.83
Edit: Damn another guy beat me to it, now I wonder how I’m so far off.
Cloudflare tunnels uses a QUIC connection between the cloudflared
on the server and Cloudflare itself, which is encrypted similarly to HTTPS.
Whatever protocol cloudflared
uses to talk to your webserver locally is configurable through the Cloudflare access web UI (just change http to https). I’ve actually got it configured to use unix sockets, which lets me treat it differently in my nginx config.
Along with VRR over HDMI not being well supported, sometimes the monitors own EDID is a little buggy and Linux can’t guarantee VRR will work properly.
I wrote a blog post a while ago on fixing EDIDs, but it was pretty much a guessing game on what to change: https://stevetech.me/posts/force-enable-vrr-edid
I’ve had to do that with both Samsung and MSI monitors so far. If you’d like to post your EDID, I could check it myself with what I know.
Epic!
I’ve never seen that on modern AMD stuff that uses radv, but I’m sure it’s probably fine.
They don’t even need to be the same process. I’m pretty sure that’s just a common practice if something needs both protocols, but there’s nothing stopping you from having a web server on TCP 443 and a VPN server on UDP 443. Ports are an abstraction brought by each protocol, they aren’t in anyway related.