Getting Keycloak and Headscale working together.
But I did it after three weeks.
I captured my efforts in a set of interdependent Ansible roles so I never have to do it again.
Getting Keycloak and Headscale working together.
But I did it after three weeks.
I captured my efforts in a set of interdependent Ansible roles so I never have to do it again.
I still have it on my Pixel 8 Pro. It requires a double tap to occur in less than 300 milliseconds.
This bug has been the bane of my existence for almost four years now: https://issuetracker.google.com/issues/204650736
It would be extremely barebones, but you can do something like this with Pandoc.
It’ll probably be there, but at least it can be disabled in the settings now. It won’t go away on its own.
This was when I stopped using it for a while. I sent multiple feedback messages as it really irritated me.
That I agree with. Microsoft drafted the recommendation to use it for local networks, and Apple ignored it or co-opted it for mDNS.
Macs aren’t the only thing that use mDNS, either. I have a host monitoring solution that I wrote that uses it.
Yeah, that’s why I started using .lan.
I was using .local, but it ran into too many conflicts with an mDNS service I host and vice versa. I switched to .lan, but I’m certainly not going to switch to .internal unless another conflict surfaces.
I’ve also developed a host-monitoring solution that uses mDNS, so I’m not about to break my own software. 😅
Coincidentally, I just found this other thread that mentions EasyEffects: https://programming.dev/post/17612973
You might be able to use a virtual device to get it working for your use case.
It depends on the model you run. Mistral, Gemma, or Phi are great for a majority of devices, even with CPU or integrated graphics inference.
I’m also going to push forward Tilda, which has been my preferred one for a while due to how minimal the UI is.
Pixel Experience is unfortunately dead now. 🙁
Yeah - the operating system (or perhaps the display hardware itself, not sure) has to stretch each software pixel to a fractional amount of larger hardware pixels. In the case of upscaling 720p to 1080p, each 720p software pixel has to stretch to 1.33 hardware pixels. This forces blending to occur, which makes the image less sharp.
The worst part of this in my opinion is reading text.
You also lose integer scaling if you need to run a game at common resolutions below 1080p. (720p/800p, etc.)
We all mess up! I hope that helps - let me know if you see improvements!
I think there was a special process to get Nvidia working in WSL. Let me check… (I’m running natively on Linux, so my experience doing it with WSL is limited.)
https://docs.nvidia.com/cuda/wsl-user-guide/index.html - I’m sure you’ve followed this already, but according to this, it looks like you don’t want to install the Nvidia drivers, and only want to install the cuda-toolkit metapackage. I’d follow the instructions from that link closely.
You may also run into performance issues within WSL due to the virtual machine overhead.
Good luck! I’m definitely willing to spend a few minutes offering advice/double checking some configuration settings if things go awry again. Let me know how things go. :-)
It does!