Docker is 80% Linux, 10% Networking, 5% Virtualization and remaining 5% is actual Docker-specific things.
If you learn Linux, networking and virtualization, Docker is just a cherry on top.
Docker is 80% Linux, 10% Networking, 5% Virtualization and remaining 5% is actual Docker-specific things.
If you learn Linux, networking and virtualization, Docker is just a cherry on top.
All of these functionalities can be provided by a simple WebSocket + REST server. The car connects to the WebSocket, and you can access these functionalities from your phone either with WebSockets or regular HTTP requests.
Cheapest servers with backend written in JS can easily handle thousands of WebSocket connections, and written in Go tens of thousands WebSocket connections. They would not ever need like over 100 of these servers GLOBALLY, which would cost them around $3000 monthly.
That’s the price of 60 subscriptions, which is freaking ridiculous.
That is because windows filesystem is mounted to WSL through NFS and while transferring large files through that is ok, transferring huge amounts of small files is really slow.
You realize that maintaining a server that would allow that costs pennies?
You wouldn’t pay $150 for a lollipop, but somehow people think this is ok.
This problem exists exactly because of people like you, thinking it’s OK to pay for the features you already paid for.
You mentioned you changed firewall rules for that device. Any chance you have set outbound rule instead of inbound rule?
Anyway, what’s the output of ip route
?
Mainly GTG response time and latency. For watching movies it’s generally not a problem, but when it comes to playing games with a mouse, latency can be a huge issue, and bad GTG response time leads to smearing.
But yeah, 4x the price is ridiculous.
And as always, DRMs fuck only legitimate customers, and pirates can watch anywhere at full quality.
That’s one of the reasons I don’t feel bad about pirating any more. Not even the cost, but the fact that if you pay you’re going to have a worse experience.
Am I too 1Gb/s fiber connected to understand that?
Better in which way? WSL2 is a VM running ALONGSIDE Windows, not inside. Its performance is basically bare metal. If you have enough RAM, there is no reason to use cygwin instead of WSL2.
https://www.oracle.com/cloud/free/
There are tutorials on youtube on how to create a VM and set up a firewall for external access.
If you’re lucky enough to successfully create an account on Oracle Cloud, you can also try Oracle Cloud Free Tier. You can have free ARM64 x4 CPU and 24 GiB RAM totally free of charge. There might be problems with availability during VM registration, but there are scripts that automate spamming for checking every 80 seconds.
I’ve been using it for 2 years and it’s great. However be aware that your VM might get erased if you have a free account. That too can be remedied if you update to a premium subscription (You still get Free Tier resources without a charge). Nobody has reported an erased VM on a premium plan yet.
Still, I am pretty sure they can erase it if you do illegal stuff with it. I’ve been using it only to host Minecraft Server, as well as other services using Docker. So far so good.
If you’re trying to self-host http service, you can use cloudflare tunnels.
If you know how to use git, you will know how to use docker (provided you know what you want to do). They are completely different programs, yet you can quickly grasp the other instinctively.
Now, Photoshop and Blender - they are also different programs, but if you know Photoshop, you still need to relearn Blender’s interface completely.
This is why I prefer terminal programs in general. Unless it’s more convenient to use GUi, i.e. Drag&Drop file manager, some git tools etc.
Learn it first.
I almost exclusively use it with my own Dockerfiles, which gives me the same flexibility I would have by just using VM, with all the benefits of being containerized and reproducible. The exceptions are images of utility stuff, like databases, reverse proxy (I use caddy btw) etc.
Without docker, hosting everything was a mess. After a month I would forget about important things I did, and if I had to do that again, I would need to basically relearn what I found out then.
If you write a Dockerfile, every configuration you did is either reflected by the bash command or adding files from the project directory to the image. You can just look at the Dockerfile and see all the configurations made to base Debian image.
Additionally with docker-compose you can use multiple containers per project with proper networking and DNS resolution between containers by their service names. Quite useful if your project sets up a few different services that communicate with each other.
Thanks to that it’s trivial to host multiple projects using for example different PHP versions for each of them.
And I haven’t even mentioned yet the best thing about docker - if you’re a developer, you can be sure that the app will run exactly the same on your machine and on the server. You can have development versions of images that extend the production image by using Dockerfile stages. You can develop a dev version with full debug/tooling support and then use a clean prod image on the server.
Java used to lack many features to make the stuff you wanted it to do, so most Java programmers adapted design patterns to solve these problems.
Honestly, older versions of Java are utter garbage DX. The only reason it got so popular was because of aggressive enterprise marketing and it worked. How can a language lack such an essential feature as default parameters?
So, anyway after the great hype Java lost its marketshare, and developers were forced to learn another technologies. And of course, instead of looking for language-native way of solving problems, they just used same design patterns.
And thus MoveAdapterStrategyFactoryFactories were in places where simple lambda function would do the same thing, just not abstracted away three layers above. Obviously used once in the entire codebase.
Imo the only really good thing about Java was JVM, while it was not perfect, it actually delivered what it promised.
Totally agree on that. When the first generation RTX cards launched, I was pretty sure that it was just going to be another gimmick like PhysX, but today, it’s an inevitable future.
I might be wrong, but even for games like Cyberpunk 2077 there is a finite set of world states that define lighting conditions (time of day weather etc.).
So prebaking lighting information for all these combinations and then figuring out a way to create transitions between them would maybe not be the perfect representation, but best of both worlds.
However, given how fast RayTracing improves hardware-wise, in my opinion it would make no sense to even consider researching and developing a solution of that kind.
Real time RT really is meh, but I like what they’re doing in CS2. Prebaked Global Illumination looks freaking fantastic.
English is not my first not language. When I write something down in my first language (polish), it feels more like I’m transcribing things I silently say to myself, while with english I’m actually thinking about every word I type.
The funny thing is, the better I am getting at English, making those types of mistakes is getting easier for me.
But idk, this is just my experience.
24.04 won’t have Plasma 6, but 24.10 will. In other words, fall 2024.
Or you can use KDE Neon, which is basically Ubuntu LTS, but with the newest Plasma.