• 1 Post
  • 129 Comments
Joined 2 years ago
cake
Cake day: July 14th, 2023

help-circle
  • Wow, there isn’t a single solution in here with the obvious answer?

    You’ll need a domain name. It doesn’t need to be paid - you can use DuckDNS. Note that whoever hosts your DNS needs to support dynamic DNS. I use Cloudflare for this for free (not their other services) even though I bought my domains from Namecheap.

    Then, you can either set up Let’s Encrypt on device and have it generate certs in a location Jellyfin knows about (not sure what this entails exactly, as I don’t use this approach) or you can do what I do:

    1. Set up a reverse proxy - I use Traefik but there are a few other solid options - and configure it to use Let’s Encrypt and your domain name.
    2. Your reverse proxy should have ports 443 and 80 exposed, but should upgrade http requests to https.
    3. Add Jellyfin as a service and route in your reverse proxy’s config.

    On your router, forward port 443 to the outbound secure port from your PI (which for simplicity’s sake should also be port 443). You likely also need to forward port 80 in order to verify Let’s Encrypt.

    If you want to use Jellyfin while on your network and your router doesn’t support NAT loopback requests, then you can use the server’s IP address and expose Jellyfin’s HTTP ports (e.g., 8080) - just make sure to not forward those ports from the router. You’ll have local unencrypted transfers if you do this, though.

    Make sure you have secure passwords in Jellyfin. Note that you are vulnerable to a Jellyfin or Traefik vulnerability if one is found, so make sure to keep your software updated.

    If you use Docker, I can share some config info with you on how to set this all up with Traefik, Jellyfin, and a dynamic dns services all up with docker-compose services.


  • Look up “LLM quantization.” The idea is that each parameter is a number; by default they use 16 bits of precision, but if you scale them into smaller sizes, you use less space and have less precision, but you still have the same parameters. There’s not much quality loss going from 16 bits to 8, but it gets more noticeable as you get lower and lower. (That said, there’s are ternary bit models being trained from scratch that use 1.58 bits per parameter and are allegedly just as good as fp16 models of the same parameter count.)

    If you’re using a 4-bit quantization, then you need about half that number in VRAM. Q4_K_M is better than Q4, but also a bit larger. Ollama generally defaults to Q4_K_M. If you can handle a higher quantization, Q6_K is generally best. If you can’t quite fit it, Q5_K_M is generally better than any other option, followed by Q5_K_S.

    For example, Llama3.3 70B, which has 70.6 billion parameters, has the following sizes for some of its quantizations:

    • q4_K_M (the default): 43 GB
    • fp16: 141 GB
    • q8: 75 GB
    • q6_K: 58 GB
    • q5_k_m: 50 GB
    • q4: 40 GB
    • q3_K_M: 34 GB
    • q2_K: 26 GB

    This is why I run a lot of Q4_K_M 70B models on two 3090s.

    Generally speaking, there’s not a perceptible quality drop going to Q6_K from 8 bit quantization (though I have heard this is less true with MoE models). Below Q6, there’s a bit of a drop between it and 5 and then 4, but the model’s still decent. Below 4-bit quantizations you can generally get better results from a smaller parameter model at a higher quantization.

    TheBloke on Huggingface has a lot of GGUF quantization repos, and most, if not all of them, have a blurb about the different quantization types and which are recommended. When Ollama.com doesn’t have a model I want, I’m generally able to find one there.


  • I recommend a used 3090, as that has 24 GB of VRAM and generally can be found for $800ish or less (at least when I last checked, in February). It’s much cheaper than a 4090 and while admittedly more expensive than the inexpensive 24GB Nvidia Tesla card (the P40?) it also has much better performance and CUDA support.

    I have dual 3090s so my performance won’t translate directly to what a single GPU would get, but it’s pretty easy to find stats on 3090 performance.







  • If they do the form correctly, then it’s just an extra step for you to confirm. One flow I’ve seen that would accomplish this is:

    1. You enter your address into a form that can be auto-filled
    2. You submit the address
    3. If the address validates, the site saves the form and shows you the address in a more readable format. You can click Edit to make changes.
    4. If the address doesn’t validate, the site displays a modal asking you to confirm the address. If another address they were able to look up looks similar, it suggests you use that instead. It’s one click to continue editing, to use the suggested address, or to use what you originally entered.

    That said, if you’re regularly seeing the wrong address pop up it may be worth submitting a request to get your address added to the database they use. That process will differ depending on your location and the address verification service(s) used by the sites that are causing issues. If you’re in the US, a first step is to confirm that the USPS database has your address listed correctly, as their database is used by some downstream address verification services like “Melissa.” I believe that requires a visit to your local post office, but you may be able to fix it by calling your region’s USPS Address Management System office.



  • Good point!

    If OP is hourly, those 3 hours should be billed as work - probably under a generic HR-related category if one is available.

    If OP is salaried exempt, then this would fall under “doing any work at all” (all that’s needed to be paid for the day) and if sick time is tracked by day and not by hour, then OP doesn’t need to use one. If it’s tracked hourly then OP should make sure to only use 5 sick hours (or less, depending on how long the work-related conversations took) and depending on employer policies may not need to use any sick time at all.

    This also cut into the time OP could have been using to rest. It would be very reasonable for OP to need an extra day to recover, as a result.



  • Generally, usage of the term “gentrification” refers to the improvement of neighborhoods - or other places where people live, like apartment complexes - and, due to increased cost of living, the displacement of the people who used to live there. Displacement of less wealthy current residents when gentrification occurs is so common that it’s implied. If it weren’t, people wouldn’t have such low opinions of gentrification.

    If a forest has been gentrified, therefore, then - if you interpret “gentrified” in the same way - it follows that people who have been living there have been displaced. And since those people were living in a forest - not in a cabin in a forest - they’re necessarily homeless. Since OP didn’t say that they were building houses or apartments in the forest, that would mean that the wealthier people who displaced them were also homeless.

    Since the context was another commenter calling “gentrified forest” a cursed phrase, I don’t think I’m alone in thinking that.



  • Further, “Whether another user actually downloaded the content that Meta made available” through torrenting “is irrelevant,” the authors alleged. “Meta ‘reproduced’ the works as soon as it made them available to other peers.”

    Is there existing case law for what making something “available” means? If I say “Alright, I’ll send you this book if you want, just ask,” have I made it available? What if, when someone asks, I don’t actually send them anything?

    I’m thinking outside of contexts of piracy and torrenting, to be clear - like if a software license requires you to make any changed versions available to anyone who uses the software. Can you say it’s available if your distribution platform is configured to prevent downloads?

    If not, then why would it be any different when torrenting?

    Meta ‘reproduced’ the works as soon as it made them available to other peers.

    The argument that a copyrighted work has been reproduced when “made available,” when “made available” has such a low bar is also perplexing. If I post an ad on Craigslist for the sale of the Mona Lisa, have I reproduced it?

    What if it was for a car?

    I’m selling a brand new 2026 Alfa Romeo 4E, DM me your offers. I’ve now “reproduced” a car - come at me, MPAA.





  • They put their repo first on the list.

    Right. And are we talking about the list for OBS or of repos in general? I doubt Fedora sets the priority on a package level. And if they don’t, and if there are some other packages in Flathub that are problematic, then it makes sense to prioritize their own repo over them.

    That said, if those problematic packages come from other repositories, or if not but there’s another alternative to putting their repo first that would have prevented unofficial builds from showing up first, but wouldn’t have deprioritized official, verified ones like OBS, then it’s a different story. I haven’t maintained a package on Flathub like the original commenter you replied to but I don’t get the impression that that’s the case.