robber

joined 2 years ago
[–] robber@lemmy.ml 4 points 3 days ago

SFTPGo is such an awesome project, never had any problems with it.

[–] robber@lemmy.ml 1 points 1 week ago

Some people on another discussion platform were praising the new Mistral Small models for agentic use. I wasn't able to try them myself yet, but with 24b params you would certainly fit a quantized version in your 24GB.

[–] robber@lemmy.ml 2 points 1 week ago

Thanks for the tip about kobold, didn't know about that.

And yeah, I can understand that building your own rig might feel overwhelming at first, but there's tons of information online that I'm sure will help you get there!

[–] robber@lemmy.ml 2 points 1 week ago (1 children)

Alright thanks! I found it somewhat difficult to find information about the hardware requirements online, but yeah, maybe I just have to try it.

[–] robber@lemmy.ml 2 points 1 week ago (3 children)

Thanks that looks cool, I'll definitely try and report back. Do you happen to know what the hardware requirements are? I have access 64GB of RAM and 48GB of VRAM across 3 RTX2000e Ada GPUs.

 

Hey fellow llama enthusiasts! Great to see that not all of lemmy is AI sceptical.

I'm in the process of upgrading my server with a bunch of GPUs. I'm really excited about the new Mistral / Magistral Small 3.2 models and would love to serve them for me and a couple of friends. My research led me to vLLM with which I was able to double inference speed compared to ollama at least for qwen3-32b-awq.

Now sadly, the most common quantization methods (GGUF, EXL, BNB) are either not fully (GGUF) or not at all (EXL) supported in vLLM, or multi-gpu inference thouth tensor parallelism is not supported (BNB). And especially for new models it's hard to find pre-quantized models in different, more broadly supported formats (AWQ, GPTQ).

Does any of you guys face a similar problem? Do you quantize models yourself? Are there any up-to-date guides you would recommend? Or did I completely overlook another, obvious solution?

It feels like when I've researched something yesterday, it's already outdated again today, since the landscape is so rapidly evolving.

Anyways, thank you for reading and sharing your thoughts or experience if you feel like it.

[–] robber@lemmy.ml 4 points 3 weeks ago* (last edited 3 weeks ago)

I'll add Pangolin to the list, it's a self-hosted Cloudflare tunnel alternative.

[–] robber@lemmy.ml 2 points 3 weeks ago

It really depends on how much you enjoy to set things up for yourself and how much it hurts you to give up control over your data with managed solutions.

If you want to do it yourself, I recommend taking a look at ZFS and its RAIDZ configurations, snapshots and replication capabilities. It's probably the most solid setup you will achieve, but possibly also a bit complicated to wrap your head around at first.

But there are a ton of options as beautifully represented by all the comments.

[–] robber@lemmy.ml 2 points 4 weeks ago (1 children)

Thanks for the hint to pocketID, haven't heard of it before. That makes me think it's time to upgrade my auth stack as well.

 

Text: Allows you to determine whether to limit CPUID maximum value. Set this to enabled for legacy operating systems such as Linux or Unix.

Found this in the BIOS of a Gigabyte Z97X-UD3H mobo.

[–] robber@lemmy.ml 1 points 1 month ago (1 children)

That sounds awesome! No issues at all so far?

[–] robber@lemmy.ml 1 points 1 month ago (1 children)

Thanks for the list! Do you use Pangolin yourself?

 

Hi fellow homelabbers! I hope your day / night is going great.

Just stubled across this self-hosted cloudflare tunnel alternernative called Pangolin.

  • Does anyone use it for exposing their homelab? It looks awesome, but I've never heard of it before.

  • Should I be reluctant since it's developed by a US-based company? I mean security-wise. (I'll remove this question if it's too political.)

  • Does anyone know of alternatives pieces or stacks or software that achieve the same without relying on cloudflare?

Your insights are highly appreciated!

[–] robber@lemmy.ml 5 points 1 month ago (1 children)

Doesn't feel gnomey enough.

[–] robber@lemmy.ml 1 points 2 months ago

TuxedoOS because my so-called "Linux-Laptop" turned out to not run mainline Linux very smoothly. But I hate that fact that it's Ubuntu-based.

I'd use Debian, Arch or dabble with Void if I could on my laptop, my servers run Debian or Alma.

 
 

Hey fellow self-hosting lemmoids

Disclaimer: not at all a network specialist

I'm currently setting up a new home server in a network where I'm given GUA IPv6 addresses in a 64 bit subnet (which means, if I understand correctly, that I can set up many devices in my network that are accessible via a fixed IP to the oustide world). Everything works so far, my services are reachable.

Now my problem is, that I need to use the router provided by my ISP, and it's - big surprise here - crap. The biggest concern for me is that I don't have fine-grained control over firewall rules. I can only open ports in groups (e.g. "Web", "All other ports") and I can only do this network-wide and not for specific IPs.

I'm thinking about getting a second router with a better IPv6 firewall and only use the ISP router as a "modem". Now I'm not sure how things would play out regarding my GUA addresses. Could a potential second router also assign addresses to devices in that globally routable space directly? Or would I need some sort of NAT? I've seen some modern routers with the capability of "pass-through" IPv6 address allocation, but I'm unsure if the firewall of the router would still work in such a configuration.

In IPv4 I used to have a similar setup, where router 1 would just forward all packets for some ports to router 2, which then would decide which device should receive them.

Has any of you experience with a similar setup? And if so, could you even recommend a router?

Many thanks!


Edit: I was able to achieve what I wanted by using OpenWrt and their IPv6 relay mode. Now my ISP router handles all IPv6 addresses directly, but I'm still able to filter the packets using the OpenWrt firewall. For IPv4 I didn't figure out how to, at the same time, use the ISP's DHCP server, so I just went with double NAT. Everything works like a charm. Thank you guys for pointing me in the right direction.

140
submitted 10 months ago* (last edited 10 months ago) by robber@lemmy.ml to c/lemmyshitpost@lemmy.world
 

Most relevant section translated to english:

If he (Trump) wins the election on November 5, his billionaire supporter Musk will chair the new board. This is to implement a full financial and performance audit of the entire government and make recommendations for drastic reforms.

Source: Swiss state media article

 

A couple of years ago, QR-bills were introduced in Switzerland as a means to make payments easier. My bank provides an app to scan the QR codes, which I prefer not to install. The only other option they provide to scan the codes is to use the webcam. Am I supposed to print my digital bills to have my webcam scan them again? Just let me upload a goddamn screenshot.

 

I've been looking into self-hosting LLMs or stable diffusion models using something like LocalAI and / or Ollama and LibreChat.

Some questions to get a nice discussion going:

  • Any of you have experience with this?
  • What are your motivations?
  • What are you using in terms of hardware?
  • Considerations regarding energy efficiency and associated costs?
  • What about renting a GPU? Privacy implications?
 
 

Just wanted to share my happiness.

AIO is the new (at least on my timeline) installation method of Nextcloud, where most of the heavy-lifting is taken care of automatically.

https://github.com/nextcloud/all-in-one

view more: next ›