this post was submitted on 12 Jan 2025
30 points (91.7% liked)

Selfhosted

41084 readers
265 users here now

A place to share alternatives to popular online services that can be self-hosted without giving up privacy or locking you into a service you don't control.

Rules:

  1. Be civil: we're here to support and learn from one another. Insults won't be tolerated. Flame wars are frowned upon.

  2. No spam posting.

  3. Posts have to be centered around self-hosting. There are other communities for discussing hardware or home computing. If it's not obvious why your post topic revolves around selfhosting, please include details to make it clear.

  4. Don't duplicate the full text of your blog or github here. Just post the link for folks to click.

  5. Submission headline should match the article title (don’t cherry-pick information from the title to fit your agenda).

  6. No trolling.

Resources:

Any issues on the community? Report it using the report flag.

Questions? DM the mods!

founded 2 years ago
MODERATORS
 

I am building a Proxmox server running on an SFF PC. Right now I have:

  • 1 x 250 GB Kingston A400 Sata SSD
  • 1 x 512 Gb Samsung NVMe 970 Evo Plus
  • 1 x 512 Gb Kingston NVMe KC3000
  • 1 x 12 Tb Seagate Ironwolf Re-certified disk

I plan to install Proxmox on the 250Gb Kingston disk using ext4 and use it only for Proxmox and nothing else.

I am thinking of configuring ZFS mirrored raid on the two NVMe disks. Here one disk is on my mobo, and the other is connected to the PCIe slot with an adapter, as I have only one M2 slot on the mobo. I plan to use this zpool for VMs and containers.

Finally, the re-certified 12 Tb disk is currently going through a long smarctl test to confirm that it is usable and it is supposed to be used primarily for storing media and non-critical data and VM snapshots, which I don't care much about it. I will in parallel most likely adopt the critical data to a cloud location as an additional way to protect my most important data.

My question is should I be really concerned about the lack of DRAM in the Kingston A400 SSD and its relatively low TBW endurance (85 TB) in case I would run it only to boot Proxmox from it and I think the wear out of the drive would be negligible.

  • I have the option to exchange the Proxmox boot drive with a proper SSD, like a Samsung 870 Evo (SATA SSD, using MLC NAND and having DRAM cache). I would of course need to pay around 60% more but I am just thinking that this might be an overkill.
  • Do you think that using ZFS pool for the two NVMe drives will wear them out very quickly? I will have 3-4 VMs and a bunch of containers.
  • Is the use of a slow Proxmox boot drive (SATA SSD) going to slow down the VMs and containers as they will run on much quicker NVMe SSDs, or it won't matter?
  • Shall I format the Seagate HDD in xfs to speed up the transfer of large files or shall I stick to ext4?
  • What other tests shall I run to confirm that the HDD is indeed fine and I can use it?
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 2 points 2 days ago (1 children)
[–] [email protected] 1 points 2 days ago (1 children)

For no particular reason, except for btrfs taking up less RAM. I don't know their specs, but the lack of RAM was my reason for deciding against btrfs for my large non-mirrored HDD.

[–] [email protected] 2 points 2 days ago

You can set the Arc cache size. Also it gets automatically freed when it is needed by applications.