Hello everyone,

I would like to get started with selfhost with two projects.

Project A (for me): A NUC with Proxmox installed on it, two VMs including a Home Assistant and a NAS system that I haven’t chosen yet.

The only question I have with this project is:

  • how to access the NAS and HA separately from the outside knowing that my access provider does not offer a static IP and that access to each VM must be differentiated from Proxmox.

Project B (for my uncle): A NUC (with Proxmox or not, I don’t know yet, perhaps simpler for making backups), with HA but especially Frigate. The goal is to use Google Coral to do recognition on 3 video surveillance cameras.

My questions are:

  • is Coral really useful with 3 cameras?
  • do you need a Coral in USB or M.2 version?
  • are there affordable NUCs with free M.2 slots?
  • won’t proxmox add a layer of complexity with Coral/Frigate/a Zigbee dongle?

Thank you in advance for your help and sorry if my post is long.

PS: if you have recommendations for cameras that work with Frigate and are self-powered with solar panels, I’ll take them!

Edit : 8 april 2024

A little update. Thank you everyone for your super quick responses!

Regarding my uncle’s project and after big discussions, he is going to buy Reolink cameras and that’s it. This will be much simpler for maintenance than building a server.

Regarding my project: I chose a Beelink Mini S12 pro with an N100 processor (for its low consumption) with a 2.5 bay for an SSD for my Nextcloud.

I wondered if I wouldn’t take the opportunity to add pihole and that’s where new questions arise…

I see a lot of people installing Pihole on Docker, should I put it on Docker? Or create a VM?

Should Docker be installed on Proxmox or on a VM?

Is Proxmox really useful, shouldn’t I better install HA/Nextcloud/Pihole under Docker directly?

Should I use LXC or Docker?

  • LifeBandit666@feddit.uk
    link
    fedilink
    English
    arrow-up
    2
    ·
    9 months ago

    I share bind-mounts currently between multiple LXC from the host Proxmox OS, configuration is pretty easy, and there are lots of tutorials online for getting started.

    Now then:

    Are you sharing SMB mounts? I have my HDDs passed through to OMV and have considered just trying to pass them through to other VMs, but never tried because I don’t wanna break anything.

    I have seen that you can share SMB to Proxmox and use them in Proxmox but don’t know if you can use them in VMs too.

    As it is I really struggled with mounting smb for a couple of weeks and then had an “aha” moment last weekend, and have it all figured out now.

    The Tailnet idea was so I can just mount everything to the Tailnet and stop worrying about whether it’s on this vlan or that. I was trying to set up an Openwrt container with VPN, which I could use for any container that needs a vpn, but then those containers couldn’t see the main network properly…

    I’ve given up on that now and have my SMB mounts all set up, but feel like pass-through would give better network speeds for moving things around.

    • dai@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      ·
      9 months ago

      Yeah there is a workaround for using bind-mounts in Proxmox VMs: https://gist.github.com/Drallas/7e4a6f6f36610eeb0bbb5d011c8ca0be

      If you wanted, and your drives are mounted to the Proxmox host (and not to a VM), try an LXC for the services you are running, if you require a VM then the above workaround would be recommended after backing up your data.

      I’ve got my drives mounted in a container as shown here:

      Basicboi config, but it’s quick and gets the job done.

      I’d originally gone down the same route as you had with VMs and shares, but it’s was all too much after a while.

      I’m almost rid of all my VMs, home assistant is currently the last package I’ve yet to migrate. Migrated my frigate to a docker container under nixos, tailscale exit node under nixos too while the vast majority of other packages are already in LXC.

      • LifeBandit666@feddit.uk
        link
        fedilink
        English
        arrow-up
        3
        ·
        9 months ago

        This all sounds awesome. So eli5 I have all my drives mounted to Proxmox, then passed through to OMV in a VM.

        I can just mount these same drives to containers no issues right now, and I can add them to VMs using your link?

        I would like to get down to LXCs too, but I’ve found VMs so much easier to set up and use. I’ll try your way

        • dai@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          ·
          edit-2
          9 months ago

          I’ve not tested the method linked but yeah it would seem like it’s possible via this method.

          My lone VM doesn’t need a connection to those drives so I’ve not had a point to.

          You could probably run OMV in an LXC and skip the overheads of a VM entirely. LXC are containers, you can just edit the config files for the containers on the host Proxmox and pass drives right through.

          Your containers will need to be privileged, you can also clone a container and make it privileged if you have something setup already as unprivileged!