this post was submitted on 14 Feb 2024
5 points (100.0% liked)

Selfhosted

40415 readers
388 users here now

A place to share alternatives to popular online services that can be self-hosted without giving up privacy or locking you into a service you don't control.

Rules:

  1. Be civil: we're here to support and learn from one another. Insults won't be tolerated. Flame wars are frowned upon.

  2. No spam posting.

  3. Posts have to be centered around self-hosting. There are other communities for discussing hardware or home computing. If it's not obvious why your post topic revolves around selfhosting, please include details to make it clear.

  4. Don't duplicate the full text of your blog or github here. Just post the link for folks to click.

  5. Submission headline should match the article title (don’t cherry-pick information from the title to fit your agenda).

  6. No trolling.

Resources:

Any issues on the community? Report it using the report flag.

Questions? DM the mods!

founded 2 years ago
MODERATORS
 

Hello !

I currently have a problem on my kubernetes cluster.

I have 3 nodes:

  • 192.168.0.16
  • 192.168.0.65
  • 192.168.0.55

I use a storage class nfs (sigs/nfs-subdir-external-provisioner) to use an NFS.

The NFS is actually set up on the 192.168.0.55 which is also a worker node then.

I noticed that i have problems mounting volumes when a pod is created on the 192.168.0.55 node. If its one of the other two, it mounts. (The error is actually a permission denied on the 192.168.0.55 node)

I would guess that something goes wrong when kube tries to mount to NFS since it’s on the same machine ?

Any idea on how i can fix this? Cheers !

all 3 comments
sorted by: hot top controversial new old
[–] theit8514@lemmy.world 6 points 9 months ago (1 children)

Could be trying to mount it loopback instead of by ip. What does your exports file look like? Can you do a mount from 192.168.0.55 manually?

[–] Burn1ngBull3t@lemmy.world 5 points 9 months ago* (last edited 9 months ago)

Hello @theit8514

You are actually spot on ^^

I did look in my exports file which was like so :/mnt/DiskArray 192.168.0.16(rw) 192.168.0.65(rw)

I added a localhost line in case: /mnt/DiskArray 127.0.0.1(rw) 192.168.0.16(rw) 192.168.0.65(rw)

It didn't solve the problem. I went to investigate with the mount command:

  • Will mount on 192.168.0.65: mount -t nfs 192.168.0.55:/mnt/DiskArray/mystuff/ /tmp/test

  • Will NOT mount on 192.168.0.55 (NAS): mount -t nfs 192.168.0.55:/mnt/DiskArray/mystuff/ /tmp/test

  • Will mount on 192.168.0.55 (NAS): mount -t nfs 127.0.0.1:/mnt/DiskArray/mystuff/ /tmp/test

The mount -t nfs 192.168.0.55 is the one that the cluster does actually. So i either need to find a way for it to use 127.0.0.1 on the NAS machine, or use a hostname that might be better to resolve

EDIT:

I was acutally WAY simpler.

I just added 192.168.0.55 to my /etc/exports file. It works fine now ^^

Thanks a lot for your help@theit8514@lemmy.world !