• r00ty@kbin.life
    link
    fedilink
    arrow-up
    0
    ·
    5 months ago

    I’ve had to boot a remote server into rescue after locking myself out.

    I think most people have done this at least once.

  • Discover5164@lemm.ee
    link
    fedilink
    arrow-up
    0
    ·
    5 months ago

    i have ssh configured on a different port,

    more than one time i enabled ssh in ufw, restarted the service… and the connection dropped

  • JATtho@sopuli.xyz
    link
    fedilink
    arrow-up
    0
    ·
    5 months ago

    It happened to me when I was configuring IP geoblocking: Only whitelist IP ranges are allowed. That was fetched from a trusted URL. If the DNS provider just happened to not be on that list, the whitelist would become empty, blocking all IPs. Literally 100% proof firewall; not even a ping gets a pass.

  • PlexSheep@feddit.de
    link
    fedilink
    arrow-up
    0
    ·
    5 months ago

    What is a good firewall that can also block ports published with docker? I’d need it to run on the same host.

    • Crack0n7uesday@lemmy.world
      link
      fedilink
      arrow-up
      0
      ·
      edit-2
      5 months ago

      You want a virtual firewall. Is this for profit or just your science project because that’s going to change the answer. You might hate me, but I’m still gonna say it, Cisco…

    • dan@upvote.au
      link
      fedilink
      arrow-up
      0
      ·
      5 months ago

      Are your Docker containers connecting to the network (eg using ipvlan or macvlan)? The default bridge network driver doesn’t expose the container publicly unless you explicitly expose a port. If you don’t expose a port, the Docker container is only accessible from the host, not from any other system on the network.

        • dan@upvote.au
          link
          fedilink
          arrow-up
          0
          ·
          5 months ago

          If you don’t want the Docker container to be accessible from other systems then just don’t publish the port.

      • PlexSheep@feddit.de
        link
        fedilink
        arrow-up
        0
        ·
        5 months ago

        I remember trying with ufw and the docker ports were still open. Iirc I’ve read somewhere that docker and ufw both use the same underlying software, so ufw cannot block docker (IP tables?)

        • iiGxC@slrpnk.net
          link
          fedilink
          arrow-up
          0
          ·
          5 months ago

          Hmm, not sure. I know with docker you can “mock” ports for the container, where the port the container sees is different than the port on the system. Maybe you can do something with that?

    • derpgon@programming.dev
      link
      fedilink
      arrow-up
      0
      ·
      5 months ago

      UFW does work with Docker, but requires some tweaking. IIRC you have to disallow Docker to modify IPTables and then add a rule to forward all traffic to the Docker network of your choice. It’s a little finicky but works.

      • JasonDJ@lemmy.zip
        link
        fedilink
        arrow-up
        0
        ·
        5 months ago

        But…why?

        Project Calico is designed for segmenting network traffic between kubernetes workloads.

        Right tool for the job.

        Also if you are a Fortinet shop, supposedly you can manage rules with FortiManager. I haven’t tried that yet but it looks really cool.

    • treechicken@lemmy.worldOP
      link
      fedilink
      arrow-up
      0
      ·
      5 months ago

      That’s exactly what I did lol. Thankfully my Pi’s just in a drawer. If this was a remote host at work I would’ve already shat myself :P

    • marcos@lemmy.world
      link
      fedilink
      arrow-up
      0
      ·
      5 months ago

      You are assuming there is a keyboard and monitor plugged to it, and that the computer is somewhere nearby.

      None of those are automatically true. And when it’s nearby, it’s usually easier to just get the SD card into another computer and edit the configuration.

  • dan@upvote.au
    link
    fedilink
    arrow-up
    0
    ·
    5 months ago

    Not exactly the same thing, but on one of my systems, eth0 and eth1 swapped position after a kernel update, so the IP was on the wrong interface. I had IPMI/BMC on the system so didn’t have to physically go to it and plug in a keyboard and monitor, but I still had to deal with manually typing a long randomly-generated password, twice (one to log in and once again for sudo).

    I’m glad “predictable” interface names are supported now. Those eth0/eth1/etc names were dangerous since the numbers were just based on the order the kernel loaded the drivers and initialized them in, which can change across reboots. The predictable names are based on physical position in the system, so they’re consistent across reboots.

  • churisotophu@feddit.de
    link
    fedilink
    arrow-up
    0
    ·
    5 months ago

    This literally happened to me yesterday. Fortunately ufw enable did not configure it as persistent across reboots 🤠

  • 7heo@lemmy.ml
    link
    fedilink
    arrow-up
    0
    ·
    5 months ago

    ufw is not a good software. I really tried to work with it. My solution was to disable it.

    • sum_yung_gai@lemm.ee
      link
      fedilink
      arrow-up
      0
      ·
      5 months ago

      UFW is a software firewall. SSH is a way to remote into computers. The joke is they turned on UFW and got locked out of the machine.

      • vort3@lemmy.ml
        link
        fedilink
        arrow-up
        0
        ·
        5 months ago

        I’m pretty sure it was a joke.

        Everyone did this at some point, but nobody would admit such a silly thing happened to them.

        • oconnordaniel@infosec.pub
          link
          fedilink
          arrow-up
          0
          ·
          5 months ago

          Never done this to a single server.

          Managed to write the “ufw enable, deny all” part of ansislbe script without the “allow 22” part and run it against all my homelab once.

  • TimTamJimJam@lemmy.world
    link
    fedilink
    arrow-up
    0
    ·
    5 months ago

    Happened to me in work once… I was connected via SSH to one of our test machines, so I could test connection disruption handling on a product we had installed.

    I had a script that added iptables rules to block all ports for 30 seconds then unblock them. Of course I didn’t add an exception for port 22, and I didn’t run it with nohup, so when I ran the script it blocked the ports, which locked me out of SSH access, and the script stopped running when the SSH session ended so never unblocked the ports. I just sat there in awe of my stupidity.

    • krash@lemmy.ml
      link
      fedilink
      arrow-up
      0
      ·
      5 months ago

      Out of curiousity, how would nohup make your situation different? As I understand, nohup makes it possible to keep terminal applications running even when the terminal session has ended.

      • octopus_ink@lemmy.ml
        link
        fedilink
        English
        arrow-up
        0
        ·
        5 months ago

        If the script was supposed to wait 30 secs and then unblock the ports, running with nohup would have allowed the ports to be unblocked 30 secs later. Instead, the script terminated when the SSH session died, and never executed the countdown nor unblock.

      • aidan@lemmy.world
        link
        fedilink
        arrow-up
        0
        ·
        5 months ago

        the script stopped running when the SSH session ended so never unblocked the ports

      • Blackmist@feddit.uk
        link
        fedilink
        English
        arrow-up
        0
        ·
        5 months ago

        Ah, if only it was a server room and not a customer 3 hours drive away. And he’d closed and gone home for the night.

        Fortunately it just needed a reboot, and I was able to talk him through that in the morning.

        • SpaceCowboy@lemmy.ca
          link
          fedilink
          arrow-up
          0
          ·
          5 months ago

          Oof… well you can just say “it must be some hardware problem or something… maybe a reboot will fix it.”

        • JasonDJ@lemmy.zip
          link
          fedilink
          arrow-up
          0
          ·
          edit-2
          5 months ago

          Oof I did a firmware upgrade on my main external firewall.

          The upgrade went fine but when we added an ISP a month or so prior, I forgot to redistribute the ISPs routes. While all my ISPs were technically working, and the firewall came back up, nothing below it could get to the internet, so it was good as down.

          Cue the 1.5 hour drive into the office…

          Had that drive to think about what went wrong. Got into the main lobby, sat down, joined the wifi, and fixed it in 3 minutes.

          Didn’t even get to my desk or the datacenter.

        • JasonDJ@lemmy.zip
          link
          fedilink
          arrow-up
          0
          ·
          5 months ago

          Tmux essentially creates a pseudo-shell that persists between sessions.

          So you can start a process, detach the session, start something else, disconnect, come back next week, and check on it.

          It does other things too. Like console tiling.

        • u/lukmly013 💾 (lemmy.sdf.org)@lemmy.sdf.org
          link
          fedilink
          English
          arrow-up
          0
          ·
          edit-2
          5 months ago

          Well, the script could keep running even after he would have detached from that tmux session due to losing ssh connection. And since that script would unblock all ports after 30 seconds…

          (Same use case as nohup that they mentioned)