Hello! I have a server that runs 24/7, and have recently started doing some stuff that requires scraping the web. The websites are detecting the server’s IP to not be residential though, and it’s causing issues.

I’d like to host a proxy server on the small server I have running 24/7 in my house, so that everything for that 1 page could be proxied through it. Does anyone have any idea how I’d set up a server like that? Thanks.

  • Anafroj
    link
    fedilink
    arrow-up
    9
    ·
    edit-2
    1 year ago

    Max-P already provided good options, but I have to ask what I, and probably other people, wonder : why don’t you just run that scrapping program from your home server, then?

      • Anafroj
        link
        fedilink
        arrow-up
        3
        ·
        1 year ago

        (sorry for the double post, the instance I’m on was throwing errors)

        Gotcha, thanks for satisfying my curiosity. :) Of course, you can plug a usb drive on the Pi, but you know better what your needs are. Good luck!

        • neoney@lemmy.neoney.devOP
          link
          fedilink
          arrow-up
          2
          ·
          1 year ago

          I’m honestly planning to stop using the Pi today, it’s been unstable and I don’t like Raspbian, but I decided it’s not worth it to reinstall after getting 3 corrupted SD cards and just bought a used thin client which will replace it.

          • Anafroj
            link
            fedilink
            arrow-up
            1
            ·
            edit-2
            1 year ago

            I feel you, been there. :) I now use Gentoo on my Pi and it is stable, but I can’t recommend that to anyone who is not already used to Gentoo, it’s challenging to install it by itself.

            Regarding the SD card, I have no problem anymore since I stopped using the cheapest brands. I now use only Sandisk Ultra microSDXC, and the oldest ones have been working for four years without issue. It’s still basically a NAND (same stuff than in SSD drives) soldered on pins, though, so it’s very fragile. Care should be taken to neve bend them : they look flexible, but the NAND really isn’t.

            It’s also a good idea to backup the whole card. As they usually weight way less than hard drives, it’s easy to backup on your system and flash them back, mounting the sdcard on your desktop/laptop:

            lsblk # find the device name, let's say it's mmcblk1 
            dd if=/dev/mmcblk1 of=./backup-file bs=1G  # making a backup
            dd if=./backup-file of=/dev/mmcblk1 bs=1G  # restoring the backup
            

            if means “input file”, of means “output file” and bs is the buffer size (how many bytes are copied at once, the more the faster, but it will use that amount or RAM at each iteration). dd is just copying input to output, bs bytes by bs bytes.

            If you do that regularly, even using cheap sdcard that fail after a year will be less of a setback : you can just flash the last saved version of the system on a new card. It’s probably better, though, to keep only the OS on the sdcard, and store important daily updated data on a usb drive or key.

  • karlthemailman
    link
    fedilink
    arrow-up
    6
    ·
    1 year ago

    How much are you scraping? You may end up getting your home up blocked.

  • Max-P@lemmy.max-p.me
    link
    fedilink
    English
    arrow-up
    5
    ·
    1 year ago

    You can pretty easily install Squid, it’s fairly simple to configure and works well for most use cases. Just a plain simple HTTP proxy.

    You could also set up a VPN to your home to achieve something similar, by binding some requests to the VPN IP. It’s a bit harder to set up however as it involves routing tables, route metrics and conditionally binding the outgoing connection to a specific interface

    • neoney@lemmy.neoney.devOP
      link
      fedilink
      English
      arrow-up
      2
      ·
      1 year ago

      Thanks, sounds like Squid will be perfect. I’ll just need to figure out some way to connect. I wish I could just open a port, but it hasn’t been working since I enabled IPv6 on my router. Do you think I could make it accessible through cloudflare tunnels?

      • Max-P@lemmy.max-p.me
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        Cloudflare tunnels won’t work as Cloudflare won’t tunnel HTTP proxy traffic, at least as far as I know.

        What you can do however is have your home server VPN into your remote server, then your remote server will have no problem connecting to Squid over the VPN link. WireGuard is very simple to configure like that, probably 5-10 lines of config on each end. You don’t need any routing or forwarding or anything, just a plain VPN with 2 peers that can ping eachother, so no ip_forward or iptables -j MASQUERADE needed or anything that most guides would include. You can also use something like Tailscale, anything that will let the two machines talk to eachother.

        Depending on the performance and reliability needs, you could even just forward a port with SSH. Connect to your remote server from the home server with something like ssh -N -R localhost:8088:localhost:8080 $remoteServer and port 8088 on the remote will forward to port 8080 on the home server as long as that SSH connection is up. -N simply makes SSH not open a shell on the remote, dedicating the SSH session to the forwarding. Nice and easy, especially for prototyping.

        • neoney@lemmy.neoney.devOP
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 year ago

          That seems overcomplicated for me honestly, but now I just thought that I actually can host the scraper on the home server, as the scraper itself only scrapes simple data, and the downloads are by a separate program.

          • neoney@lemmy.neoney.devOP
            link
            fedilink
            English
            arrow-up
            1
            ·
            1 year ago

            The downloader talks to the scraper through HTTP, which I can publish through CF Tunnels, so it’s perfect.

  • Illecors@lemmy.cafe
    link
    fedilink
    English
    arrow-up
    4
    ·
    edit-2
    1 year ago

    Besides other answers - you could just use SSH port forwarding. Remote would be your home server, local would be your “cloud” server. You should initiate the connection from the cloud server to your home server. Playing with local ports would enable you filter what domains are used for the proxy.

    I rarely use it, so the exact syntax is gone from my memory. It is a bit tricky at first, but definitely not rocket science to figure out.

    Once the connection is established - you would point your scraper to https://localhost:localport

    • lynny@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      This is the most simple solution and probably a lot safer than the alternatives. Another good option would be to use OpenVPN.

  • neoney@lemmy.neoney.devOP
    link
    fedilink
    arrow-up
    1
    ·
    1 year ago

    I ended up hosting the scraper service on my home server and exposing it through Cloudflare Tunnels, as the service is pretty much just an API that doesn’t really do much data downloads