• jarfil@beehaw.org
    link
    fedilink
    arrow-up
    2
    ·
    edit-2
    4 months ago

    URL shorteners in general, or just Google?

    https://wiki.archiveteam.org/index.php/URLTeam

    Goo.gl has a namespace for 10 billion entries, it used to keep tracking/analytics data for each link, with a user interface, and it would happily generate them for stuff like Google Maps links.

    How much money would you say it takes to even maintain a system like that, plus update its security, not to mention account for changing web standards, at that scale?

    • jonne@infosec.pub
      link
      fedilink
      arrow-up
      4
      ·
      4 months ago

      Probably like $50/month in cloud resources if you turned off all the extra stuff and only did redirects and kept it around in read only mode. You’d need to do some dev work up front and price that in as well, obviously.

      • jarfil@beehaw.org
        link
        fedilink
        arrow-up
        4
        ·
        edit-2
        4 months ago

        $50/month would barely scratch the surface.

        Let’s take a conservative approach, and say there are:

        • only 1 billion links
        • each link only points to a URL of up to 100 characters in length on average (some will be 1000 or longer, but let’s hope some are 50 or shorter)
        • less than 10 billion daily hits total (that’s an average of 10/link)
        • the response time should be well under 50ms.

        Now you’re looking at 100GB of raw data to put into a database, that needs to return 100K answers/second, in less than 50ms each, worldwide, 24/7.

        What is your estimated cloud cost for something like 256GB of RAM, 128 cores, 10Gbps connect, replicated across several zones, and 1TB/day outgoing transfer?

        That’s only for the redirect responses in read-only mode, nothing else. You will also need some maintenance to keep it 24/7, for when the server catches fire, or gets obsoleted, and when new exploits come up against your software stack.