One of Google Search’s oldest and best-known features, cache links, are being retired. Best known by the “Cached” button, those are a snapshot of a web page the last time Google indexed it. However, according to Google, they’re no longer required.

“It was meant for helping people access pages when way back, you often couldn’t depend on a page loading,” Google’s Danny Sullivan wrote. “These days, things have greatly improved. So, it was decided to retire it.”

    • aname@lemmy.one
      link
      fedilink
      English
      arrow-up
      153
      ·
      9 months ago

      They are an Ad company, and using cached page doesn’t bring ad money to their clients

      • kratoz29@lemm.ee
        link
        fedilink
        English
        arrow-up
        8
        ·
        9 months ago

        Make sense, it seems that they have been having lots of meetings regarding how to maximize its revenue

    • lemmyvore@feddit.nl
      link
      fedilink
      English
      arrow-up
      18
      ·
      9 months ago

      They may not have a choice in the matter. AI-generated pages are set to completely destroy the noise to signal ratio on the web.

      Google’s business has two aspects, collecting user data and serving ads. If Search stops being relevant people will stop using it, which impacts both aspects negatively.

  • NoRodent@lemmy.world
    link
    fedilink
    English
    arrow-up
    187
    ·
    9 months ago

    Well that really sucks because it was often the only way to actually find the content on the page that the Google results “promised”. For numerous reasons - sometimes the content simply changes, gets deleted or is made inaccessible because of geo-fencing or the site is straight up broken and so on.

    Yes, there’s archive.org but believe it or not, not everything is there.

  • Toes♀@ani.social
    link
    fedilink
    English
    arrow-up
    147
    arrow-down
    1
    ·
    9 months ago

    That’s bs, it’s one of the best features Google has and they’ve been ruining it. Wayback machine wished it could be that comprehensive.

  • _number8_@lemmy.world
    link
    fedilink
    English
    arrow-up
    118
    ·
    9 months ago

    of course it is. why have anything good on there, no point reminding me of the old days when the internet was actually fucking useful

      • Saik0@lemmy.saik0.com
        link
        fedilink
        English
        arrow-up
        48
        ·
        9 months ago

        Literally yesterday. What source is sufficient to tell you first hand that I used the feature yesterday?

        You want proof that it’s useful. Go look at waybackmachine. Literally millions of users using a cached web page feature.

        • Jakeroxs@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          23
          ·
          9 months ago

          I also literally used it yesterday, mostly because my work has an insanely over the top site blocking situation, and rather then having to input (and likely get a rejection) to allow the site, cached page usually works good and gets me the info I need.

          • Kite@sh.itjust.works
            link
            fedilink
            English
            arrow-up
            9
            ·
            9 months ago

            That is exactly why I use it. I need to access pages for work, our internet security is ridiculously overdone and so many sites don’t load… but the cached versions do. Fml

        • Guru_Insights99@lemm.ee
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          34
          ·
          9 months ago

          Photo / visual evidence would be fine, I am not picky. I would just like to be sure you are telling the truth, a lot of fraud on the internet nowadays 😒😒

      • Emerald@lemmy.world
        link
        fedilink
        English
        arrow-up
        14
        ·
        edit-2
        9 months ago

        I last used the feature to view deleted reddit posts.

        Another time I used something similar (the wayback machine) to view long gone websites about a postcard

      • nixcamic@lemmy.world
        link
        fedilink
        English
        arrow-up
        7
        ·
        9 months ago

        I’ve used it three times today. Site down, geo-blocked, and a forum post with info I needed deleted.

  • Raiderkev@lemmy.world
    link
    fedilink
    English
    arrow-up
    115
    arrow-down
    3
    ·
    9 months ago

    Without getting into too much detail, a cached site saved my ass in a court case. Fuck you Google.

    • lud@lemm.ee
      link
      fedilink
      English
      arrow-up
      14
      ·
      9 months ago

      It sucks because it’s sometimes (but not very often) useful but it’s not like they are under any obligation to support it or are getting any money from doing it.

        • megaman@discuss.tchncs.de
          link
          fedilink
          English
          arrow-up
          8
          ·
          9 months ago

          At least some of these tools change their “user agent” to be whatever google’s crawler is.

          When you browse in, say, Firefox, one of the headers that firefox sends to the website is “I am using Firefox” which might affect how the website should display to you or let the admin knkw they need firefox compatibility (or be used to fingerprint you…).

          You can just lie on that, though. Some privacy tools will change it to Chrome, since that’s the most common.

          Or, you say “i am the google web crawler”, which they let past the paywall so it can be added to google.

          • sfgifz@lemmy.world
            link
            fedilink
            English
            arrow-up
            2
            ·
            edit-2
            9 months ago

            Or, you say “i am the google web crawler”, which they let past the paywall so it can be added to google.

            If I’m not wrong, Google has a set range of IP addresses for their crawlers, so not all sites will let you through just because your UA claims to be Googlebot

        • lud@lemm.ee
          link
          fedilink
          English
          arrow-up
          5
          ·
          9 months ago

          I dunno, but I suspect that they aren’t using Google’s cache if that’s the case.

          My guess is that the site uses its own scrapper that acts like a search engine and because websites want to be seen to search engines they allow them to see everything. This is just my guess, so it might very well be completely wrong.

      • icedterminal@lemmy.world
        link
        fedilink
        English
        arrow-up
        10
        ·
        9 months ago

        Depends. Not every site, or its pages, will be crawled by the Internet Archive. Many pages are available only because someone has submitted it to be archived. Whereas Google search will typically cache after indexed.

    • Neato@ttrpg.network
      link
      fedilink
      English
      arrow-up
      8
      ·
      9 months ago

      I can’t imagine there was even that much lost revenue. Cached pages are good for seeing basic content in that page but you can’t click through links or interact with the page in any way. Were so many people using it to avoid ads?

      • NoRodent@lemmy.world
        link
        fedilink
        English
        arrow-up
        14
        ·
        edit-2
        9 months ago

        Were so many people using it to avoid ads?

        I doubt that as well. There are much better ways to deal with ads. I always only used it when the content on the page didn’t exist anymore or couldn’t be accessed for whatever reason.

        But I suspected this was coming, they’ve been hiding this feature deeper and deeper in the last few years.

      • db2@lemmy.world
        link
        fedilink
        English
        arrow-up
        4
        ·
        9 months ago

        but you can’t click through links or interact with the page in any way

        Most of the time that’s exactly what I want. I hate hunting through 473 pages of stupid bullshit in some janky forum to try to find the needle in that haystack.

      • bjorney@lemmy.ca
        link
        fedilink
        English
        arrow-up
        1
        ·
        edit-2
        9 months ago

        I feel like 99% of its usage was to avoid ads/paywalls/geo/account restrictions on news and social media sites

  • _haha_oh_wow_@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    59
    arrow-down
    2
    ·
    9 months ago

    The enshittification will continue until quarterly reports improve.

    Just kidding, it will continue regardless.

  • rhabarba@feddit.de
    link
    fedilink
    English
    arrow-up
    57
    ·
    9 months ago

    These days, things have greatly improved.

    Websites will never change their URLs today.

    • ares35@kbin.social
      link
      fedilink
      arrow-up
      21
      ·
      9 months ago

      i maintain redirects for old URLs for which the content still exists at another address. i’ve been doing that since i started working on web sites 20-some years ago. not many take the time to do that, but i do. so there’s at least a few web sites out there that if you have a 20 year old bookmark to, chances are it still works.

  • NoRodent@lemmy.world
    link
    fedilink
    English
    arrow-up
    50
    ·
    edit-2
    9 months ago

    By they way, I just found out that they removed the button, but typing cache:www.example.com into Google still redirects you to the cached version (if it exists). But who knows for how long. And there’s the question whether they’ll continue to cache new pages.

    • _number8_@lemmy.world
      link
      fedilink
      English
      arrow-up
      35
      ·
      9 months ago

      they’ve broken / ignored every modifier besides site: in the last few years, god knows how long that’ll work

      • Raiderkev@lemmy.world
        link
        fedilink
        English
        arrow-up
        9
        ·
        9 months ago

        Quotes are fucking awful now. You have to change the search terms to verbatim now which takes way fucking longer. Google has enshittified almost everything. I’m just waiting for them to ruin Maps.

        • Flax@feddit.uk
          link
          fedilink
          English
          arrow-up
          3
          ·
          9 months ago

          Remember when Google Now was intelligently selected data and not an endless scroll of paywalled news articles?

    • Appoxo@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      6
      ·
      9 months ago

      I hope they only kill the announced feature but keep the cache part.
      Just today I had to use it because some random rss aggregator website had the search result I wanted but redirected me somewhere completely different…

    • AnAngryAlpaca@feddit.de
      link
      fedilink
      English
      arrow-up
      5
      ·
      9 months ago

      My guess is that a cached page is just a byproduct when the page is indexed by the crawler. The need a local copy to parse text, links etc. and see the difference to the previous page.

  • 🇰 🌀 🇱 🇦 🇳 🇦 🇰 ℹ️@yiffit.net
    link
    fedilink
    English
    arrow-up
    49
    ·
    9 months ago

    It was meant for helping people access pages when way back, you often couldn’t depend on a page loading,” Google’s Danny Sullivan wrote. “These days, things have greatly improved. So, it was decided to retire it."

    They still go down, Danny. And fairly frequently at that. Y’all are fuckin’ stupid.

    • merc@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      11
      ·
      9 months ago

      I’d say things are much worse than they used to be. Sure, in the past sites would disappear or completely fail more often. But, because most sites were static, those were the only ways they could fail. These days the cache feature is useful for websites that have javascript bugs preventing them from displaying properly, or where the content-management-system still pretends the link works but where it silently just loads different content.

  • EnderMB@lemmy.world
    link
    fedilink
    English
    arrow-up
    50
    arrow-down
    3
    ·
    9 months ago

    How has no one worked on a new search engine over the last decade or so where Google has been on a clear decline in its flagship product!

    I know of the likes of DDG, and Bing has worked hard to catch up, but I’m genuinely surprised that a startup hasn’t risen to find a novel way of attacking reliable web search. Some will say it’s a “solved problem”, but I’d argue that it was, but no longer.

    A web search engine that crawls and searches historic versions of a web page could be an incredibly useful resource. If someone can also find a novel way to rank and crawl web applications or to find ways to “open” the closed web, it could pair with web search to be a genuine Google killer.

    • mlg@lemmy.world
      link
      fedilink
      English
      arrow-up
      20
      ·
      9 months ago
      • Google invents, invests, or previously invested into some ground breaking technology
      • They buy out competition and throw tons of effort into making superior product
      • Eventually Google becomes defacto standard
      • Like a few years pass
      • Google hands off project to fresh interns to reduce the crap out of the cloud usage to decrease cost
      • Any viable alternatives are immediately bought out by Google
      • Anything left over is either struggling FOSS or another crappy corporate attempt (cough cough Microsoft)
      • Repeat

      My favorite case in point being Google Maps.

    • OsrsNeedsF2P@lemmy.ml
      link
      fedilink
      English
      arrow-up
      17
      arrow-down
      2
      ·
      9 months ago

      There’s a lot of startups trying to make better search engines. Brave for example is one of them. There’s even one Lemmy user, but I forget what the name of theirs is.

      But it’s borderline impossible. In the old days, Google used webscrapers and key word search. When people started uploading the whole dictionary in white text on their pages, Google added some antispam and context logic. When that got beat, they handled web credibility by the number of “inlinks” from other websites. Then SEO came out to beat link farmers, and you know the rest from there.

      An indexable version of Archive.org is feasible, borderline trivial with ElasticSearch, but the problem is who wants that? Sure you want I may, but no one else cares. Also, let’s say you want to search up something specific - each page could be indexed, with slight differences, thousands of times. Which one will you pick? Maybe you’ll want to set your “search date” to a specific year? Well guess what, Google has that feature as well.

      • TWeaK@lemm.ee
        link
        fedilink
        English
        arrow-up
        8
        arrow-down
        1
        ·
        9 months ago

        Brave is not a business that should be supported. Also, I’m pretty sure they just use Bing for a back end.

        There are also a few paid search engines that people say are good.

          • TWeaK@lemm.ee
            link
            fedilink
            English
            arrow-up
            3
            ·
            9 months ago

            They’ve had a history of controversy over their life, ranging from replacing ads with their own affiliate links to bundling an opt-out crypto miner. Every time something like this happened, the CEO went on a marketing campaign across social media, effectively drowning out the controversial story with an influx of new users. The CEO meanwhile has got in trouble for his comments on same-sex marriage and covid-19.

            In general, it’s always seemed like it would take a very small sack of money for Brave to sell out its users. Also, their browser is Chromium based, so it’s still contributing to Google’s market dominance and dictatorial position over web technologies.

    • gunslingerfry@lemmy.world
      link
      fedilink
      English
      arrow-up
      12
      arrow-down
      3
      ·
      9 months ago

      I recommend Kagi. Bought a family plan and it feels like I’ve gone back to 2016 when the search engines weren’t a dumpster fire.

    • piecat@lemmy.world
      link
      fedilink
      English
      arrow-up
      14
      arrow-down
      8
      ·
      9 months ago

      The next revolutionary search engine will be an AI that understands you. Like what a librarian is… Not just ads served.

      • spujb@lemmy.cafe
        link
        fedilink
        English
        arrow-up
        30
        arrow-down
        1
        ·
        9 months ago

        i don’t need a search engine that understand me i need a search engine that finds sites and pages based on a string of text i provide it

        we should be calling the future piss the way it’s going down the toilet

        • piecat@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          1
          ·
          9 months ago

          Well, at the least, you need something to filter out the shit trying to game seo. To me it seems that AI is the easiest approach.

    • sgtgig@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      arrow-down
      1
      ·
      9 months ago

      Bing’s copilot is genuinely pretty good, the AI answer is often pretty accurate and the way it’s able to weave links into its answer is handy. I find it way more useful than Google search these days and I’m pretty much just using it on principle as Google is just pissing me off with killing their services, a few of which I’ve used.

      I don’t think Microsoft is some saint but copilot is just a good product.

    • AAA@feddit.de
      link
      fedilink
      English
      arrow-up
      2
      ·
      9 months ago

      Yes, that would be a Google killer. If you somehow find the money to provide it for free.

      Finding a novel way of searching is one thing. Finding a novel way of financing the whole endeavor (and not going the exact route Google is) is another.