Jump to content

Recommended Posts

Posted

I was wondering why there are so many dead links in the directory?

 

A few days ago I downloaded the raw directory data and loaded it into our web safe bot, it searches for... (popups, unsafe script, 3rd party cookies, tracking).

 

Now I am about 60% done with fetching all the pages in the directory, but I have over 50,000 pages so far, that return 404. I have personally tested half the ones that return 404 and all of them are real 404 returns!

 

So my question is...

 

When is the directory cleared of 404(s) and do you even check links before releasing the new data. I ask this because over 80% of the dead links have been in previous raw data releases.

 

Thanks

 

j

  • Meta
Posted

We have a linkchecker called RoboZilla which runs from time to time (invoked manually, no fixed schedule). As you can imagine it takes a while to finish and needs a lot of server time while running.

 

Or technical staff is aware that the situation could be better than it is today. But I can't offer a timeframe when the situation may change.

Curlie Meta/kMeta Editor windharp

 

d9aaee9797988d021d7c863cef1d0327.gif

Create an account or sign in to comment

You need to be a member in order to leave a comment

Create an account

Sign up for a new account in our community. It's easy!

Register a new account

Sign in

Already have an account? Sign in here.

Sign In Now
×
×
  • Create New...