yum-jelly Posted November 25, 2005 Posted November 25, 2005 I was wondering why there are so many dead links in the directory? A few days ago I downloaded the raw directory data and loaded it into our web safe bot, it searches for... (popups, unsafe script, 3rd party cookies, tracking). Now I am about 60% done with fetching all the pages in the directory, but I have over 50,000 pages so far, that return 404. I have personally tested half the ones that return 404 and all of them are real 404 returns! So my question is... When is the directory cleared of 404(s) and do you even check links before releasing the new data. I ask this because over 80% of the dead links have been in previous raw data releases. Thanks j
Meta windharp Posted November 25, 2005 Meta Posted November 25, 2005 We have a linkchecker called RoboZilla which runs from time to time (invoked manually, no fixed schedule). As you can imagine it takes a while to finish and needs a lot of server time while running. Or technical staff is aware that the situation could be better than it is today. But I can't offer a timeframe when the situation may change. Curlie Meta/kMeta Editor windharp
Recommended Posts
Create an account or sign in to comment
You need to be a member in order to leave a comment
Create an account
Sign up for a new account in our community. It's easy!
Register a new accountSign in
Already have an account? Sign in here.
Sign In Now