Steppenwolf Posted May 25, 2006 Report Share Posted May 25, 2006 I just found out that with some pretty simple google "hacks" it is very easy to find and filter thousands of unprotected mapfiles (*.vmt, *.map etc.), custom textures etc. I'm sure some of this stuff is for mods and by no means intented for public use. When you store backups of your files on web servers only upload them on pw protectet servers/folders or at least within pw protected zips or rars if you dont want that everybody has access to it. Quote Link to comment Share on other sites More sharing options...
dissonance Posted May 25, 2006 Report Share Posted May 25, 2006 or just make a no-robots file. http://www.robotstxt.org/wc/norobots.html eh, still, there are some search engines that don't obey those. Quote Link to comment Share on other sites More sharing options...
Section_Ei8ht Posted May 26, 2006 Report Share Posted May 26, 2006 no-robots files are rather useless these days, imho. Its not a standard, and its pretty easy to make a spider that will ignore those. Hell, even Google's ignore those. The only way to get Google spiders to stop hitting certain areas, or all of your page is to call them up and tell them to stop (They're actually really good about that sort of thing). Quote Link to comment Share on other sites More sharing options...
Minos Posted May 26, 2006 Report Share Posted May 26, 2006 Most of those map files are rubbish tho stepp Quote Link to comment Share on other sites More sharing options...
user Posted May 26, 2006 Report Share Posted May 26, 2006 so unless you either link to the files or to the folder, how do these bot-spiders even find the files anyways? Quote Link to comment Share on other sites More sharing options...
Minos Posted May 26, 2006 Report Share Posted May 26, 2006 You can search for specific file extesions, they don't necessarily need to be linked in a page for google to find them. Quote Link to comment Share on other sites More sharing options...
csharp Posted May 26, 2006 Report Share Posted May 26, 2006 when we did GTA:F we had passworprotected FTP, Every user had own folder, which was allso password protected (every one had own pw).. Quote Link to comment Share on other sites More sharing options...
skdr Posted May 26, 2006 Report Share Posted May 26, 2006 when we did GTA:F we had passworprotected FTP, Every user had own folder, which was allso password protected (every one had own pw).. That's awesome. Quote Link to comment Share on other sites More sharing options...
D3ads Posted May 26, 2006 Report Share Posted May 26, 2006 GTA:F ? Quote Link to comment Share on other sites More sharing options...
curman Posted May 26, 2006 Report Share Posted May 26, 2006 GTA:F ? Quote Link to comment Share on other sites More sharing options...
Thrik Posted May 27, 2006 Report Share Posted May 27, 2006 no-robots files are rather useless these days, imho. Its not a standard, and its pretty easy to make a spider that will ignore those. Hell, even Google's ignore those. The only way to get Google spiders to stop hitting certain areas, or all of your page is to call them up and tell them to stop (They're actually really good about that sort of thing). While you're right in saying that it's not a standard, I must point out that your claim of Google not obeying robots.txt is nonsense. Google does obey them; have a look in the official support database and you will see evidence of this, and Google even uses robots.txt on their own sites at times. Almost all respectable engines will obey robots.txt, only leaving the smaller ones that much fewer people use and various other crawlers such as those operated by Blog scourers. I'm not sure where exactly you got your facts, but they're wrong. But, still, don't rely on a robot denial file if absolute privacy is vital. Instead, use a .htaccess password (Google for “htpasswd tool” for an easy-to-use utility for creating one) or something else that's more physical. Quote Link to comment Share on other sites More sharing options...
csharp Posted May 27, 2006 Report Share Posted May 27, 2006 GTA:F = Grand Theft Auto: Finland.. but like allmost 90% of mods, this one died.. But atleast we made 0.1A (and 0.1B internal beta).. Damn we had great team, but we were lazy If someone wants 0.1b i can upload it (it just replaces couple of vehicles and gangs (every gang skin is made by me o//).. I think there were allso couple of weapons in 0.1b.. And yeah, it is to san andreas Quote Link to comment Share on other sites More sharing options...
Section_Ei8ht Posted May 27, 2006 Report Share Posted May 27, 2006 While you're right in saying that it's not a standard, I must point out that your claim of Google not obeying robots.txt is nonsense. Google does obey them; have a look in the official support database and you will see evidence of this, and Google even uses robots.txt on their own sites at times. Almost all respectable engines will obey robots.txt, only leaving the smaller ones that much fewer people use and various other crawlers such as those operated by Blog scourers. I'm not sure where exactly you got your facts, but they're wrong. Yeah, they probably are. I'm basing this off of something that happend at work a few years ago. A client site kept getting thrown offline because the spider would poll every single page in the inventory (poorly written asp code). Now that I think about it, the sysadmin there was probably too stupid to config the norobots file correctly. But it wasnt my job to fix it, I'm just going off of what he told me (You know, one of those bestfriend's-sister's-boyfriend's-uncle's-cousin's stories ) Quote Link to comment Share on other sites More sharing options...
Recommended Posts
Join the conversation
You can post now and register later. If you have an account, sign in now to post with your account.
Note: Your post will require moderator approval before it will be visible.