Petalbot |
Petalbot |
Brian Chandler |
Jan 13 2022, 11:47 PM
Post
#1
|
Jocular coder Group: Members Posts: 2,460 Joined: 31-August 06 Member No.: 43 |
My error log is full of accesses to the nonexistent https://imaginatorium.com/addbskt.php from something identifying itself as Petalbot. This links to a page here:
https://webmaster.petalsearch.com/site/petalbot This explains that Petalbot follows the robots.txt protocol, and describes how to block it by (e.g.) CODE User-agent: PetalBot Disallow: /*.php But https://imaginatorium.com/robots.txt already includes CODE User-agent: * Allow: /*.html Disallow: /*.php Unless I misunderstand something, if Petalbot followed the robots.txt protocol it would not attempt to access this page. Or do I have to go around adding in the names of all the robots I want to exclude? |
Christian J |
Jan 14 2022, 08:41 AM
Post
#2
|
. Group: WDG Moderators Posts: 9,686 Joined: 10-August 06 Member No.: 7 |
My error log is full of accesses to the nonexistent https://imaginatorium.com/addbskt.php from something identifying itself as Petalbot. How did it find addbskt.php, has that page existed previously? If Petalbot is just guessing URLs I would regard it as malicious and ban. QUOTE CODE User-agent: * Allow: /*.html Disallow: /*.php According to http://www.robotstxt.org/robotstxt.html things like "/*.php" seems invalid: QUOTE Specifically, you cannot have lines like "User-agent: *bot*", "Disallow: /tmp/*" or "Disallow: *.gif". Also, an Allow field does not exist, except as an extension: https://en.wikipedia.org/wiki/Robots_exclus...Allow_directive (I have no idea if Petalbot supports it). Personally I use a spider trap that auto-bans anyone that does not follow my robots.txt directives. Usually there are a few catches every month. |
Brian Chandler |
Jan 15 2022, 02:19 AM
Post
#3
|
Jocular coder Group: Members Posts: 2,460 Joined: 31-August 06 Member No.: 43 |
Thanks Christian. I don't know where I got my version from; everywhere generally says something different...
My error log is full of accesses to the nonexistent https://imaginatorium.com/addbskt.php from something identifying itself as Petalbot. How did it find addbskt.php, has that page existed previously? If Petalbot is just guessing URLs I would regard it as malicious and ban. There is an addbskt.php on the old site, and I suppose at some stage during development I had references to it on the new site. I intended to have robots banned at that stage, but who knows. It seemed neat just to block all php pages, which are 'operations' not appropriate for search targeting, and allow all html pages, which are 'content'. I can just make an explicit list of *.php. Petalbot does seem to be a bit slow, because it is trawling though pages from months back, but in the end that's not a problem. |
Lo-Fi Version | Time is now: 15th June 2024 - 01:40 PM |