fast-esp

Robots.txt to disallow everything and allow only specific parts of the site/pages. Is “allow” supported by crawlers like Ultraseek and FAST?

左心房为你撑大大i 提交于 2019-12-23 12:03:58
问题 Just wanted to know if it is possible to disallow the whole site for crawlers and allow only specific webpages or sections? Is "allow" supported by crawlers like FAST and Ultraseek? 回答1: There is an Allow Directive however there's no guarantee that a particular bot will support it (much like there's no guarantee a bot will even check your robots.txt to begin with). You could probably tell by examining your weblogs whether or not specific bots were indexing only the parts of your website that

Google Search, FAST ESP and Lucene

孤街浪徒 提交于 2019-12-11 02:23:24
问题 Has anybody come across or created a functionality comparison document between the various enterprise search options like Google Search Appliance, Fast ESP, Lucene? Any helpful pointers or links much appreciated. 回答1: Strangely enough I have, sort of. It's so dependent on your use case and $$$ requirements though that such an open ended question is pretty useless. If you need XML indexing, with things like XPath type queries, then FAST is the only way to go. Trouble is, it costs &&&. If you