Crawl just stops

Need a little help

Got a Content Source that crawls a Custom Repository - the protocol handler is installed and all worked earlier on..

Now when I start a Full Crawl, it just stop after 2 min.

It's not the process (mssdmn.exe) that doing a crash - nothing to see in the event log, nor the crawl log  - the crawl log just have warning, no top-level error, errors at all

Suddenly the other day, it crawled for almost an hour and crawled over 100.000 items for then just to stop

now we're back to where it stops after 2 min - and the +100.000 items is moved to Deletes in the crawl log.

Under warnings I see this:

Item not crawled due to one of the following reasons: Preventive crawl rule; Specified content source hops/depth exceeded; URL has query string parameter; Required protocol handler not found; Preventive robots directive. ( This item was deleted because it was excluded by a crawl rule. )

Nothing has changed on the server nor on the repository source.   

January 23rd, 2015 9:59am

Check your crawl rules; because there appears to be one that is preventing the crawl.  And with crawl rules when one is met, the rules are not processed further.  So, in this case the preventing rule is stopping the crawl.

there is a 256 character limit

http://support.microsoft.com/kb/894630/en-us?p=1


http://technet.microsoft.com/en-us/library/ff919564(v=office.14).aspx

you could enumerate your webs and find it, or it might be a list item, or something else,

Or, you have some type of file in the SharePoint that was at one point allowed but is now not allowed.


Free Windows Admin Tool Kit Click here and download it now
January 24th, 2015 5:07pm

Hi Stacy 

I've no crawl rules defined.. I'll look into the file type thing.. 

January 25th, 2015 8:38am

Hey Jm,

Also check your site for links that have a search query, as part of the link,

"URL has query string parameter;"  and start looking in the site that is associated with the

Free Windows Admin Tool Kit Click here and download it now
January 25th, 2015 3:58pm

is there a way to find this quick  there's a lot of Site collections, sites and subsites...  
January 25th, 2015 7:13pm

there should be a link in your crawl log to the item it stopped on
Free Windows Admin Tool Kit Click here and download it now
January 25th, 2015 9:58pm

Hi JmATK,

Based on the error message, it indicates that the item is excluded from the crawl rules, so I recommend to use PowerShell command to re-check if there are any crawl rules:

Get-SPEnterpriseSearchCrawlRule SearchApplication Search Service Application.

Please also make sure that the default content access account has Full Read permission on the web application and disable loopback checking.(please refer to this link).

Thanks,

Victoria

Forum Support

Please remember to mark the replies as answers if they help and unmark them if they provide no help. If you have feedback for TechNet Subscriber Support, contact tnmff@microsoft.com.

January 26th, 2015 1:29pm

You might try enabling BackConnectionHostNames too http://support.microsoft.com/kb/896861
Free Windows Admin Tool Kit Click here and download it now
January 26th, 2015 2:09pm

No crawl rules

and BackConnectionHostNames - that's already done.. 

  • Edited by JmATK Monday, January 26, 2015 12:48 PM
January 26th, 2015 3:48pm

No crawl rules

and BackConnectionHostNames - that's already done.. 

  • Edited by JmATK Monday, January 26, 2015 12:48 PM
Free Windows Admin Tool Kit Click here and download it now
January 26th, 2015 3:48pm

No crawl rules

and BackConnectionHostNames - that's already done.. 

  • Edited by JmATK Monday, January 26, 2015 12:48 PM
January 26th, 2015 3:48pm

If you're not using siteDataServer, make sure you have an entry in the hosts file for each web app that is getting crawled, and then also test access to the web apps from the server.
Free Windows Admin Tool Kit Click here and download it now
January 26th, 2015 7:40pm

This topic is archived. No further replies will be accepted.

Other recent topics Other recent topics