FAST 搜索爬网在本地文件系统上失败,错误为 0x80040d07

发布于 2024-10-20 19:15:58 字数 951 浏览 5 评论 0原文

我已在我的计算机上成功安装 Sharepoint 2010 和 FAST Search for Sharepoint 2010。爬取共享点站点的情况似乎工作正常,但爬取本地文件系统始终失败,并显示以下内容:

The start address unknown://win-fs/C:/Users/ravish/FAST_TEST_SET cannot be crawled.

Context: Application 'FAST_Content_SSA', Catalog 'Portal_Content'

Details:
    Item not crawled due to one of the following reasons: Preventive crawl rule; Specified content source hops/depth exceeded; URL has query string parameter; Required protocol handler not found; Preventive robots directive.   (0x80040d07)

我从事件视图应用程序日志中得到了此信息。我已确保权限设置为允许运行服务的用户进行读取、列出和执行访问。还尝试更新爬网规则等,使它们非常自由。

我在内容源中尝试过的 URL 是: //win-fs/C:/Users/ravish/FAST_TEST_SET 文件://win-fs/C:/Users/ravish/FAST_TEST_SET file:\win-fs\C:\Users\ravish\FAST_TEST_SET

到目前为止似乎没有任何效果。一直在网络上努力寻找答案,发现很多用户都有类似的问题,但没有具体的直接回复。我花了几个小时,所以任何帮助将不胜感激。

我的系统的详细信息: Windows Server 2008 R2 Sharepoint 企业版 2010 FAST Search 2010

(全部安装在一台笔记本电脑上)

I've successfully installed Sharepoint 2010 and FAST Search for Sharepoint 2010 on my machine. Things seem to work fine with crawl of sharepoint sites but crawling local file system keeps failing with below:

The start address unknown://win-fs/C:/Users/ravish/FAST_TEST_SET cannot be crawled.

Context: Application 'FAST_Content_SSA', Catalog 'Portal_Content'

Details:
    Item not crawled due to one of the following reasons: Preventive crawl rule; Specified content source hops/depth exceeded; URL has query string parameter; Required protocol handler not found; Preventive robots directive.   (0x80040d07)

I got this from Event View Application log. I have made sure permissions are set to allow read, list and execute access to users running the services. Also tried updating crawl rules etc. to make them very liberal.

The URLs I have tried in Content Source are:
//win-fs/C:/Users/ravish/FAST_TEST_SET
file://win-fs/C:/Users/ravish/FAST_TEST_SET
file:\win-fs\C:\Users\ravish\FAST_TEST_SET

Nothing seems to work so far. Have been toiling the web trying to find answers, found many users with similar problems but nothing concrete with a straight reply to it. I've spent hours so any help will be appreciated.

Details of my system:
Windows Server 2008 R2
Sharepoint Enterprise 2010
FAST Search 2010

(all installed on single laptop)

如果你对这篇内容有疑问,欢迎到本站社区发帖提问 参与讨论,获取更多帮助,或者扫码二维码加入 Web 技术交流群。

扫码二维码加入Web技术交流群

发布评论

需要 登录 才能够评论, 你可以免费 注册 一个本站的账号。

评论(2

ゞ记忆︶ㄣ 2024-10-27 19:15:58

我没有使用过sharepoint版本的fast,但我认为爬虫与之前版本的独立fast中的爬虫大致相同。在这种情况下 - 我猜想它本身将无法抓取文件系统,因为它是为了通过 http 和 http 进行处理和抓取而构建的。 html...解析页面、跟踪链接等。

您也许可以设置本地网络服务器&允许目录索引,然后以这种方式抓取“站点”。或者您可以使用 Everything 中嵌入的网络服务器,然后再次...抓取本地站点(或所需的子集)。

I haven't used the sharepoint version of fast, but i think the crawler is mostly the same as the one in previous version of standalone fast. in which case - i'd guess that by itself it won't be able to crawl the filesystem as it's built to process and crawl via http & html... parsing pages, following links, etc.

You might be able to setup a local webserver & allow directory indexes and then crawl the 'site' that way. or you might be able to use the webserver embedded in Everything and again... crawl the local site (or desired subset).

故人爱我别走 2024-10-27 19:15:58

好吧,我想我现在明白了,重新启动机器并尝试以下格式,瞧,它成功了!

file://WIN-FS/Users/ravish/FAST_TEST_SET

不敢相信 SP 对路径有多么挑剔!

OK I think I got it now, rebooted the machine and tried following format and voila it worked!

file://WIN-FS/Users/ravish/FAST_TEST_SET

Can't believe how fussy SP is about paths!

~没有更多了~
我们使用 Cookies 和其他技术来定制您的体验包括您的登录状态等。通过阅读我们的 隐私政策 了解更多相关信息。 单击 接受 或继续使用网站,即表示您同意使用 Cookies 和您的相关数据。
原文