nutch-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From zo tiger <>
Subject Help me, No urls to fetch.
Date Wed, 02 Sep 2009 10:36:15 GMT


I have installed nutch1.0 and hadoop1.9 successfully.

There is no error.

I followed a tutorial 

I used three multinode, one is master node and other two are slave nodes.

But I just do crawled , no data showed.

$ bin/nutch crawl urls -dir crawled -depth 3
crawl started in: crawled
rootUrlDir = urls
threads = 10
depth = 3
Injector: starting
Injector: crawlDb: crawled/crawldb
Injector: urlDir: urls
Injector: Converting injected urls to crawl db entries.
Injector: Merging injected urls into crawl db.
Injector: done
Generator: Selecting best-scoring urls due for fetch.
Generator: starting
Generator: segment: crawled/segments/20090902102133
Generator: filtering: true
Generator: 0 records selected for fetching, exiting ...
Stopping at depth=0 - no more URLs to fetch.
No URLs to fetch - check your seed list and URL filters.
crawl finished: crawled

At last i ran bin/nutch crawl command but it gives

No urls to fetch check your filter and seed list error 

I am sure there is no problem in crawl-url filter and other configuration
xml files

İs anyone know any possible problem???? 

help me...
View this message in context:
Sent from the Nutch - User mailing list archive at

View raw message