Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- bin/nutch index http://localhost:8983/solr crawl/crawldb/ -linkdb crawl/linkdb/ crawl/segments/*
- The input path at crawldb is not a segment... skipping
- Segment dir is complete: crawl/segments/20161214143435.
- Segment dir is complete: crawl/segments/20161214144230.
- Indexer: starting at 2016-12-15 10:55:35
- Indexer: deleting gone documents: false
- Indexer: URL filtering: false
- Indexer: URL normalizing: false
- Active IndexWriters :
- SOLRIndexWriter
- solr.server.url : URL of the SOLR instance
- solr.zookeeper.hosts : URL of the Zookeeper quorum
- solr.commit.size : buffer size when sending to SOLR (default 1000)
- solr.mapping.file : name of the mapping file for fields (default solrindex-mapping.xml)
- solr.auth : use authentication (default false)
- solr.auth.username : username for authentication
- solr.auth.password : password for authentication
- Indexer: java.io.IOException: No FileSystem for scheme: http
- at org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2385)
- at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2392)
- at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:89)
- at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2431)
- at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2413)
- at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:368)
- at org.apache.hadoop.fs.Path.getFileSystem(Path.java:296)
- at org.apache.hadoop.mapred.FileInputFormat.singleThreadedListStatus(FileInputFormat.java:256)
- at org.apache.hadoop.mapred.FileInputFormat.listStatus(FileInputFormat.java:228)
- at org.apache.hadoop.mapred.SequenceFileInputFormat.listStatus(SequenceFileInputFormat.java:45)
- at org.apache.hadoop.mapred.FileInputFormat.getSplits(FileInputFormat.java:304)
- at org.apache.hadoop.mapreduce.JobSubmitter.writeOldSplits(JobSubmitter.java:520)
- at org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:512)
- at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:394)
- at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1285)
- at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1282)
- at java.security.AccessController.doPrivileged(Native Method)
- at javax.security.auth.Subject.doAs(Subject.java:422)
- at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
- at org.apache.hadoop.mapreduce.Job.submit(Job.java:1282)
- at org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:562)
- at org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:557)
- at java.security.AccessController.doPrivileged(Native Method)
- at javax.security.auth.Subject.doAs(Subject.java:422)
- at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
- at org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:557)
- at org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:548)
- at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:833)
- at org.apache.nutch.indexer.IndexingJob.index(IndexingJob.java:145)
- at org.apache.nutch.indexer.IndexingJob.run(IndexingJob.java:228)
- at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
- at org.apache.nutch.indexer.IndexingJob.main(IndexingJob.java:237)
- bin/nutch solrindex http://localhost:8983/solr crawl/crawldb/ -linkdb crawl/linkdb/ crawl/segments/*
- Segment dir is complete: crawl/segments/20161214143435.
- Segment dir is complete: crawl/segments/20161214144230.
- Indexer: starting at 2016-12-15 10:54:07
- Indexer: deleting gone documents: false
- Indexer: URL filtering: false
- Indexer: URL normalizing: false
- Active IndexWriters :
- SOLRIndexWriter
- solr.server.url : URL of the SOLR instance
- solr.zookeeper.hosts : URL of the Zookeeper quorum
- solr.commit.size : buffer size when sending to SOLR (default 1000)
- solr.mapping.file : name of the mapping file for fields (default solrindex-mapping.xml)
- solr.auth : use authentication (default false)
- solr.auth.username : username for authentication
- solr.auth.password : password for authentication
- Indexing 250/250 documents
- Deleting 0 documents
- Indexing 250/250 documents
- Deleting 0 documents
- Indexer: java.io.IOException: Job failed!
- at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:836)
- at org.apache.nutch.indexer.IndexingJob.index(IndexingJob.java:145)
- at org.apache.nutch.indexer.IndexingJob.run(IndexingJob.java:228)
- at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
- at org.apache.nutch.indexer.IndexingJob.main(IndexingJob.java:237)
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement