nutch index路径找不到问题记录

nutch 在抓取网页的时代报如下错误:

2015-06-30 16:25:53,420 DEBUG security.UserGroupInformation - PrivilegedActionException as:hadoop (auth:SIMPLE) cause:org.apache.hadoop.mapred.InvalidInputException: Input path does not exist: file:/app/apache-nutch-1.9/data/segments/20150630152657/parse_data/crawl_fetch
Input path does not exist: file:/app/apache-nutch-1.9/data/segments/20150630152657/parse_data/crawl_parse
Input path does not exist: file:/app/apache-nutch-1.9/data/segments/20150630152657/parse_data/parse_data
Input path does not exist: file:/app/apache-nutch-1.9/data/segments/20150630152657/parse_data/parse_text
Input path does not exist: file:/app/apache-nutch-1.9/data/segments/20150630152657/content/crawl_fetch
Input path does not exist: file:/app/apache-nutch-1.9/data/segments/20150630152657/content/crawl_parse
Input path does not exist: file:/app/apache-nutch-1.9/data/segments/20150630152657/content/parse_data
Input path does not exist: file:/app/apache-nutch-1.9/data/segments/20150630152657/content/parse_text
Input path does not exist: file:/app/apache-nutch-1.9/data/segments/20150630152657/crawl_generate/crawl_fetch
Input path does not exist: file:/app/apache-nutch-1.9/data/segments/20150630152657/crawl_generate/crawl_parse
Input path does not exist: file:/app/apache-nutch-1.9/data/segments/20150630152657/crawl_generate/parse_data
Input path does not exist: file:/app/apache-nutch-1.9/data/segments/20150630152657/crawl_generate/parse_text
Input path does not exist: file:/app/apache-nutch-1.9/data/segments/20150630152657/crawl_parse/crawl_fetch
Input path does not exist: file:/app/apache-nutch-1.9/data/segments/20150630152657/crawl_parse/crawl_parse
Input path does not exist: file:/app/apache-nutch-1.9/data/segments/20150630152657/crawl_parse/parse_data
Input path does not exist: file:/app/apache-nutch-1.9/data/segments/20150630152657/crawl_parse/parse_text
Input path does not exist: file:/app/apache-nutch-1.9/data/segments/20150630152657/parse_text/crawl_fetch
Input path does not exist: file:/app/apache-nutch-1.9/data/segments/20150630152657/parse_text/crawl_parse
Input path does not exist: file:/app/apache-nutch-1.9/data/segments/20150630152657/parse_text/parse_data
Input path does not exist: file:/app/apache-nutch-1.9/data/segments/20150630152657/parse_text/parse_text
Input path does not exist: file:/app/apache-nutch-1.9/data/segments/20150630152657/crawl_fetch/crawl_fetch
Input path does not exist: file:/app/apache-nutch-1.9/data/segments/20150630152657/crawl_fetch/crawl_parse
Input path does not exist: file:/app/apache-nutch-1.9/data/segments/20150630152657/crawl_fetch/parse_data
Input path does not exist: file:/app/apache-nutch-1.9/data/segments/20150630152657/crawl_fetch/parse_text

路径找到不到 。该问题在升级nutch 到1.9的时候出现。
通过定位,要找的该路径不对,file:/app/apache-nutch-1.9/data/segments/20150630152657/crawl_fetch/parse_text ,每个路径后面多了一个层冗余路径。正确的路径应该是:
/app/apache-nutch-1.9/data/segments/20150630152657/crawl_fetch/ 到20150630152657时间下的一层就可以了。那肯定是路径寻找出问题了。

nutch index的时候找segment路径,是IndexingJob的如下代码实现

 for (int i = 1; i < args.length; i++) {
            if (args[i].equals("-linkdb")) {
                linkDb = new Path(args[++i]);
                //这里开始解析传入的参数路径
            } else if (args[i].equals("-dir")) {
                Path dir = new Path(args[++i]);
                FileSystem fs = dir.getFileSystem(getConf());
                //这里列出了所有的子目录,
                FileStatus[] fstats = fs.listStatus(dir,
                        HadoopFSUtil.getPassDirectoriesFilter(fs));
                Path[] files = HadoopFSUtil.getPaths(fstats);
                for (Path p : files) {
                    segments.add(p);
                }
            } else if (args[i].equals("-noCommit")) {
                noCommit = true;
            } else if (args[i].equals("-deleteGone")) {
                deleteGone = true;
            } else if (args[i].equals("-filter")) {
                filter = true;
            } else if (args[i].equals("-normalize")) {
                normalize = true;
            } else if (args[i].equals("-params")) {
                params = args[++i];
            } else {
                segments.add(new Path(args[i]));
            }
        }

实际上对调用程序传入的 segment path参数又找了一次子目录,而crawl脚本程序传入的路径已经到20150630152657时间这一层了,导致路径多了一层。
crawl脚本如下:

  echo "Indexing $SEGMENT on SOLR index -> $SOLRURL"
  "$bin/nutch" index -D solr.server.url=$SOLRURL "$CRAWL_PATH"/crawldb -linkdb "$CRAWL_PATH"/linkdb "$CRAWL_PATH"/segments/$SEGMENT  ---$SEGMENT即是生成的时间路径
   //IndexerMapReduce 初始化job类负责加上最后的 crawl_fetch、crawl_parse 、parse_data、parse_text 路径

  public static void initMRJob(Path crawlDb, Path linkDb,
                           Collection<Path> segments,
                           JobConf job) {

    LOG.info("IndexerMapReduce: crawldb: " + crawlDb);

    if (linkDb!=null)
      LOG.info("IndexerMapReduce: linkdb: " + linkDb);

    for (final Path segment : segments) {
      LOG.info("IndexerMapReduces: adding segment: " + segment);
      //
      FileInputFormat.addInputPath(job, new Path(segment, CrawlDatum.FETCH_DIR_NAME));
      FileInputFormat.addInputPath(job, new Path(segment, CrawlDatum.PARSE_DIR_NAME));
      FileInputFormat.addInputPath(job, new Path(segment, ParseData.DIR_NAME));
      FileInputFormat.addInputPath(job, new Path(segment, ParseText.DIR_NAME));
    }

    FileInputFormat.addInputPath(job, new Path(crawlDb, CrawlDb.CURRENT_NAME));

    if (linkDb!=null)
      FileInputFormat.addInputPath(job, new Path(linkDb, LinkDb.CURRENT_NAME));

    job.setInputFormat(SequenceFileInputFormat.class);

    job.setMapperClass(IndexerMapReduce.class);
    job.setReducerClass(IndexerMapReduce.class);

    job.setOutputFormat(IndexerOutputFormat.class);
    job.setOutputKeyClass(Text.class);
    job.setMapOutputValueClass(NutchWritable.class);
    job.setOutputValueClass(NutchWritable.class);
  }

修改方法:

修改crawl脚本 去掉最后的时间路径 或者修改代码不在找子路径

Python网络爬虫与推荐算法新闻推荐平台:网络爬虫:通过Python实现新浪新闻的爬取,可爬取新闻页面上的标题、文本、图片、视频链接(保留排版) 推荐算法:权重衰减+标签推荐+区域推荐+热点推荐.zip项目工程资源经过严格测试可直接运行成功且功能正常的情况才上传,可轻松复刻,拿到资料包后可轻松复现出一样的项目,本人系统开发经验充足(全领域),有任何使用问题欢迎随与我联系,我会及为您解惑,提供帮助。 【资源内容】:包含完整源码+工程文件+说明(如有)等。答辩评审平均分达到96分,放心下载使用!可轻松复现,设计报告也可借鉴此项目,该资源内项目代码都经过测试运行成功,功能ok的情况下才上传的。 【提供帮助】:有任何使用问题欢迎随与我联系,我会及解答解惑,提供帮助 【附带帮助】:若还需要相关开发工具、学习资料等,我会提供帮助,提供资料,鼓励学习进步 【项目价值】:可用在相关项目设计中,皆可应用在项目、毕业设计、课程设计、期末/期中/大作业、工程实训、大创等学科竞赛比赛、初期项目立项、学习/练手等方面,可借鉴此优质项目实现复刻,设计报告也可借鉴此项目,也可基于此项目来扩展开发出更多功能 下载后请首先打开README文件(如有),项目工程可直接复现复刻,如果基础还行,也可在此程序基础上进行修改,以实现其它功能。供开源学习/技术交流/学习参考,勿用于商业用途。质量优质,放心下载使用。
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值