使用SolrJ生成索引

这个例子使用两种方式来演示如何生成全量索引:
一个是从db中通过sql生成全量索引
一个是通过tika解析文件生成全量索引

  1 package SolrJExample;
  2 
  3 import org.apache.solr.client.solrj.SolrServerException;
  4 import org.apache.solr.client.solrj.impl.StreamingUpdateSolrServer;
  5 import org.apache.solr.client.solrj.impl.XMLResponseParser;
  6 import org.apache.solr.client.solrj.response.UpdateResponse;
  7 import org.apache.solr.common.SolrInputDocument;
  8 import org.apache.tika.metadata.Metadata;
  9 import org.apache.tika.parser.AutoDetectParser;
 10 import org.apache.tika.parser.ParseContext;
 11 import org.apache.tika.sax.BodyContentHandler;
 12 import org.xml.sax.ContentHandler;
 13 
 14 import java.io.File;
 15 import java.io.FileInputStream;
 16 import java.io.IOException;
 17 import java.io.InputStream;
 18 import java.sql.*;
 19 import java.util.ArrayList;
 20 import java.util.Collection;
 21 
 22 /* Example class showing the skeleton of using Tika and
 23    Sql on the client to index documents from
 24    both structured documents and a SQL database.
 25 
 26    NOTE: The SQL example and the Tika example are entirely orthogonal.
 27    Both are included here to make a
 28    more interesting example, but you can omit either of them.
 29 
 30  */
 31 public class SqlTikaExample {
 32   private StreamingUpdateSolrServer _server;
 33   private long _start = System.currentTimeMillis();
 34   private AutoDetectParser _autoParser;
 35   private int _totalTika = 0;
 36   private int _totalSql = 0;
 37 
 38   private Collection _docs = new ArrayList();
 39 
 40   public static void main(String[] args) {
 41     try {
 42       SqlTikaExample idxer = new SqlTikaExample("http://localhost:8983/solr");
 43 
 44       idxer.doTikaDocuments(new File("/Users/Erick/testdocs"));
 45       idxer.doSqlDocuments();
 46 
 47       idxer.endIndexing();
 48     } catch (Exception e) {
 49       e.printStackTrace();
 50     }
 51   }
 52 
 53   private SqlTikaExample(String url) throws IOException, SolrServerException {
 54       // Create a multi-threaded communications channel to the Solr server.
 55       // Could be CommonsHttpSolrServer as well.
 56       //
 57     _server = new StreamingUpdateSolrServer(url, 10, 4);
 58 
 59     _server.setSoTimeout(1000);  // socket read timeout
 60     _server.setConnectionTimeout(1000);
 61     _server.setMaxRetries(1); // defaults to 0.  > 1 not recommended.
 62          // binary parser is used by default for responses
 63     _server.setParser(new XMLResponseParser()); 
 64 
 65       // One of the ways Tika can be used to attempt to parse arbitrary files.
 66     _autoParser = new AutoDetectParser();
 67   }
 68 
 69     // Just a convenient place to wrap things up.
 70   private void endIndexing() throws IOException, SolrServerException {
 71     if (_docs.size() > 0) { // Are there any documents left over?
 72       _server.add(_docs, 300000); // Commit within 5 minutes
 73     }
 74     _server.commit(); // Only needs to be done at the end,
 75                       // commitWithin should do the rest.
 76                       // Could even be omitted
 77                       // assuming commitWithin was specified.
 78     long endTime = System.currentTimeMillis();
 79     log("Total Time Taken: " + (endTime - _start) +
 80          " milliseconds to index " + _totalSql +
 81         " SQL rows and " + _totalTika + " documents");
 82   }
 83 
 84   // I hate writing System.out.println() everyplace,
 85   // besides this gives a central place to convert to true logging
 86   // in a production system.
 87   private static void log(String msg) {
 88     System.out.println(msg);
 89   }
 90 
 91   /**
 92    * ***************************Tika processing here
 93    */
 94   // Recursively traverse the filesystem, parsing everything found.
 95   private void doTikaDocuments(File root) throws IOException, SolrServerException {
 96 
 97     // Simple loop for recursively indexing all the files
 98     // in the root directory passed in.
 99     for (File file : root.listFiles()) {
100       if (file.isDirectory()) {
101         doTikaDocuments(file);
102         continue;
103       }
104         // Get ready to parse the file.
105       ContentHandler textHandler = new BodyContentHandler();
106       Metadata metadata = new Metadata();
107       ParseContext context = new ParseContext();
108 
109       InputStream input = new FileInputStream(file);
110 
111         // Try parsing the file. Note we haven't checked at all to
112         // see whether this file is a good candidate.
113       try {
114         _autoParser.parse(input, textHandler, metadata, context);
115       } catch (Exception e) {
116           // Needs better logging of what went wrong in order to
117           // track down "bad" documents.
118         log(String.format("File %s failed", file.getCanonicalPath()));
119         e.printStackTrace();
120         continue;
121       }
122       // Just to show how much meta-data and what form it's in.
123       dumpMetadata(file.getCanonicalPath(), metadata);
124 
125       // Index just a couple of the meta-data fields.
126       SolrInputDocument doc = new SolrInputDocument();
127 
128       doc.addField("id", file.getCanonicalPath());
129 
130       // Crude way to get known meta-data fields.
131       // Also possible to write a simple loop to examine all the
132       // metadata returned and selectively index it and/or
133       // just get a list of them.
134       // One can also use the LucidWorks field mapping to
135       // accomplish much the same thing.
136       String author = metadata.get("Author");
137 
138       if (author != null) {
139         doc.addField("author", author);
140       }
141 
142       doc.addField("text", textHandler.toString());
143 
144       _docs.add(doc);
145       ++_totalTika;
146 
147       // Completely arbitrary, just batch up more than one document
148       // for throughput!
149       if (_docs.size() >= 1000) {
150           // Commit within 5 minutes.
151         UpdateResponse resp = _server.add(_docs, 300000);
152         if (resp.getStatus() != 0) {
153           log("Some horrible error has occurred, status is: " +
154                   resp.getStatus());
155         }
156         _docs.clear();
157       }
158     }
159   }
160 
161     // Just to show all the metadata that's available.
162   private void dumpMetadata(String fileName, Metadata metadata) {
163     log("Dumping metadata for file: " + fileName);
164     for (String name : metadata.names()) {
165       log(name + ":" + metadata.get(name));
166     }
167     log("\n\n");
168   }
169 
170   /**
171    * ***************************SQL processing here
172    */
173   private void doSqlDocuments() throws SQLException {
174     Connection con = null;
175     try {
176       Class.forName("com.mysql.jdbc.Driver").newInstance();
177       log("Driver Loaded");
178 
179       con = DriverManager.getConnection("jdbc:mysql://192.168.1.103:3306/test?"
180                 + "user=testuser&password=test123");
181 
182       Statement st = con.createStatement();
183       ResultSet rs = st.executeQuery("select id,title,text from test");
184 
185       while (rs.next()) {
186         // DO NOT move this outside the while loop
187         // or be sure to call doc.clear()
188         SolrInputDocument doc = new SolrInputDocument(); 
189         String id = rs.getString("id");
190         String title = rs.getString("title");
191         String text = rs.getString("text");
192 
193         doc.addField("id", id);
194         doc.addField("title", title);
195         doc.addField("text", text);
196 
197         _docs.add(doc);
198         ++_totalSql;
199 
200         // Completely arbitrary, just batch up more than one
201         // document for throughput!
202         if (_docs.size() > 1000) {
203              // Commit within 5 minutes.
204           UpdateResponse resp = _server.add(_docs, 300000);
205           if (resp.getStatus() != 0) {
206             log("Some horrible error has occurred, status is: " +
207                   resp.getStatus());
208           }
209           _docs.clear();
210         }
211       }
212     } catch (Exception ex) {
213       ex.printStackTrace();
214     } finally {
215       if (con != null) {
216         con.close();
217       }
218     }
219   }
220 }

 

转载于:https://www.cnblogs.com/SuperBing/archive/2013/01/30/2882820.html

评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值