package com.gitblit.utils;
|
|
import java.io.ByteArrayOutputStream;
|
import java.io.File;
|
import java.io.IOException;
|
import java.io.InputStream;
|
import java.text.ParseException;
|
import java.util.ArrayList;
|
import java.util.Arrays;
|
import java.util.HashMap;
|
import java.util.LinkedHashSet;
|
import java.util.List;
|
import java.util.Map;
|
import java.util.Set;
|
import java.util.TreeSet;
|
import java.util.concurrent.ConcurrentHashMap;
|
|
import org.apache.lucene.analysis.standard.StandardAnalyzer;
|
import org.apache.lucene.document.DateTools;
|
import org.apache.lucene.document.DateTools.Resolution;
|
import org.apache.lucene.document.Document;
|
import org.apache.lucene.document.Field;
|
import org.apache.lucene.document.Field.Index;
|
import org.apache.lucene.document.Field.Store;
|
import org.apache.lucene.index.IndexReader;
|
import org.apache.lucene.index.IndexWriter;
|
import org.apache.lucene.index.IndexWriterConfig;
|
import org.apache.lucene.index.IndexWriterConfig.OpenMode;
|
import org.apache.lucene.index.MultiReader;
|
import org.apache.lucene.index.Term;
|
import org.apache.lucene.queryParser.QueryParser;
|
import org.apache.lucene.search.BooleanClause.Occur;
|
import org.apache.lucene.search.BooleanQuery;
|
import org.apache.lucene.search.IndexSearcher;
|
import org.apache.lucene.search.Query;
|
import org.apache.lucene.search.ScoreDoc;
|
import org.apache.lucene.search.TopScoreDocCollector;
|
import org.apache.lucene.store.Directory;
|
import org.apache.lucene.store.FSDirectory;
|
import org.apache.lucene.util.Version;
|
import org.eclipse.jgit.diff.DiffEntry.ChangeType;
|
import org.eclipse.jgit.lib.Constants;
|
import org.eclipse.jgit.lib.FileMode;
|
import org.eclipse.jgit.lib.ObjectId;
|
import org.eclipse.jgit.lib.ObjectLoader;
|
import org.eclipse.jgit.lib.Repository;
|
import org.eclipse.jgit.revwalk.RevCommit;
|
import org.eclipse.jgit.revwalk.RevObject;
|
import org.eclipse.jgit.revwalk.RevWalk;
|
import org.eclipse.jgit.treewalk.TreeWalk;
|
|
import com.gitblit.models.IssueModel;
|
import com.gitblit.models.IssueModel.Attachment;
|
import com.gitblit.models.PathModel.PathChangeModel;
|
import com.gitblit.models.RefModel;
|
import com.gitblit.models.SearchResult;
|
|
/**
|
* A collection of utility methods for indexing and querying a Lucene repository
|
* index.
|
*
|
* @author James Moger
|
*
|
*/
|
public class LuceneUtils {
|
|
/**
|
* The types of objects that can be indexed and queried.
|
*/
|
public static enum ObjectType {
|
commit, blob, issue;
|
|
static ObjectType fromName(String name) {
|
for (ObjectType value : values()) {
|
if (value.name().equals(name)) {
|
return value;
|
}
|
}
|
return null;
|
}
|
}
|
|
private static final Version LUCENE_VERSION = Version.LUCENE_35;
|
|
private static final String FIELD_OBJECT_TYPE = "type";
|
private static final String FIELD_OBJECT_ID = "id";
|
private static final String FIELD_BRANCH = "branch";
|
private static final String FIELD_REPOSITORY = "repository";
|
private static final String FIELD_SUMMARY = "summary";
|
private static final String FIELD_CONTENT = "content";
|
private static final String FIELD_AUTHOR = "author";
|
private static final String FIELD_COMMITTER = "committer";
|
private static final String FIELD_DATE = "date";
|
private static final String FIELD_LABEL = "label";
|
private static final String FIELD_ATTACHMENT = "attachment";
|
|
private static Set<String> excludedExtensions = new TreeSet<String>(
|
Arrays.asList("7z", "arc", "arj", "bin", "bmp", "dll", "doc",
|
"docx", "exe", "gif", "gz", "jar", "jpg", "lib", "lzh",
|
"odg", "pdf", "ppt", "png", "so", "swf", "xcf", "xls",
|
"xlsx", "zip"));
|
|
private static Set<String> excludedBranches = new TreeSet<String>(
|
Arrays.asList("/refs/heads/gb-issues"));
|
|
private static final Map<File, IndexSearcher> SEARCHERS = new ConcurrentHashMap<File, IndexSearcher>();
|
private static final Map<File, IndexWriter> WRITERS = new ConcurrentHashMap<File, IndexWriter>();
|
|
/**
|
* Returns the name of the repository.
|
*
|
* @param repository
|
* @return the repository name
|
*/
|
private static String getName(Repository repository) {
|
if (repository.isBare()) {
|
return repository.getDirectory().getName();
|
} else {
|
return repository.getDirectory().getParentFile().getName();
|
}
|
}
|
|
/**
|
* Deletes the Lucene index for the specified repository.
|
*
|
* @param repository
|
* @return true, if successful
|
*/
|
public static boolean deleteIndex(Repository repository) {
|
try {
|
File luceneIndex = new File(repository.getDirectory(), "lucene");
|
if (luceneIndex.exists()) {
|
org.eclipse.jgit.util.FileUtils.delete(luceneIndex,
|
org.eclipse.jgit.util.FileUtils.RECURSIVE);
|
}
|
return true;
|
} catch (IOException e) {
|
throw new RuntimeException(e);
|
}
|
}
|
|
/**
|
* This completely indexes the repository and will destroy any existing
|
* index.
|
*
|
* @param repository
|
* @return true if the indexing has succeeded
|
*/
|
public static boolean index(Repository repository) {
|
try {
|
String repositoryName = getName(repository);
|
Set<String> indexedCommits = new TreeSet<String>();
|
IndexWriter writer = getIndexWriter(repository, true);
|
// build a quick lookup of tags
|
Map<String, List<String>> tags = new HashMap<String, List<String>>();
|
for (RefModel tag : JGitUtils.getTags(repository, false, -1)) {
|
if (!tags.containsKey(tag.getObjectId())) {
|
tags.put(tag.getReferencedObjectId().getName(), new ArrayList<String>());
|
}
|
tags.get(tag.getReferencedObjectId().getName()).add(tag.displayName);
|
}
|
|
// walk through each branch
|
List<RefModel> branches = JGitUtils.getLocalBranches(repository, true, -1);
|
for (RefModel branch : branches) {
|
if (excludedBranches.contains(branch.getName())) {
|
continue;
|
}
|
String branchName = branch.getName();
|
RevWalk revWalk = new RevWalk(repository);
|
RevCommit rev = revWalk.parseCommit(branch.getObjectId());
|
|
// index the blob contents of the tree
|
ByteArrayOutputStream os = new ByteArrayOutputStream();
|
byte[] tmp = new byte[32767];
|
TreeWalk treeWalk = new TreeWalk(repository);
|
treeWalk.addTree(rev.getTree());
|
treeWalk.setRecursive(true);
|
String revDate = DateTools.timeToString(rev.getCommitTime() * 1000L,
|
Resolution.MINUTE);
|
while (treeWalk.next()) {
|
Document doc = new Document();
|
doc.add(new Field(FIELD_OBJECT_TYPE, ObjectType.blob.name(), Store.YES,
|
Index.NOT_ANALYZED_NO_NORMS));
|
doc.add(new Field(FIELD_REPOSITORY, repositoryName, Store.YES,
|
Index.NOT_ANALYZED));
|
doc.add(new Field(FIELD_BRANCH, branchName, Store.YES,
|
Index.NOT_ANALYZED));
|
doc.add(new Field(FIELD_OBJECT_ID, treeWalk.getPathString(), Store.YES,
|
Index.NOT_ANALYZED));
|
doc.add(new Field(FIELD_DATE, revDate, Store.YES, Index.NO));
|
doc.add(new Field(FIELD_AUTHOR, rev.getAuthorIdent().getName(), Store.YES,
|
Index.NOT_ANALYZED_NO_NORMS));
|
doc.add(new Field(FIELD_COMMITTER, rev.getCommitterIdent().getName(),
|
Store.YES, Index.NOT_ANALYZED_NO_NORMS));
|
doc.add(new Field(FIELD_LABEL, branch.getName(), Store.YES, Index.ANALYZED));
|
|
// determine extension to compare to the extension
|
// blacklist
|
String ext = null;
|
String name = treeWalk.getPathString().toLowerCase();
|
if (name.indexOf('.') > -1) {
|
ext = name.substring(name.lastIndexOf('.') + 1);
|
}
|
|
if (StringUtils.isEmpty(ext) || !excludedExtensions.contains(ext)) {
|
// read the blob content
|
ObjectId entid = treeWalk.getObjectId(0);
|
FileMode entmode = treeWalk.getFileMode(0);
|
RevObject ro = revWalk.lookupAny(entid, entmode.getObjectType());
|
revWalk.parseBody(ro);
|
ObjectLoader ldr = repository.open(ro.getId(), Constants.OBJ_BLOB);
|
InputStream in = ldr.openStream();
|
os.reset();
|
int n = 0;
|
while ((n = in.read(tmp)) > 0) {
|
os.write(tmp, 0, n);
|
}
|
in.close();
|
byte[] content = os.toByteArray();
|
String str = new String(content, "UTF-8");
|
doc.add(new Field(FIELD_CONTENT, str, Store.NO, Index.ANALYZED));
|
writer.addDocument(doc);
|
}
|
}
|
|
os.close();
|
treeWalk.release();
|
|
// index the head commit object
|
String head = rev.getId().getName();
|
if (indexedCommits.add(head)) {
|
Document doc = createDocument(rev, tags.get(head));
|
doc.add(new Field(FIELD_REPOSITORY, repositoryName, Store.YES,
|
Index.NOT_ANALYZED));
|
doc.add(new Field(FIELD_BRANCH, branchName, Store.YES,
|
Index.NOT_ANALYZED));
|
writer.addDocument(doc);
|
}
|
|
// traverse the log and index the previous commit objects
|
revWalk.markStart(rev);
|
while ((rev = revWalk.next()) != null) {
|
String hash = rev.getId().getName();
|
if (indexedCommits.add(hash)) {
|
Document doc = createDocument(rev, tags.get(hash));
|
doc.add(new Field(FIELD_REPOSITORY, repositoryName, Store.YES,
|
Index.NOT_ANALYZED));
|
doc.add(new Field(FIELD_BRANCH, branchName, Store.YES,
|
Index.NOT_ANALYZED));
|
writer.addDocument(doc);
|
}
|
}
|
|
// finished
|
revWalk.dispose();
|
}
|
|
// this repository has a gb-issues branch, index all issues
|
if (IssueUtils.getIssuesBranch(repository) != null) {
|
List<IssueModel> issues = IssueUtils.getIssues(repository, null);
|
for (IssueModel issue : issues) {
|
Document doc = createDocument(issue);
|
doc.add(new Field(FIELD_REPOSITORY, repositoryName, Store.YES,
|
Index.NOT_ANALYZED));
|
writer.addDocument(doc);
|
}
|
}
|
|
// commit all changes and reset the searcher
|
resetIndexSearcher(repository);
|
writer.commit();
|
return true;
|
} catch (Exception e) {
|
e.printStackTrace();
|
}
|
return false;
|
}
|
|
/**
|
* Incrementally update the index with the specified commit for the
|
* repository.
|
*
|
* @param repository
|
* @param branch
|
* the fully qualified branch name (e.g. refs/heads/master)
|
* @param commit
|
* @return true, if successful
|
*/
|
public static boolean index(Repository repository, String branch, RevCommit commit) {
|
try {
|
if (excludedBranches.contains(branch)) {
|
if (IssueUtils.GB_ISSUES.equals(branch)) {
|
// index an issue
|
String issueId = commit.getShortMessage().substring(2).trim();
|
IssueModel issue = IssueUtils.getIssue(repository, issueId);
|
return index(repository, issue, true);
|
}
|
return false;
|
}
|
List<PathChangeModel> changedPaths = JGitUtils.getFilesInCommit(repository, commit);
|
String repositoryName = getName(repository);
|
String revDate = DateTools.timeToString(commit.getCommitTime() * 1000L,
|
Resolution.MINUTE);
|
IndexWriter writer = getIndexWriter(repository, false);
|
for (PathChangeModel path : changedPaths) {
|
// delete the indexed blob
|
writer.deleteDocuments(new Term(FIELD_OBJECT_TYPE, ObjectType.blob.name()),
|
new Term(FIELD_BRANCH, branch),
|
new Term(FIELD_OBJECT_ID, path.path));
|
|
// re-index the blob
|
if (!ChangeType.DELETE.equals(path.changeType)) {
|
Document doc = new Document();
|
doc.add(new Field(FIELD_OBJECT_TYPE, ObjectType.blob.name(), Store.YES,
|
Index.NOT_ANALYZED_NO_NORMS));
|
doc.add(new Field(FIELD_REPOSITORY, repositoryName, Store.YES,
|
Index.NOT_ANALYZED));
|
doc.add(new Field(FIELD_BRANCH, branch, Store.YES, Index.NOT_ANALYZED));
|
doc.add(new Field(FIELD_OBJECT_ID, path.path, Store.YES,
|
Index.NOT_ANALYZED));
|
doc.add(new Field(FIELD_DATE, revDate, Store.YES, Index.NO));
|
doc.add(new Field(FIELD_AUTHOR, commit.getAuthorIdent().getName(), Store.YES,
|
Index.NOT_ANALYZED_NO_NORMS));
|
doc.add(new Field(FIELD_COMMITTER, commit.getCommitterIdent().getName(),
|
Store.YES, Index.NOT_ANALYZED_NO_NORMS));
|
doc.add(new Field(FIELD_LABEL, branch, Store.YES, Index.ANALYZED));
|
|
// determine extension to compare to the extension
|
// blacklist
|
String ext = null;
|
String name = path.name.toLowerCase();
|
if (name.indexOf('.') > -1) {
|
ext = name.substring(name.lastIndexOf('.') + 1);
|
}
|
|
if (StringUtils.isEmpty(ext) || !excludedExtensions.contains(ext)) {
|
// read the blob content
|
String str = JGitUtils.getStringContent(repository,
|
commit.getTree(), path.path);
|
doc.add(new Field(FIELD_CONTENT, str, Store.NO, Index.ANALYZED));
|
writer.addDocument(doc);
|
}
|
}
|
}
|
writer.commit();
|
|
Document doc = createDocument(commit, null);
|
return index(repository, doc);
|
} catch (Exception e) {
|
e.printStackTrace();
|
}
|
return false;
|
}
|
|
/**
|
* Incrementally update the index with the specified issue for the
|
* repository.
|
*
|
* @param repository
|
* @param issue
|
* @param reindex
|
* if true, the old index entry for this issue will be deleted.
|
* This is only appropriate for pre-existing/indexed issues.
|
* @return true, if successful
|
*/
|
public static boolean index(Repository repository, IssueModel issue, boolean reindex) {
|
try {
|
Document doc = createDocument(issue);
|
if (reindex) {
|
// delete the old issue from the index, if exists
|
IndexWriter writer = getIndexWriter(repository, false);
|
writer.deleteDocuments(new Term(FIELD_OBJECT_TYPE, ObjectType.issue.name()),
|
new Term(FIELD_OBJECT_ID, String.valueOf(issue.id)));
|
writer.commit();
|
}
|
return index(repository, doc);
|
} catch (Exception e) {
|
e.printStackTrace();
|
}
|
return false;
|
}
|
|
/**
|
* Creates a Lucene document from an issue.
|
*
|
* @param issue
|
* @return a Lucene document
|
*/
|
private static Document createDocument(IssueModel issue) {
|
Document doc = new Document();
|
doc.add(new Field(FIELD_OBJECT_TYPE, ObjectType.issue.name(), Store.YES,
|
Field.Index.NOT_ANALYZED_NO_NORMS));
|
doc.add(new Field(FIELD_OBJECT_ID, issue.id, Store.YES, Index.NOT_ANALYZED));
|
doc.add(new Field(FIELD_BRANCH, IssueUtils.GB_ISSUES, Store.YES, Index.NOT_ANALYZED));
|
doc.add(new Field(FIELD_DATE, DateTools.dateToString(issue.created, Resolution.MINUTE),
|
Store.YES, Field.Index.NO));
|
doc.add(new Field(FIELD_AUTHOR, issue.reporter, Store.YES, Index.NOT_ANALYZED_NO_NORMS));
|
List<String> attachments = new ArrayList<String>();
|
for (Attachment attachment : issue.getAttachments()) {
|
attachments.add(attachment.name.toLowerCase());
|
}
|
doc.add(new Field(FIELD_ATTACHMENT, StringUtils.flattenStrings(attachments), Store.YES,
|
Index.ANALYZED));
|
doc.add(new Field(FIELD_SUMMARY, issue.summary, Store.YES, Index.ANALYZED));
|
doc.add(new Field(FIELD_CONTENT, issue.toString(), Store.NO, Index.ANALYZED));
|
doc.add(new Field(FIELD_LABEL, StringUtils.flattenStrings(issue.getLabels()), Store.YES,
|
Index.ANALYZED));
|
return doc;
|
}
|
|
/**
|
* Creates a Lucene document for a commit
|
*
|
* @param commit
|
* @param tags
|
* @return a Lucene document
|
*/
|
private static Document createDocument(RevCommit commit, List<String> tags) {
|
Document doc = new Document();
|
doc.add(new Field(FIELD_OBJECT_TYPE, ObjectType.commit.name(), Store.YES,
|
Index.NOT_ANALYZED_NO_NORMS));
|
doc.add(new Field(FIELD_OBJECT_ID, commit.getName(), Store.YES, Index.NOT_ANALYZED));
|
doc.add(new Field(FIELD_DATE, DateTools.timeToString(commit.getCommitTime() * 1000L,
|
Resolution.MINUTE), Store.YES, Index.NO));
|
doc.add(new Field(FIELD_AUTHOR, commit.getCommitterIdent().getName(), Store.YES,
|
Index.NOT_ANALYZED_NO_NORMS));
|
doc.add(new Field(FIELD_SUMMARY, commit.getShortMessage(), Store.YES, Index.ANALYZED));
|
doc.add(new Field(FIELD_CONTENT, commit.getFullMessage(), Store.NO, Index.ANALYZED));
|
if (!ArrayUtils.isEmpty(tags)) {
|
if (!ArrayUtils.isEmpty(tags)) {
|
doc.add(new Field(FIELD_LABEL, StringUtils.flattenStrings(tags), Store.YES,
|
Index.ANALYZED));
|
}
|
}
|
return doc;
|
}
|
|
/**
|
* Incrementally index an object for the repository.
|
*
|
* @param repository
|
* @param doc
|
* @return true, if successful
|
*/
|
private static boolean index(Repository repository, Document doc) {
|
try {
|
String repositoryName = getName(repository);
|
doc.add(new Field(FIELD_REPOSITORY, repositoryName, Store.YES,
|
Index.NOT_ANALYZED));
|
IndexWriter writer = getIndexWriter(repository, false);
|
writer.addDocument(doc);
|
resetIndexSearcher(repository);
|
writer.commit();
|
return true;
|
} catch (Exception e) {
|
e.printStackTrace();
|
}
|
return false;
|
}
|
|
private static SearchResult createSearchResult(Document doc, float score) throws ParseException {
|
SearchResult result = new SearchResult();
|
result.score = score;
|
result.date = DateTools.stringToDate(doc.get(FIELD_DATE));
|
result.summary = doc.get(FIELD_SUMMARY);
|
result.author = doc.get(FIELD_AUTHOR);
|
result.committer = doc.get(FIELD_COMMITTER);
|
result.type = ObjectType.fromName(doc.get(FIELD_OBJECT_TYPE));
|
result.repository = doc.get(FIELD_REPOSITORY);
|
result.branch = doc.get(FIELD_BRANCH);
|
result.id = doc.get(FIELD_OBJECT_ID);
|
if (doc.get(FIELD_LABEL) != null) {
|
result.labels = StringUtils.getStringsFromValue(doc.get(FIELD_LABEL));
|
}
|
return result;
|
}
|
|
private static void resetIndexSearcher(Repository repository) throws IOException {
|
IndexSearcher searcher = SEARCHERS.get(repository.getDirectory());
|
if (searcher != null) {
|
SEARCHERS.remove(repository.getDirectory());
|
searcher.close();
|
}
|
}
|
|
/**
|
* Gets an index searcher for the repository.
|
*
|
* @param repository
|
* @return
|
* @throws IOException
|
*/
|
private static IndexSearcher getIndexSearcher(Repository repository) throws IOException {
|
IndexSearcher searcher = SEARCHERS.get(repository.getDirectory());
|
if (searcher == null) {
|
IndexWriter writer = getIndexWriter(repository, false);
|
searcher = new IndexSearcher(IndexReader.open(writer, true));
|
SEARCHERS.put(repository.getDirectory(), searcher);
|
}
|
return searcher;
|
}
|
|
/**
|
* Gets an index writer for the repository. The index will be created if it
|
* does not already exist or if forceCreate is specified.
|
*
|
* @param repository
|
* @param forceCreate
|
* @return an IndexWriter
|
* @throws IOException
|
*/
|
private static IndexWriter getIndexWriter(Repository repository, boolean forceCreate)
|
throws IOException {
|
IndexWriter indexWriter = WRITERS.get(repository.getDirectory());
|
File indexFolder = new File(repository.getDirectory(), "lucene");
|
Directory directory = FSDirectory.open(indexFolder);
|
if (forceCreate || !indexFolder.exists()) {
|
// if the writer is going to blow away the existing index and create
|
// a new one then it should not be cached. instead, close any open
|
// writer, create a new one, and return.
|
if (indexWriter != null) {
|
indexWriter.close();
|
indexWriter = null;
|
WRITERS.remove(repository.getDirectory());
|
}
|
indexFolder.mkdirs();
|
IndexWriterConfig config = new IndexWriterConfig(LUCENE_VERSION, new StandardAnalyzer(
|
LUCENE_VERSION));
|
config.setOpenMode(OpenMode.CREATE);
|
IndexWriter writer = new IndexWriter(directory, config);
|
writer.close();
|
}
|
|
if (indexWriter == null) {
|
IndexWriterConfig config = new IndexWriterConfig(LUCENE_VERSION, new StandardAnalyzer(
|
LUCENE_VERSION));
|
config.setOpenMode(OpenMode.APPEND);
|
indexWriter = new IndexWriter(directory, config);
|
WRITERS.put(repository.getDirectory(), indexWriter);
|
}
|
return indexWriter;
|
}
|
|
/**
|
* Searches the specified repositories for the given text or query
|
*
|
* @param text
|
* if the text is null or empty, null is returned
|
* @param maximumHits
|
* the maximum number of hits to collect
|
* @param repositories
|
* a list of repositories to search. if no repositories are
|
* specified null is returned.
|
* @return a list of SearchResults in order from highest to the lowest score
|
*
|
*/
|
public static List<SearchResult> search(String text, int maximumHits,
|
Repository... repositories) {
|
if (StringUtils.isEmpty(text)) {
|
return null;
|
}
|
if (repositories.length == 0) {
|
return null;
|
}
|
Set<SearchResult> results = new LinkedHashSet<SearchResult>();
|
StandardAnalyzer analyzer = new StandardAnalyzer(LUCENE_VERSION);
|
try {
|
// default search checks summary and content
|
BooleanQuery query = new BooleanQuery();
|
QueryParser qp;
|
qp = new QueryParser(LUCENE_VERSION, FIELD_SUMMARY, analyzer);
|
qp.setAllowLeadingWildcard(true);
|
query.add(qp.parse(text), Occur.SHOULD);
|
|
qp = new QueryParser(LUCENE_VERSION, FIELD_CONTENT, analyzer);
|
qp.setAllowLeadingWildcard(true);
|
query.add(qp.parse(text), Occur.SHOULD);
|
|
IndexSearcher searcher;
|
if (repositories.length == 1) {
|
// single repository search
|
searcher = getIndexSearcher(repositories[0]);
|
} else {
|
// multiple repository search
|
List<IndexReader> readers = new ArrayList<IndexReader>();
|
for (Repository repository : repositories) {
|
IndexSearcher repositoryIndex = getIndexSearcher(repository);
|
readers.add(repositoryIndex.getIndexReader());
|
}
|
IndexReader [] rdrs = readers.toArray(new IndexReader[readers.size()]);
|
MultiReader reader = new MultiReader(rdrs);
|
searcher = new IndexSearcher(reader);
|
}
|
Query rewrittenQuery = searcher.rewrite(query);
|
TopScoreDocCollector collector = TopScoreDocCollector.create(maximumHits, true);
|
searcher.search(rewrittenQuery, collector);
|
ScoreDoc[] hits = collector.topDocs().scoreDocs;
|
for (int i = 0; i < hits.length; i++) {
|
int docId = hits[i].doc;
|
Document doc = searcher.doc(docId);
|
SearchResult result = createSearchResult(doc, hits[i].score);
|
results.add(result);
|
}
|
} catch (Exception e) {
|
e.printStackTrace();
|
}
|
return new ArrayList<SearchResult>(results);
|
}
|
|
/**
|
* Close all the index writers and searchers
|
*/
|
public static void close() {
|
// close writers
|
for (File file : WRITERS.keySet()) {
|
try {
|
WRITERS.get(file).close(true);
|
} catch (Throwable t) {
|
t.printStackTrace();
|
}
|
}
|
WRITERS.clear();
|
|
// close searchers
|
for (File file : SEARCHERS.keySet()) {
|
try {
|
SEARCHERS.get(file).close();
|
} catch (Throwable t) {
|
t.printStackTrace();
|
}
|
}
|
SEARCHERS.clear();
|
}
|
}
|