热门标签 | HotTags
当前位置:  开发笔记 > 编程语言 > 正文

org.apache.lucene.index.LogMergePolicy.setMergeFactor()方法的使用及代码示例

本文整理了Java中org.apache.lucene.index.LogMergePolicy.setMergeFactor()方法的一些代码示例,展示了

本文整理了Java中org.apache.lucene.index.LogMergePolicy.setMergeFactor()方法的一些代码示例,展示了LogMergePolicy.setMergeFactor()的具体用法。这些代码示例主要来源于Github/Stackoverflow/Maven等平台,是从一些精选项目中提取出来的代码,具有较强的参考意义,能在一定程度帮忙到你。LogMergePolicy.setMergeFactor()方法的具体详情如下:
包路径:org.apache.lucene.index.LogMergePolicy
类名称:LogMergePolicy
方法名:setMergeFactor

LogMergePolicy.setMergeFactor介绍

[英]Determines how often segment indices are merged by addDocument(). With smaller values, less RAM is used while indexing, and searches are faster, but indexing speed is slower. With larger values, more RAM is used during indexing, and while searches is slower, indexing is faster. Thus larger values ( > 10) are best for batch index creation, and smaller values ( <10) for indices that are interactively maintained.
[中]确定addDocument()合并段索引的频率。值越小,索引时使用的内存越少,搜索速度越快,但索引速度越慢。值越大,索引过程中使用的RAM越多,搜索越慢,索引速度越快。因此,较大的值(>10)最好用于批量索引创建,较小的值(<10)最好用于交互式维护的索引。

代码示例

代码示例来源:origin: com.strapdata.elasticsearch.test/framework

public static MergePolicy newLogMergePolicy(int mergeFactor) {
LogMergePolicy logmp = newLogMergePolicy();
logmp.setMergeFactor(mergeFactor);
return logmp;
}

代码示例来源:origin: org.apache.lucene/lucene-core-jfrog

/** Determines how often segment indices are merged by addDocument(). With
* smaller values, less RAM is used while indexing, and searches on
* unoptimized indices are faster, but indexing speed is slower. With larger
* values, more RAM is used during indexing, and while searches on unoptimized
* indices are slower, indexing is faster. Thus larger values (> 10) are best
* for batch index creation, and smaller values (<10) for indices that are
* interactively maintained.
*
*

Note that this method is a convenience method: it
* just calls mergePolicy.setMergeFactor as long as
* mergePolicy is an instance of {@link LogMergePolicy}.
* Otherwise an IllegalArgumentException is thrown.


*
*

This must never be less than 2. The default value is 10.
*/
public void setMergeFactor(int mergeFactor) {
getLogMergePolicy().setMergeFactor(mergeFactor);
}

代码示例来源:origin: org.apache.lucene/com.springsource.org.apache.lucene

/** Determines how often segment indices are merged by addDocument(). With
* smaller values, less RAM is used while indexing, and searches on
* unoptimized indices are faster, but indexing speed is slower. With larger
* values, more RAM is used during indexing, and while searches on unoptimized
* indices are slower, indexing is faster. Thus larger values (> 10) are best
* for batch index creation, and smaller values (<10) for indices that are
* interactively maintained.
*
*

Note that this method is a convenience method: it
* just calls mergePolicy.setMergeFactor as long as
* mergePolicy is an instance of {@link LogMergePolicy}.
* Otherwise an IllegalArgumentException is thrown.


*
*

This must never be less than 2. The default value is 10.
*/
public void setMergeFactor(int mergeFactor) {
getLogMergePolicy().setMergeFactor(mergeFactor);
}

代码示例来源:origin: com.strapdata.elasticsearch.test/framework

public static MergePolicy newLogMergePolicy(boolean useCFS, int mergeFactor) {
LogMergePolicy logmp = newLogMergePolicy();
logmp.setNoCFSRatio(useCFS ? 1.0 : 0.0);
logmp.setMergeFactor(mergeFactor);
return logmp;
}

代码示例来源:origin: org.apache.oodt/cas-filemgr

public void doOptimize() {
IndexWriter writer = null;
boolean createIndex = false;
try {
writer = new IndexWriter(reader.directory(), config);
LogMergePolicy lmp =new LogDocMergePolicy();
lmp.setMergeFactor(this.mergeFactor);
config.setMergePolicy(lmp);
long timeBefore = System.currentTimeMillis();
//TODO http://blog.trifork.com/2011/11/21/simon-says-optimize-is-bad-for-you/
//writer.optimize();
long timeAfter = System.currentTimeMillis();
double numSecOnds= ((timeAfter - timeBefore) * 1.0) / DOUBLE;
LOG.log(Level.INFO, "LuceneCatalog: [" + this.catalogPath
+ "] optimized: took: [" + numSeconds + "] seconds");
} catch (IOException e) {
LOG.log(Level.WARNING, "Unable to optimize lucene index: ["
+ catalogPath + "]: Message: " + e.getMessage());
} finally {
try {
writer.close();
} catch (Exception ignore) {
}
}
}

代码示例来源:origin: psidev.psi.mi/psimitab-search

public void index(Directory directory, InputStream is, boolean createIndex, boolean hasHeaderLine) throws IOException, ConverterException, MitabLineException {
IndexWriterConfig writerCOnfig= new IndexWriterConfig(Version.LUCENE_30, new StandardAnalyzer(Version.LUCENE_30));
LogMergePolicy policy = new LogDocMergePolicy();
policy.setMergeFactor(MERGE_FACTOR);
policy.setMaxMergeDocs(Integer.MAX_VALUE);
writerConfig.setMergePolicy(policy);
IndexWriter indexWriter = new IndexWriter(directory, writerConfig);
if (createIndex){
indexWriter.commit();
indexWriter.deleteAll();
indexWriter.commit();
}
index(indexWriter, is, hasHeaderLine);
indexWriter.close();
}

代码示例来源:origin: apache/oodt

public void doOptimize() {
IndexWriter writer = null;
boolean createIndex = false;
try {
writer = new IndexWriter(reader.directory(), config);
LogMergePolicy lmp =new LogDocMergePolicy();
lmp.setMergeFactor(this.mergeFactor);
config.setMergePolicy(lmp);
long timeBefore = System.currentTimeMillis();
//TODO http://blog.trifork.com/2011/11/21/simon-says-optimize-is-bad-for-you/
//writer.optimize();
long timeAfter = System.currentTimeMillis();
double numSecOnds= ((timeAfter - timeBefore) * 1.0) / DOUBLE;
LOG.log(Level.INFO, "LuceneCatalog: [" + this.catalogPath
+ "] optimized: took: [" + numSeconds + "] seconds");
} catch (IOException e) {
LOG.log(Level.WARNING, "Unable to optimize lucene index: ["
+ catalogPath + "]: Message: " + e.getMessage());
} finally {
try {
writer.close();
} catch (Exception ignore) {
}
}
}

代码示例来源:origin: INL/BlackLab

public static IndexWriterConfig getIndexWriterConfig(Analyzer analyzer, boolean create) {
IndexWriterConfig cOnfig= new IndexWriterConfig(analyzer);
config.setOpenMode(create ? OpenMode.CREATE : OpenMode.CREATE_OR_APPEND);
config.setRAMBufferSizeMB(150); // faster indexing
// Set merge factor (if using LogMergePolicy, which is the default up to version LUCENE_32,
// so yes)
MergePolicy mp = config.getMergePolicy();
if (mp instanceof LogMergePolicy) {
((LogMergePolicy) mp).setMergeFactor(40); // faster indexing
}
return config;
}

代码示例来源:origin: com.strapdata.elasticsearch.test/framework

public static LogMergePolicy newLogMergePolicy(Random r) {
LogMergePolicy logmp = r.nextBoolean() ? new LogDocMergePolicy() : new LogByteSizeMergePolicy();
logmp.setCalibrateSizeByDeletes(r.nextBoolean());
if (rarely(r)) {
logmp.setMergeFactor(TestUtil.nextInt(r, 2, 9));
} else {
logmp.setMergeFactor(TestUtil.nextInt(r, 10, 50));
}
configureRandom(r, logmp);
return logmp;
}

代码示例来源:origin: apache/oodt

config.setOpenMode(IndexWriterConfig.OpenMode.CREATE_OR_APPEND);
LogMergePolicy lmp =new LogDocMergePolicy();
lmp.setMergeFactor(mergeFactor);
config.setMergePolicy(lmp);

代码示例来源:origin: org.apache.oodt/cas-workflow

config.setOpenMode(IndexWriterConfig.OpenMode.CREATE_OR_APPEND);
LogMergePolicy lmp =new LogDocMergePolicy();
lmp.setMergeFactor(mergeFactor);
config.setMergePolicy(lmp);

代码示例来源:origin: apache/oodt

private synchronized void addWorkflowInstanceToCatalog(
WorkflowInstance wInst) throws InstanceRepositoryException {
IndexWriter writer = null;
try {
IndexWriterConfig cOnfig= new IndexWriterConfig(new StandardAnalyzer());
config.setOpenMode(IndexWriterConfig.OpenMode.CREATE_OR_APPEND);
LogMergePolicy lmp =new LogDocMergePolicy();
lmp.setMergeFactor(mergeFactor);
config.setMergePolicy(lmp);
writer = new IndexWriter(indexDir, config);
Document doc = toDoc(wInst);
writer.addDocument(doc);
} catch (IOException e) {
LOG.log(Level.WARNING, "Unable to index workflow instance: ["
+ wInst.getId() + "]: Message: " + e.getMessage());
throw new InstanceRepositoryException(
"Unable to index workflow instance: [" + wInst.getId()
+ "]: Message: " + e.getMessage());
} finally {
try {
writer.close();
} catch (Exception e) {
System.out.println(e);
}
}
}

代码示例来源:origin: org.apache.oodt/cas-workflow

private synchronized void addWorkflowInstanceToCatalog(
WorkflowInstance wInst) throws InstanceRepositoryException {
IndexWriter writer = null;
try {
IndexWriterConfig cOnfig= new IndexWriterConfig(new StandardAnalyzer());
config.setOpenMode(IndexWriterConfig.OpenMode.CREATE_OR_APPEND);
LogMergePolicy lmp =new LogDocMergePolicy();
lmp.setMergeFactor(mergeFactor);
config.setMergePolicy(lmp);
writer = new IndexWriter(indexDir, config);
Document doc = toDoc(wInst);
writer.addDocument(doc);
} catch (IOException e) {
LOG.log(Level.WARNING, "Unable to index workflow instance: ["
+ wInst.getId() + "]: Message: " + e.getMessage());
throw new InstanceRepositoryException(
"Unable to index workflow instance: [" + wInst.getId()
+ "]: Message: " + e.getMessage());
} finally {
try {
writer.close();
} catch (Exception e) {
System.out.println(e);
}
}
}

代码示例来源:origin: apache/oodt

lmp.setMergeFactor(mergeFactor);
config.setMergePolicy(lmp);

代码示例来源:origin: org.apache.oodt/cas-workflow

lmp.setMergeFactor(mergeFactor);
config.setMergePolicy(lmp);

代码示例来源:origin: org.apache.oodt/cas-filemgr

lmp.setMergeFactor(mergeFactor);
config.setMergePolicy(lmp);

代码示例来源:origin: apache/oodt

lmp.setMergeFactor(mergeFactor);
config.setMergePolicy(lmp);

代码示例来源:origin: org.apache.oodt/cas-filemgr

lmp.setMergeFactor(mergeFactor);
config.setMergePolicy(lmp);

代码示例来源:origin: apache/oodt

lmp.setMergeFactor(mergeFactor);
config.setMergePolicy(lmp);

代码示例来源:origin: org.apache.lucene/lucene-benchmark

if (iwConf.getMergePolicy() instanceof LogMergePolicy) {
LogMergePolicy logMergePolicy = (LogMergePolicy) iwConf.getMergePolicy();
logMergePolicy.setMergeFactor(config.get("merge.factor",OpenIndexTask.DEFAULT_MERGE_PFACTOR));

推荐阅读
  • 本文介绍了Java高并发程序设计中线程安全的概念与synchronized关键字的使用。通过一个计数器的例子,演示了多线程同时对变量进行累加操作时可能出现的问题。最终值会小于预期的原因是因为两个线程同时对变量进行写入时,其中一个线程的结果会覆盖另一个线程的结果。为了解决这个问题,可以使用synchronized关键字来保证线程安全。 ... [详细]
  • Iamtryingtomakeaclassthatwillreadatextfileofnamesintoanarray,thenreturnthatarra ... [详细]
  • 本文详细介绍了Java中vector的使用方法和相关知识,包括vector类的功能、构造方法和使用注意事项。通过使用vector类,可以方便地实现动态数组的功能,并且可以随意插入不同类型的对象,进行查找、插入和删除操作。这篇文章对于需要频繁进行查找、插入和删除操作的情况下,使用vector类是一个很好的选择。 ... [详细]
  • Java太阳系小游戏分析和源码详解
    本文介绍了一个基于Java的太阳系小游戏的分析和源码详解。通过对面向对象的知识的学习和实践,作者实现了太阳系各行星绕太阳转的效果。文章详细介绍了游戏的设计思路和源码结构,包括工具类、常量、图片加载、面板等。通过这个小游戏的制作,读者可以巩固和应用所学的知识,如类的继承、方法的重载与重写、多态和封装等。 ... [详细]
  • 本文介绍了使用Java实现大数乘法的分治算法,包括输入数据的处理、普通大数乘法的结果和Karatsuba大数乘法的结果。通过改变long类型可以适应不同范围的大数乘法计算。 ... [详细]
  • 本文介绍了一个Java猜拳小游戏的代码,通过使用Scanner类获取用户输入的拳的数字,并随机生成计算机的拳,然后判断胜负。该游戏可以选择剪刀、石头、布三种拳,通过比较两者的拳来决定胜负。 ... [详细]
  • Java容器中的compareto方法排序原理解析
    本文从源码解析Java容器中的compareto方法的排序原理,讲解了在使用数组存储数据时的限制以及存储效率的问题。同时提到了Redis的五大数据结构和list、set等知识点,回忆了作者大学时代的Java学习经历。文章以作者做的思维导图作为目录,展示了整个讲解过程。 ... [详细]
  • 本文讨论了如何优化解决hdu 1003 java题目的动态规划方法,通过分析加法规则和最大和的性质,提出了一种优化的思路。具体方法是,当从1加到n为负时,即sum(1,n)sum(n,s),可以继续加法计算。同时,还考虑了两种特殊情况:都是负数的情况和有0的情况。最后,通过使用Scanner类来获取输入数据。 ... [详细]
  • 本文介绍了在Java中gt、gtgt、gtgtgt和lt之间的区别。通过解释符号的含义和使用例子,帮助读者理解这些符号在二进制表示和移位操作中的作用。同时,文章还提到了负数的补码表示和移位操作的限制。 ... [详细]
  • 本文介绍了如何在给定的有序字符序列中插入新字符,并保持序列的有序性。通过示例代码演示了插入过程,以及插入后的字符序列。 ... [详细]
  • JavaSE笔试题-接口、抽象类、多态等问题解答
    本文解答了JavaSE笔试题中关于接口、抽象类、多态等问题。包括Math类的取整数方法、接口是否可继承、抽象类是否可实现接口、抽象类是否可继承具体类、抽象类中是否可以有静态main方法等问题。同时介绍了面向对象的特征,以及Java中实现多态的机制。 ... [详细]
  • 本文讨论了一个关于cuowu类的问题,作者在使用cuowu类时遇到了错误提示和使用AdjustmentListener的问题。文章提供了16个解决方案,并给出了两个可能导致错误的原因。 ... [详细]
  • 从零学Java(10)之方法详解,喷打野你真的没我6!
    本文介绍了从零学Java系列中的第10篇文章,详解了Java中的方法。同时讨论了打野过程中喷打野的影响,以及金色打野刀对经济的增加和线上队友经济的影响。指出喷打野会导致线上经济的消减和影响队伍的团结。 ... [详细]
  • 猜字母游戏
    猜字母游戏猜字母游戏——设计数据结构猜字母游戏——设计程序结构猜字母游戏——实现字母生成方法猜字母游戏——实现字母检测方法猜字母游戏——实现主方法1猜字母游戏——设计数据结构1.1 ... [详细]
  • [大整数乘法] java代码实现
    本文介绍了使用java代码实现大整数乘法的过程,同时也涉及到大整数加法和大整数减法的计算方法。通过分治算法来提高计算效率,并对算法的时间复杂度进行了研究。详细代码实现请参考文章链接。 ... [详细]
author-avatar
lucia_8899_458
这个家伙很懒,什么也没留下!
PHP1.CN | 中国最专业的PHP中文社区 | DevBox开发工具箱 | json解析格式化 |PHP资讯 | PHP教程 | 数据库技术 | 服务器技术 | 前端开发技术 | PHP框架 | 开发工具 | 在线工具
Copyright © 1998 - 2020 PHP1.CN. All Rights Reserved | 京公网安备 11010802041100号 | 京ICP备19059560号-4 | PHP1.CN 第一PHP社区 版权所有