Elasticsearch问题汇总

2022-01-12 12:22:05 浏览数 (1)

前言

本文主要基于Elasticsearch 6.5.4版本:

1 2 3 4 5

<dependency> <groupId>org.elasticsearch</groupId> <artifactId>elasticsearch</artifactId> <version>6.5.4</version> </dependency>

too_many_clauses问题

Elasticsearch查询时报错如下:

1 2

"caused_by":{"type":"too_many_clauses","reason":"maxClauseCount is set to 1024"}}}], "caused_by":{"type":"query_shard_exception","reason":"failed to create query:

这是bool查询的条件超过了默认的1024上限,可以通过修改全局配置来增加上限,需要注意的是别设置太高,会消耗太多的CPU资源和内存。

打开ES的配置文件/config/elasticsearch.yml,增加配置:

1

indices.query.bool.max_clause_count: 2048

修改全局配置后需要重启ES才能生效。如果不允许重启ES集群,就只能从查询语句入手了,要么削减查询条件的数量,要么将查询条件转移到must_notterms查询中。

must_notterms查询可以超过默认的1024上限,对于肯定条件可以用must_not嵌套must_not来实现。(这种做法是其他博主验证的,这里只提一嘴,在短期内无法重启ES集群时可以作为临时方案使用。)

修改jvm参数

Elasticsearch是用Java开发的,默认会配置1G的jvm堆的初始值和最大值,该jvm参数被配置在/config/jvm.options里:

1 2

-Xms1g -Xmx1g

如果只是个人开发小项目,可以把参数改小些(当然不能调整太小,内存对于ES的性能影响较大),比如:

1 2

-Xms512m -Xmx512m

这个jvm.options用来配置各种jvm参数,比如GC、GC logging、heap dumps等。

cannot write xcontent for unknown value of type class java.math.BigDecimal

Elasticsearch在索引数据时报错如下:

1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19

java.lang.IllegalArgumentException: cannot write xcontent for unknown value of type class java.math.BigDecimal at org.elasticsearch.common.xcontent.XContentBuilder.unknownValue(XContentBuilder.java:755) at org.elasticsearch.common.xcontent.XContentBuilder.value(XContentBuilder.java:726) at org.elasticsearch.common.xcontent.XContentBuilder.field(XContentBuilder.java:711) at org.elasticsearch.index.query.BaseTermQueryBuilder.doXContent(BaseTermQueryBuilder.java:154) at org.elasticsearch.index.query.AbstractQueryBuilder.toXContent(AbstractQueryBuilder.java:82) at org.elasticsearch.index.query.BoolQueryBuilder.doXArrayContent(BoolQueryBuilder.java:275) at org.elasticsearch.index.query.BoolQueryBuilder.doXContent(BoolQueryBuilder.java:256) at org.elasticsearch.index.query.AbstractQueryBuilder.toXContent(AbstractQueryBuilder.java:82) at org.elasticsearch.common.xcontent.XContentBuilder.value(XContentBuilder.java:779) at org.elasticsearch.common.xcontent.XContentBuilder.value(XContentBuilder.java:772) at org.elasticsearch.common.xcontent.XContentBuilder.field(XContentBuilder.java:764) at org.elasticsearch.search.builder.SearchSourceBuilder.toXContent(SearchSourceBuilder.java:1184) at org.elasticsearch.common.xcontent.XContentHelper.toXContent(XContentHelper.java:349) at org.elasticsearch.search.builder.SearchSourceBuilder.toString(SearchSourceBuilder.java:1558) at org.elasticsearch.search.builder.SearchSourceBuilder.toString(SearchSourceBuilder.java:1553) at java.lang.String.valueOf(String.java:2994) at java.lang.StringBuilder.append(StringBuilder.java:131) at org.elasticsearch.action.search.SearchRequest.toString(SearchRequest.java:516)

从异常信息看,显然ES无法接受BigDecimal类型,经过百度,也确实如此。在一篇博文评论中解释如下:

应该是客户端代码里将查询的数值定义成了java.math.BigDecimal,而ES不支持这个类型。之所以2.2没有问题,是因为之前的transport client发送数据之前将其序列化成了json,而在5.x以后,使用的内部的transport protocol,数据类型如果不匹配会抛错误。 所以数据类型的定义上,需要使用ES支持的类型。

解决方案一:转变成其他ES支持的数据类型

我使用的是6.5.4版本的Elasticsearch,该版本尚不支持BigDecimal或者BigInteger的数据类型,所以在index到Elasticsearch之前,需要转换成其他数据类型,这里要注意不要数据溢出了:

  1. BigDecimal要转变成Double类型
  2. BigInteger要转变成Long类型

解决方案二:使用更高版本的ES

我在看6.7.1版本的Elasticsearch源码时发现已经可以支持BigDecimal或者BigInteger的数据类型了,所以直接使用该版本或更高版本的就行了。

下面附上两个版本的支持的数据类型的源码:

  • 6.5.4版本的Elasticsearch相关源码

1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22

Map<Class<?>, Writer> writers = new HashMap<>(); writers.put(Boolean.class, (b, v) -> b.value((Boolean) v)); writers.put(Byte.class, (b, v) -> b.value((Byte) v)); writers.put(byte[].class, (b, v) -> b.value((byte[]) v)); writers.put(Date.class, XContentBuilder::timeValue); writers.put(Double.class, (b, v) -> b.value((Double) v)); writers.put(double[].class, (b, v) -> b.values((double[]) v)); writers.put(Float.class, (b, v) -> b.value((Float) v)); writers.put(float[].class, (b, v) -> b.values((float[]) v)); writers.put(Integer.class, (b, v) -> b.value((Integer) v)); writers.put(int[].class, (b, v) -> b.values((int[]) v)); writers.put(Long.class, (b, v) -> b.value((Long) v)); writers.put(long[].class, (b, v) -> b.values((long[]) v)); writers.put(Short.class, (b, v) -> b.value((Short) v)); writers.put(short[].class, (b, v) -> b.values((short[]) v)); writers.put(String.class, (b, v) -> b.value((String) v)); writers.put(String[].class, (b, v) -> b.values((String[]) v)); writers.put(Locale.class, (b, v) -> b.value(v.toString())); writers.put(Class.class, (b, v) -> b.value(v.toString())); writers.put(ZonedDateTime.class, (b, v) -> b.value(v.toString())); writers.put(Calendar.class, XContentBuilder::timeValue); writers.put(GregorianCalendar.class, XContentBuilder::timeValue);

  • 6.7.1版本的Elasticsearch相关源码

1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24

Map<Class<?>, Writer> writers = new HashMap<>(); writers.put(Boolean.class, (b, v) -> b.value((Boolean) v)); writers.put(Byte.class, (b, v) -> b.value((Byte) v)); writers.put(byte[].class, (b, v) -> b.value((byte[]) v)); writers.put(Date.class, XContentBuilder::timeValue); writers.put(Double.class, (b, v) -> b.value((Double) v)); writers.put(double[].class, (b, v) -> b.values((double[]) v)); writers.put(Float.class, (b, v) -> b.value((Float) v)); writers.put(float[].class, (b, v) -> b.values((float[]) v)); writers.put(Integer.class, (b, v) -> b.value((Integer) v)); writers.put(int[].class, (b, v) -> b.values((int[]) v)); writers.put(Long.class, (b, v) -> b.value((Long) v)); writers.put(long[].class, (b, v) -> b.values((long[]) v)); writers.put(Short.class, (b, v) -> b.value((Short) v)); writers.put(short[].class, (b, v) -> b.values((short[]) v)); writers.put(String.class, (b, v) -> b.value((String) v)); writers.put(String[].class, (b, v) -> b.values((String[]) v)); writers.put(Locale.class, (b, v) -> b.value(v.toString())); writers.put(Class.class, (b, v) -> b.value(v.toString())); writers.put(ZonedDateTime.class, (b, v) -> b.value(v.toString())); writers.put(Calendar.class, XContentBuilder::timeValue); writers.put(GregorianCalendar.class, XContentBuilder::timeValue); writers.put(BigInteger.class, (b, v) -> b.value((BigInteger) v)); writers.put(BigDecimal.class, (b, v) -> b.value((BigDecimal) v));

可以发现,在6.7.1版本的源码里,多出了最后的两种数据类型的支持:BigInteger和BigDecimal。

Limit of total fields [1000] in index has been exceeded

在索引数据时ES抛出异常:

1

cause: ElasticsearchException[Elasticsearch exception [type=illegal_argument_exception, reason=Limit of total fields [1000] in index [item] has been exceeded]]

这是由于被索引的文档字段数量超过了默认的1000上限,两种解决方法,要么减少文档的字段,要么增加字段上限。

增加字段上限可以只设置某个索引,也可以设置为全局的配置,对所有已存在的索引生效,但对之后新建的索引是无效的。

1 2 3 4 5 6 7 8 9 10 11 12 13

// 只设置test索引的配置 PUT http://localhost:9200/test/_settings { "index.mapping.total_fields.limit": 5000 } // 全局的配置 PUT http://localhost:9200/_settings { "index.mapping.total_fields.limit": 5000 }

FORBIDDEN/12/index read-only / allow delete (api)

索引数据时报错如下:

1

cause: ElasticsearchException[Elasticsearch exception [type=cluster_block_exception, reason=blocked by: [FORBIDDEN/12/index read-only / allow delete (api)];]]

这是ES节点的数据目录data磁盘空间使用率超过90%导致的,为了保护数据,ES将索引变为只读模式,只允许删除。

此时需要增大磁盘的使用空间,有如下多种方法:

  1. 集群增加节点
  2. 降低集群的索引副本数量
  3. 清理磁盘无用的数据,比如日志等

ES应该尽量别和其他项目部署在一起,磁盘容易被其他项目的日志挤占。此外,ES本身的日志和数据存储目录也可以配置在不同的目录,需要更改配置文件/config/elasticsearch.yml

1 2 3 4 5 6 7 8 9

# ----------------------------------- Paths ------------------------------------ # # Path to directory where to store the data (separate multiple locations by comma): # #path.data: /path/to/data # # Path to log files: # #path.logs: /path/to/logs

在增大了磁盘的使用空间后,索引的只读状态需要手动更改回来,可以更改所有索引,也可以只指定某个索引(用对应的索引名字取代_all_all表示所有索引,如果不指定索引名,也不使用_all,同样表示修改全局配置):

1 2 3 4 5 6 7

// curl方式 curl -XPUT -H "Content-Type: application/json" http://localhost:9200/_all/_settings -d '{"index.blocks.read_only_allow_delete": null}' // RESTful方式 PUT http://localhost:9200/_all/_settings {"index.blocks.read_only_allow_delete": null}

Result window is too large

报错如下:

1 2 3 4

"root_cause": [{ "type": "illegal_argument_exception", "reason": "Result window is too large, from size must be less than or equal to: [10000] but was [80000]. See the scroll api for a more efficient way to request large data sets. This limit can be set by changing the [index.max_result_window] index level setting." }]

ES分页查询(from size)默认的最大查询结果数量为10000,可以通过修改max_result_window的值来提高上限:

1 2 3 4 5 6 7

// curl方式 curl -XPUT -H "Content-Type: application/json" http://localhost:9200/_all/_settings -d '{"index.max_result_window" :"100000"}' // RESTful方式 PUT http://localhost:9200/_all/_settings {"index.max_result_window" :"100000"}

failed to parse date field

在查询es时报错如下:

1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26

{ "error": { "root_cause": [{ "type": "parse_exception", "reason": "failed to parse date field [2021-06-15 00:00:00] with format [strict_date_optional_time||epoch_millis]" }], "type": "search_phase_execution_exception", "reason": "all shards failed", "phase": "query", "grouped": true, "failed_shards": [{ "shard": 0, "index": "inspectbooking_kmt", "node": "Its-Juf2QXKEwmYYNu_aBQ", "reason": { "type": "parse_exception", "reason": "failed to parse date field [2021-06-15 00:00:00] with format [strict_date_optional_time||epoch_millis]", "caused_by": { "type": "illegal_argument_exception", "reason": "Unrecognized chars at the end of [2021-06-15 00:00:00]: [ 00:00:00]" } } }] }, "status": 400 }

这是因为es的日期默认使用strict_date_optional_timeepoch_millis的format来匹配,前者是严格的ISO日期格式,后者是毫秒值格式。

这里由于搜索日期值使用的是2021-06-15 00:00:00这种格式,无法被es的日期解析器解析成上述的两种格式,因此抛出异常。要避免这种异常,要么修改mapping中日期字段的format,比如说用||添加新的格式;要么修改搜索日期时输入的值。

由于mapping一旦确定就无法更改,因此更推荐改变被搜索的日期值格式这种做法:

1 2 3 4 5

DateTimeFormatter dateTimePattern = DateTimeFormatter.ofPattern("yyyy-MM-dd HH:mm:ss"); // 日期字符串是从db中获取的零时区日期 TemporalAccessor parseDateTime = dateTimePattern.parse("2021-06-15 00:00:00"); LocalDateTime localDateTime = LocalDateTime.from(parseDateTime); DateTimeFormatter.ISO_LOCAL_DATE_TIME.format(localDateTime);

Request size exceeded 104857600 bytes

在使用bulk批处理ES请求时,报错如下:

1 2 3 4 5 6

URI [/_bulk?timeout=1m], status line [HTTP/1.1 413 Request Entity Too Large] {"Message":"Request size exceeded 104857600 bytes"} at org.elasticsearch.client.RestClient$SyncResponseListener.get(RestClient.java:926) at org.elasticsearch.client.RestClient.performRequest(RestClient.java:229) at org.elasticsearch.client.RestHighLevelClient.internalPerformRequest(RestHighLevelClient.java:1593) ...

ES默认的请求内容最大值为100mb,超过100mb就会报错,可以在/config/elasticsearch.yml中添加如下配置:

1

http.max_content_length: 200mb

注:AWS的ES似乎没有提供修改ES配置文件的服务,在请求最大值上也只有10mb或者100mb的两种选项。

  • AWS Elasticsearch cluster method to update http.max_content_length?

参考链接

  • Elasticsearch Guide 6.7 - Search Settings
  • ES 问题 : too_many_clauses maxClauseCount is set to 1024
  • elastic search 5.4.版本,java api 调用出现:can not write type 【class java.math.BigDecimal】
  • java.lang.IllegalArgumentException: Limit of total fields 【1000】 in index 索引名称
  • ES 写索引报错 FORBIDDEN/12/index read-only / allow delete (api)解决方案
  • ES集群修改index副本数报错 :index read-only / allow delete
  • ES更改参数max_result_window
  • Elasticsearch date 类型详解
  • hive向ES中插入数据量过大时出错:HTTP content length exceeded 104857600 bytes.

0 人点赞