spark 怎么读写 elasticsearch
参考这两篇文章: https://www.bmc.com/blogs/spark-elasticsearch-hadoop/ https://blog.pythian.com/updating-elasticsearch-indexes-spark/ Pre-requisite: 先装上 elasticsearch-hadoop 包 Step-by-Step guide 1. 先在ES创建3个document [mshuai@node1 ~]$ curl -XPUT --header 'Content-Type: application/json' http://your_ip:your_port/school/doc/1 -d '{ "school" : "Clemson" }' [mshuai@node1 ~]$ curl -XPUT --header 'Content-Type: application/json' http://your_ip:your_port/school/doc/2 -d '{ "school" : "Harvard"}' 2. Spark 里面去读,这里是 pyspark 代码 reader = spark.read.format("org.elasticsearch.spark.sql").option("es.read.metadata",