datastax-enterprise

How to create a graph and its schema without using Datastax Studio but through Java?

老子叫甜甜 提交于 2020-01-03 02:22:13
问题 I was trying to create my first connection with DSE Graph through java.. public static void main(String args[]){ DseCluster dseCluster = null; try { dseCluster = DseCluster.builder() .addContactPoint("192.168.1.43") .build(); DseSession dseSession = dseCluster.connect(); GraphTraversalSource g = DseGraph.traversal(dseSession, new GraphOptions().setGraphName("graph")); GraphStatement graphStatement = DseGraph.statementFromTraversal(g.addV("test")); GraphResultSet grs = dseSession.executeGraph

Datastax opscenter upgrade fails to display stats

。_饼干妹妹 提交于 2020-01-02 22:34:10
问题 I upgraded Opscenter from 5.1.2 to 5.2.0 yesterday, and now none of the graphs on the Dashboard are showing any statistics. My cluster is datastax enterprise 4.5.1 with the following versions: cqlsh 4.1.1 | Cassandra 2.0.8.39 | CQL spec 3.1.1 | Thrift protocol 19.39.0 I'm using this cluster for a search workload with solr. The agent.log is filled with the following: INFO [qtp1313948736-24] 2015-08-06 12:30:52,211 New JMX connection (127.0.0.1:7199) INFO [qtp1313948736-24] 2015-08-06 12:30:52

How to test a Spark SQL Query without Scala

一曲冷凌霜 提交于 2020-01-02 20:43:32
问题 I am trying to figure out how to test Spark SQL queries against a Cassandra database -- kind of like you would in SQL Server Management Studio. Currently I have to open the Spark Console and type Scala commands which is really tedious and error prone. Something like: scala > var query = csc.sql("select * from users"); scala > query.collect().foreach(println) Especially with longer queries this can be a real pain. This seems like a terribly inefficient way to test if your query is correct and

How to submit a job via REST API?

依然范特西╮ 提交于 2020-01-01 06:55:29
问题 I'm using Datastax Enterprise 4.8.3. I'm trying to implement a Quartz based application to remotely submit Spark jobs. During my research I have stumbled upon the following links: Apache Spark Hidden REST API Spark feature - Provide a stable application submission gateway in standalone cluster mode To test out the theory I tried executing the below code snippet on the master node (IP: "spark-master-ip"; directly on the shell) of my 2 node cluster (as provided in link #1 above): curl -X POST

How to submit a job via REST API?

瘦欲@ 提交于 2020-01-01 06:55:02
问题 I'm using Datastax Enterprise 4.8.3. I'm trying to implement a Quartz based application to remotely submit Spark jobs. During my research I have stumbled upon the following links: Apache Spark Hidden REST API Spark feature - Provide a stable application submission gateway in standalone cluster mode To test out the theory I tried executing the below code snippet on the master node (IP: "spark-master-ip"; directly on the shell) of my 2 node cluster (as provided in link #1 above): curl -X POST

How can I filter `filter(lambda x:len(x[1])>=2)` in dataframe?

强颜欢笑 提交于 2019-12-25 15:40:29
问题 I am not sure about how to filter(lambda x:len(x[1])>=2) in dataframe. I would like to improve the speed of my spark app. Thanks for your help! This some context from my spark app: article_ids = sqlContext.read.format("org.apache.spark.sql.cassandra").options(table="article_by_created_at", keyspace=source).load().where(range_expr).select('article','created_at').repartition(64*2) axes = sqlContext.read.format("org.apache.spark.sql.cassandra").options(table="axes", keyspace=source).load() speed

Is `nodetool disableautocompaction` persisted between restarts of Cassandra?

妖精的绣舞 提交于 2019-12-25 11:50:36
问题 Is nodetool disableautocompaction persisted between restarts of Cassandra? Useful to know in case Cassandra would run out of memory or something. 回答1: No, to disable it across restarts you need to set it in the cassandra.yaml 来源: https://stackoverflow.com/questions/47834953/is-nodetool-disableautocompaction-persisted-between-restarts-of-cassandra

How can we set nodetool and cqlsh to be run from anywhere and by any user on linux server

无人久伴 提交于 2019-12-25 07:58:26
问题 I am trying to setup environment variables so that any user on a particular server can run commands like nodetool or cqlsh from any where in linux file system . The effort to traverse to bin directory everytime should be saved . How can we achieve this ? My DSE 4.8 is a tarball install . 回答1: Nodetool is usually available to any user that has execution privileges in your linux boxes For cqlsh, you can set any configuration inside the cqlshrc file (usually found in $HOME/.cassandra/cqlshrc; we

Datastax OpsCenter upgrade from 6.0.3 to 6.0.5: migration of lcm.db

♀尐吖头ヾ 提交于 2019-12-25 07:35:53
问题 I upgraded OpsCenter from 6.0.3 to 6.0.5. Is there a way to migrate the LCM data? Is it as simple as copying the file lcm.db? Thanks. 回答1: OpsCenter developer here. 6.0.3 to 6.0.5 is a safe upgrade to do "in-place". So if it's a package you can just 'yum upgrade' or 'apt-get upgrade'. If you do want to backup/restore your settings for whatever reason, the best docs to look at are our high-availability docs (HA failover is basically a backup/restore): https://docs.datastax.com/en/opscenter/6.0

Fault tolerance and topology transparency of multi-node DSE Cluster

妖精的绣舞 提交于 2019-12-25 04:45:30
问题 I have the following DSE cluster setup: DC Cassandra Cassandra node 1 DC Solr Solr node 1 Solr node 2 Solr node 3 Solr node 4 The replication factor is 1 for each DC My questions: To perform a search, I send a Solr SELECT query to a specific node. This introduces a single point of failure. If the node is down, the query fails. Is there a way to "query the cluster/DC" instead of querying a specific node? In order for the result-set to be complete, I need to manually specify the other nodes via