I have a cassandra table containing 3 million rows. Now I am trying to fetch all the rows and write them to several csv files. I know it is impossible to perform selec
You could use Pig to read the data and store it into HDFS, then copy it out as a single file:
In Pig:
data = LOAD 'cql://your_ksp/your_table' USING CqlStorage();
STORE data INTO '/path/to/output' USING PigStorage(',');
From OS shell:
hadoop fs -copyToLocal hdfs://hadoop_url/path/to/output /path/to/local/storage