I want to use Apache Spark and connect to Vertica by JDBC.
In Vertica database, I have 100 million records and spark code runs on another server.
When I run
After your spark jobs finishes logon to the Vertica database using the same credentials that the spark job used and run:
SELECT * FROM v_monitor.query_requests ORDER BY start_timetamp DESC LIMIT 10000;
This will show you the queries sent to the database by the spark job, allowing you to see if it pushed down the count(*) to the database or if it indeed tried to retrieve the entire table across the network.