Exploding Array in Batches of size 'n'
问题 Looking to explode a nested array w/ Spark into batches. The column below is a nested array from an XML files. Now attempting to write the time series data into batches in order to write over to a NoSQL database. For example: +-------+-----------------------+ | ID | Example | +-------+-----------------------+ | A| [[1,2],[3,4],[5,6]] | +-------+-----------------------+ Output with batches of size 2 +-------+-----------------------+ | ID | Example | +-------+-----------------------+ | A| [[1,2