Is it possible to partition data based on column which may contain empty value sometimes? how to handle it?

回眸只為那壹抹淺笑 提交于 2020-05-17 07:02:09

问题


I am using spark-sql-2.4.1v. I need to join two datasets ds1 and ds2 (new-column) based on some field. val resultDs = ds1.join(ds2, , "inner");

resultDs now contain "new-column" but for the records which not met does not have this "new-column". Hence for those records I need to set this "new-column" null/empty.

But as per my business requirement I need to parition the resultDs on "new-column". so how generally these kind of scenarios are handled ? pleas advice.

来源:https://stackoverflow.com/questions/61607228/is-it-possible-to-partition-data-based-on-column-which-may-contain-empty-value-s

易学教程内所有资源均来自网络或用户发布的内容,如有违反法律规定的内容欢迎反馈
该文章没有解决你所遇到的问题?点击提问,说说你的问题,让更多的人一起探讨吧!