Spark specify multiple column conditions for dataframe join
How to give more column conditions when joining two dataframes. For example I want to run the following : val Lead_all = Leads.join(Utm_Master, Leaddetails.columns("LeadSource","Utm_Source","Utm_Medium","Utm_Campaign") == Utm_Master.columns("LeadSource","Utm_Source","Utm_Medium","Utm_Campaign"), "left") I want to join only when these columns match. But above syntax is not valid as cols only takes one string. So how do I get what I want. rchukh There is a Spark column/expression API join for such case: Leaddetails.join( Utm_Master, Leaddetails("LeadSource") <=> Utm_Master("LeadSource") &&