In SQL we are able to select entries from a TABLE A based on a column in TABLE B.
Please see below:
SELECT * FROM TABLE A
WHERE NAME in (SELECT NAME FROM TABLE B)
How do I replicate this piece of code in pyspark without using a sql context?
In SQL we are able to select entries from a TABLE A based on a column in TABLE B.
Please see below:
SELECT * FROM TABLE A
WHERE NAME in (SELECT NAME FROM TABLE B)
How do I replicate this piece of code in pyspark without using a sql context?
Df=A.join(B, on =['Name'], how ='inner' ).select(A.columns)