I just start learning spark , I'm a bit confused by the this concept, so from the spark installation , we get the pyspark under the spark installation sub-folders , I understand it's a shell, and from the python package we also can also install the python package thru pip install pyspark, so we can run the python code instead of submitting it to the cluster , so what's the difference between these two ? also in anaconda we can use findspark and use the pyspark from there , so does that mean it's not using the pyspark from the python package ?
Plus in the real world spark application development , what is being used in which scenario ? thanks in advance .