Skip to content

Getting pyspark working #7543

@thomashan

Description

@thomashan

Hi there.
I couldn't find any documentation for initializing pyspark.

I've initialized the spark sessions with

%%classpath add mvn
org.apache.spark spark-sql_2.11 2.3.1

followed by

%%spark

Everything is all good.

Then when I go to link the spark session in python I don't have an entry point to the spark because spark has been initialized through the jar files and not through python.

Is there a tutorial or documentation which outlines how to use spark in python or through pyspark?

Cheers.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions