[DOCS] Fix Missing Documentation for SparkSession in Declarative Pipelines (Python) #53283
+19
−5
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Signed-off-by: Avril Aysha [email protected]
What changes were proposed in this pull request?
I’m experimenting with the new Declarative Pipelines API in Spark 4.1 (preview), and I noticed a gap in the documentation regarding how Python transformations are expected to obtain a SparkSession.
This PR adds documentation to the Python transformation section to show how to access the existing Spark session inside each transformation.
Why are the changes needed?
Without the
getActiveSessionline, thespark-pipelines run...command throws aNameError: name 'spark' is not defined.Does this PR introduce any user-facing change?
No
How was this patch tested?
Docs were built locally
Was this patch authored or co-authored using generative AI tooling?
No