Skip to content
This repository was archived by the owner on Jan 9, 2020. It is now read-only.
This repository was archived by the owner on Jan 9, 2020. It is now read-only.

Documentation plan for Spark #262

@foxish

Description

@foxish

As discussed in the SIG meeting earlier today, we need documentation coverage which is much more comprehensive to ensure that people have a good kick-the-tyres experience when using Spark on Kubernetes.
This will especially be a concern after our documentation change goes live.

Various broad categories of items that come to mind:

  • High level overview of Kubernetes as a cluster scheduler backend for Spark
  • Running a SparkJob from the beta release
  • Workflow
    • Building custom images for drivers and executors
    • Using the file submission server and local submissions
    • Using an external storage mechanism (File Server/S3/GCS/external HDFS cluster)
  • Running a development build
    • Developer workflow
    • Integration tests
  • Debugging a SparkJob
    • Using kubectl to identify memory, disk or other issues causing failure
    • Using the Spark UI and dashboard
  • Best Practices and FAQ
  • Reporting issues and submitting PRs

Thoughts on any items we've missed? I expect that each of the top-level items will be pages of their own.
cc @apache-spark-on-k8s/contributors

Metadata

Metadata

Assignees

Labels

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions