diff --git a/src/docs/ocean-spark/_sidebar.md b/src/docs/ocean-spark/_sidebar.md index 7a525c4e93..21761ebedc 100644 --- a/src/docs/ocean-spark/_sidebar.md +++ b/src/docs/ocean-spark/_sidebar.md @@ -29,6 +29,8 @@ - [Run Apps from Airflow](ocean-spark/tools-integrations/run-apps-from-airflow) - [Spark Connect](ocean-spark/tools-integrations/spark-connect) - [JDBC](ocean-spark/tools-integrations/jdbc) + - [JupyterLab](ocean-spark/tools-integrations/jupyter-lab) + - [VS Code](ocean-spark/tools-integrations/vs-code) - [Hive Metastore](ocean-spark/tools-integrations/hive-metastore) - [AWS Glue Data Catalog](ocean-spark/tools-integrations/aws-glue-catalog) - [Docker Images Release Notes](ocean-spark/docker-images-release-notes/) diff --git a/src/docs/ocean-spark/tools-integrations/jupyter-lab.md b/src/docs/ocean-spark/tools-integrations/jupyter-lab.md new file mode 100644 index 0000000000..29dee7b39b --- /dev/null +++ b/src/docs/ocean-spark/tools-integrations/jupyter-lab.md @@ -0,0 +1,38 @@ +# Jupyter Lab + +Install and launch JupyterLab server on any Spark application image with Python installed. + +## Configuration + +To start a Spark application with SparkConnect server, either run the mainClass SparkConnectServer or enable the SparkConnect plugin. Using the Spark Connect plugin, the application can run other tasks or services while enabling Spark Connect. + +### Launch JupyterLab using the JupyterLab main class + +```json +"mainClass": "com.netapp.spark.JupyterLab", +"deps": { + "packages": ["com.netapp.spark:jupyter-plugin:1.3.0"], + "repositories": ["https://us-central1-maven.pkg.dev/ocean-spark/ocean-spark-adapters"] +}, +``` + +### Launch JupyterLab using the spark plugin + +```json +"sparkConf": { + "spark.plugins": "com.netapp.spark.JupyterPlugin", + "spark.jupyter.work.dir": "/opt/spark/work-dir" +}, +"deps": { + "packages": ["com.netapp.spark:jupyter-plugin:1.3.0"], + "repositories": ["https://us-central1-maven.pkg.dev/ocean-spark/ocean-spark-adapters"] +}, +``` + +## Access + +Access the JupyterLab server from a URL in the following format + +``` +https://console.spotinst.com/api/ocean/spark/cluster/osc-mycluster/app/spark-myapp/notebook/ +``` \ No newline at end of file diff --git a/src/docs/ocean-spark/tools-integrations/vs-code.md b/src/docs/ocean-spark/tools-integrations/vs-code.md new file mode 100644 index 0000000000..e1990848e5 --- /dev/null +++ b/src/docs/ocean-spark/tools-integrations/vs-code.md @@ -0,0 +1,38 @@ +# VS Code Server + +Install and launch VS Code server on any Spark application image with Python installed. + +## Configuration + +To start a Spark application with SparkConnect server, either run the mainClass SparkConnectServer or enable the SparkConnect plugin. Using the Spark Connect plugin, the application can run other tasks or services while enabling Spark Connect. + +### Launch VS Code Server using the VSCodeServer main class + +```json +"mainClass": "com.netapp.spark.VSCodeServer", +"deps": { + "packages": ["com.netapp.spark:vscode-plugin:1.3.0"], + "repositories": ["https://us-central1-maven.pkg.dev/ocean-spark/ocean-spark-adapters"] +}, +``` + +### Launch VS Code Server using the spark plugin + +```json +"sparkConf": { + "spark.plugins": "com.netapp.spark.SparkCodeServerPlugin", + "spark.jupyter.work.dir": "/opt/spark/work-dir" +}, +"deps": { + "packages": ["com.netapp.spark:jupyter-plugin:1.3.0"], + "repositories": ["https://us-central1-maven.pkg.dev/ocean-spark/ocean-spark-adapters"] +}, +``` + +## Access + +Access the JupyterLab server from a URL in the following format + +``` +https://console.spotinst.com/api/ocean/spark/cluster/osc-mycluster/app/spark-myapp/code/ +``` \ No newline at end of file