How to submit a Python job on the Data Processing platform using the OVHcloud manager
Find out how to create a cluster and run your Apache Spark Python job with Data Processing platform using the OVHcloud manager
Find out how to create a cluster and run your Apache Spark Python job with Data Processing platform using the OVHcloud manager
Last updated 04th May, 2020
This guide helps you to upload an Apache Spark job in Python to your OVHcloud Object Storage and run your job with the Data Processing page in the Manager.
If you want to submit an Apache Spark job in Java or Scala language, you can read this document: How to submit a Java/Scala job using Data Processing Manager
In this guide, we are assuming that you're using the OVHcloud Manager to use the Data Processing platform.
To read an introduction about the Data Processing service you can visit Data Processing Overview.
Before running your job in the Data Processing platform, you will need to create a container in your Object Storage. Then you will need to upload your application Python files and environment.yml file at the root of this container. You can work with your Object Storage using either the OVHcloud manager or the Openstack Horizon dashboard.
Please see Creating Storage Containers in Customer Panel or Create an object container in Horizon for more details.
If you don’t currently have an application code and you still would like to try OVHcloud Data Processing, you can download and use the PI sample from Apache Spark repository.
If your application has some package requirements or needs a specific version of Python to run your job, make sure that you mention them in your Conda environment.yml file. You can learn how to generate environment file here.
To submit your job with your required parameters follow these steps:
Public Cloud
Data Processing
from the left panelSubmit a new job
Submit job
button to submit your Apache Spark job.Please see How to fill job submit form in Data Processing Manager for more details.
In the Data Processing section of the OVHcloud Manager you can see the list of all the jobs that you have submitted so far. If you click on a job's name, you can see detailed information on it, including its status. Then you can click on the Logs
to see the live logs while the job is running.
If your jobs are stuck in "Running", you probably forgot to stop the spark context in your code. To stop it, please refer to pyspark package documentation.
Once the job will be finished, the complete logs will be saved to your Object Storage container. You will be able to download it from your account whenever you would like.
Please see How to check your job's logs in the Data Processing manager page for more details.
After your Spark job is finished, you will be able to check the results from your logs as well as in any connected storage your job was designed to update.
To learn more about using Data Processing and how to submit a job and process your data, we invite you to look at Data Processing documentations page.
You can send your questions, suggestions or feedbacks in our community of users on https://community.ovh.com/en/ or on our Discord in the channel #dataprocessing-spark
Prima di inviare la valutazione, proponici dei suggerimenti per migliorare la documentazione.
Immagini, contenuti, struttura... Spiegaci perché, così possiamo migliorarla insieme!
Le richieste di assistenza non sono gestite con questo form. Se ti serve supporto, utilizza il form "Crea un ticket" .
Grazie per averci inviato il tuo feedback.
Accedi al tuo spazio nella Community Fai domande, cerca informazioni, pubblica contenuti e interagisci con gli altri membri della Community OVHcloud
Discuss with the OVHcloud community