How to fill the job submit form in the Data Processing page from the OVHcloud Manager
Find out how to fill the job submit form from the Data Processing page in the OVHcloud Manager
Find out how to fill the job submit form from the Data Processing page in the OVHcloud Manager
Last updated 06th March, 2020
This guide will help you to understand the different parameters that you can set when submitting a new job to the Data Processing platform
In this guide, we are assuming that you're using the OVHcloud Manager to use the Data Processing platform.
To read an introduction about Data Processing service you can visit Data Processing Overview.
To submit your job with your required parameters follow these steps:
Public Cloud
Data Processing
from the left panel. Submit a new job
Next
button.In this step you will have to select the region in which you would like your Data Processing job to be run. The region list you will see will only contain the currently supported regions. Select a region and then click on Next
button.
An OVHcloud Data Processing job being executed in a distributed environment, you will have to specify the amount of resources you would like your job to use. The resources you will have to specify are going to depend on the engine you selected previously.
For the Apache Spark Engine, the resources you will have to specify are:
If you want to know more about how to size your resources or how Apache Spark works, visit Apache Spark Documentation.
You can choose how to size your resources by either selecting some templates from the default view or clicking on Advanced configuration and setting everything by hand.
If you click on Advanced configuration
, you will have more options to configure your Driver and Executors and also you are not limited to some pre-defined templates. In advanced mode, you can change memory overhead for Drivers and Executors as well. Memory overhead is the amount of memory that each node of cluster requires for running Apache Spark processes itself.
When you configured compute and memory of your cluster, click on Next
button to go to the next step.
Follow these steps to configure your job before submitting it to the Data Processing service:
For Apache Spark, for example, you will also have to :
Submit Job
button and your application will be sent to the Data Processing platform and should start shortly after that.The arguments of the application are stored in plain text. It is advised that you store your credentials in configuration files instead of using arguments in the Manager. You need to upload the configuration files in the same Object Storage container that you upload the code so they will be downloaded to the data processing cluster together when you submit the job.
To learn more about using Data Processing and how to submit a job and process your data, we invite you to look at Data Processing documentations page.
You can send your questions, suggestions or feedbacks in our community of users on https://community.ovh.com/en/ or on our Discord in the channel #dataprocessing-spark
Please feel free to give any suggestions in order to improve this documentation.
Whether your feedback is about images, content, or structure, please share it, so that we can improve it together.
Your support requests will not be processed via this form. To do this, please use the "Create a ticket" form.
Thank you. Your feedback has been received.
Access your community space. Ask questions, search for information, post content, and interact with other OVHcloud Community members.
Discuss with the OVHcloud community