mirror of
				https://github.com/clearml/clearml-docs
				synced 2025-06-26 18:17:44 +00:00 
			
		
		
		
	Small edits (#987)
This commit is contained in:
		
							parent
							
								
									058e78e313
								
							
						
					
					
						commit
						a1d2843a3b
					
				@ -65,7 +65,7 @@ See the [HyperParameterOptimizer SDK reference page](../references/sdk/hpo_optim
 | 
			
		||||
### Pipeline
 | 
			
		||||
 | 
			
		||||
ClearML's `automation` module includes classes that support creating pipelines: 
 | 
			
		||||
  * [PipelineController](../pipelines/pipelines_sdk_tasks.md) - A pythonic interface for 
 | 
			
		||||
  * [PipelineController](../pipelines/pipelines_sdk_tasks.md) - A Pythonic interface for 
 | 
			
		||||
    defining and configuring a pipeline controller and its steps. The controller and steps can be functions in your 
 | 
			
		||||
    python code, or ClearML [tasks](../fundamentals/task.md).
 | 
			
		||||
  * [PipelineDecorator](../pipelines/pipelines_sdk_function_decorators.md) - A set 
 | 
			
		||||
 | 
			
		||||
@ -170,7 +170,7 @@ If the `secure.conf` file does not exist, create your own in ClearML Server's `/
 | 
			
		||||
an alternate folder you configured), and input the modified configuration
 | 
			
		||||
:::
 | 
			
		||||
 | 
			
		||||
The default secret for the system's apiserver component can be overridden by setting the following environment variable: 
 | 
			
		||||
You can override the default secret for the system's `apiserver` component by setting the following environment variable: 
 | 
			
		||||
`CLEARML__SECURE__CREDENTIALS__APISERVER__USER_SECRET="my-new-secret"`
 | 
			
		||||
 | 
			
		||||
:::note
 | 
			
		||||
 | 
			
		||||
@ -21,7 +21,7 @@ During early stages of model development, while code is still being modified hea
 | 
			
		||||
  - **Workstation with a GPU**, usually with a limited amount of memory for small batch-sizes. Use this workstation to train 
 | 
			
		||||
    the model and ensure that you choose a model that makes sense, and the training procedure works. Can be used to provide initial models for testing. 
 | 
			
		||||
 | 
			
		||||
The abovementioned setups might be folded into each other and that's great! If you have a GPU machine for each researcher, that's awesome! 
 | 
			
		||||
These setups can be folded into each other and that's great! If you have a GPU machine for each researcher, that's awesome! 
 | 
			
		||||
The goal of this phase is to get a code, dataset, and environment set up, so you can start digging to find the best model!
 | 
			
		||||
 | 
			
		||||
- [ClearML SDK](../../clearml_sdk/clearml_sdk.md) should be integrated into your code (check out [Getting Started](ds_first_steps.md)). 
 | 
			
		||||
 | 
			
		||||
@ -6,7 +6,7 @@ title: First Steps
 | 
			
		||||
## Install ClearML
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
First, [sign up for free](https://app.clear.ml)
 | 
			
		||||
First, [sign up for free](https://app.clear.ml).
 | 
			
		||||
 | 
			
		||||
Install the `clearml` python package:
 | 
			
		||||
```bash
 | 
			
		||||
 | 
			
		||||
@ -46,7 +46,7 @@ We can change the task’s name by clicking it here, and add a description or ge
 | 
			
		||||
 | 
			
		||||
First of all, source code is captured. If you’re working in a git repository we’ll save your git information along with any uncommitted changes. If you’re running an unversioned script, `clearml` will save the script instead.
 | 
			
		||||
 | 
			
		||||
Together with the python packages your coded uses, this’ll allow you to recreate your experiment on any machine.
 | 
			
		||||
Together with the Python packages your code uses, this will allow you to recreate your experiment on any machine.
 | 
			
		||||
 | 
			
		||||
Similarly, all of the output the code produces will also be captured.
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
@ -58,7 +58,7 @@ to open the app's instance launch form.
 | 
			
		||||
  * **Base Docker Image** (optional) - Available when `Use docker mode` is selected: Default Docker image in which the 
 | 
			
		||||
  ClearML Agent will run. Provide an image stored in a Docker artifactory so instances can automatically fetch it
 | 
			
		||||
* **Compute Resources**
 | 
			
		||||
    * Resource Name - Assign a name to the resource type. This name will appear in the Autoscaler dashboard
 | 
			
		||||
    * Resource Name - Assign a name to the resource type. This name will appear in the autoscaler dashboard
 | 
			
		||||
    * EC2 Instance Type - See [Instance Types](https://aws.amazon.com/ec2/instance-types) for full list of types
 | 
			
		||||
    * Run in CPU mode - Check box to run with CPU only
 | 
			
		||||
    * Use Spot Instance - Select to use a spot instance. Otherwise, a reserved instance is used.
 | 
			
		||||
@ -98,7 +98,7 @@ to open the app's instance launch form.
 | 
			
		||||
      instance. Read more [here](https://docs.aws.amazon.com/vpc/latest/userguide/VPC_SecurityGroups.html) 
 | 
			
		||||
    * VPC Subnet ID - The subnet ID For the created instance. If more than one ID is provided, the instance will be started in the first available subnet. For more information, see [What is Amazon VPC?](https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html)
 | 
			
		||||
    * \+ Add Item - Define another resource type
 | 
			
		||||
* **IAM Instance Profile** (optional) - Set an IAM instance profile for all instances spun by the Autoscaler 
 | 
			
		||||
* **IAM Instance Profile** (optional) - Set an IAM instance profile for all instances spun by the autoscaler 
 | 
			
		||||
    * Arn - Amazon Resource Name specifying the instance profile
 | 
			
		||||
    * Name - Name identifying the instance profile
 | 
			
		||||
* **Autoscaler Instance Name** (optional) - Name for the Autoscaler instance. This will appear in the instance list
 | 
			
		||||
@ -129,7 +129,7 @@ The Configuration Vault is available under the ClearML Enterprise plan.
 | 
			
		||||
 | 
			
		||||
You can utilize the [configuration vault](../settings/webapp_settings_profile.md#configuration-vault) to set the following: 
 | 
			
		||||
* `aws_region`
 | 
			
		||||
* `aws_credentials_key_id` and `aws_secret_access_key` - AWS credentials for the Autoscaler
 | 
			
		||||
* `aws_credentials_key_id` and `aws_secret_access_key` - AWS credentials for the autoscaler
 | 
			
		||||
* `extra_vm_bash_script` - A bash script to execute after launching the EC2 instance. This script will be appended to
 | 
			
		||||
the one set in the `Init script` field of the instance launch form
 | 
			
		||||
* `extra_clearml_conf` - ClearML configuration to use by the ClearML Agent when executing your experiments. This 
 | 
			
		||||
@ -202,7 +202,7 @@ auto_scaler.v1.aws {
 | 
			
		||||
#### Configure Instances Spawned by the Autoscaler
 | 
			
		||||
To configure instances spawned by the autoscaler, do any of the following:
 | 
			
		||||
* Add the configuration in the `auto_scaler.v1.aws.extra_clearml_conf` field of the configuration vault
 | 
			
		||||
* Run the Autoscaler using a [ClearML Service Account](../settings/webapp_settings_users.md#service-accounts). Add the 
 | 
			
		||||
* Run the autoscaler using a [ClearML Service Account](../settings/webapp_settings_users.md#service-accounts). Add the 
 | 
			
		||||
configuration to the service account's configuration vault, and set the autoscaler to run under that account
 | 
			
		||||
in the `Run with Service Account` field
 | 
			
		||||
* Admins can add the configuration to a [ClearML Administrator Vault](../settings/webapp_settings_admin_vaults.md)
 | 
			
		||||
 | 
			
		||||
@ -58,7 +58,7 @@ to open the app's instance launch form.
 | 
			
		||||
* **Base Docker Image** (optional) - Available when `Use docker mode` is selected. Default Docker image in which the ClearML Agent will run. Provide an image stored in a 
 | 
			
		||||
  Docker artifactory so VM instances can automatically fetch it
 | 
			
		||||
* **Compute Resources**
 | 
			
		||||
    * Resource Name - Assign a name to the resource type. This name will appear in the Autoscaler dashboard
 | 
			
		||||
    * Resource Name - Assign a name to the resource type. This name will appear in the autoscaler dashboard
 | 
			
		||||
    * GCP Machine Type - See list of [machine types](https://cloud.google.com/compute/docs/machine-types)
 | 
			
		||||
    * Run in CPU mode - Select to have the autoscaler utilize only CPU VM instances
 | 
			
		||||
    * GPU Type - See list of [supported GPUs by instance](https://cloud.google.com/compute/docs/gpus)
 | 
			
		||||
@ -106,7 +106,7 @@ to open the app's instance launch form.
 | 
			
		||||
 | 
			
		||||
:::important Enterprise Feature
 | 
			
		||||
You can utilize the [configuration vault](../settings/webapp_settings_profile.md#configuration-vault) to configure GCP 
 | 
			
		||||
credentials for the Autoscaler in the following format: 
 | 
			
		||||
credentials for the autoscaler in the following format: 
 | 
			
		||||
 | 
			
		||||
```
 | 
			
		||||
auto_scaler.v1 {
 | 
			
		||||
 | 
			
		||||
		Loading…
	
		Reference in New Issue
	
	Block a user