diff --git a/docs/clearml_sdk/clearml_sdk.md b/docs/clearml_sdk/clearml_sdk.md index d706a671..bb17e631 100644 --- a/docs/clearml_sdk/clearml_sdk.md +++ b/docs/clearml_sdk/clearml_sdk.md @@ -65,7 +65,7 @@ See the [HyperParameterOptimizer SDK reference page](../references/sdk/hpo_optim ### Pipeline ClearML's `automation` module includes classes that support creating pipelines: - * [PipelineController](../pipelines/pipelines_sdk_tasks.md) - A pythonic interface for + * [PipelineController](../pipelines/pipelines_sdk_tasks.md) - A Pythonic interface for defining and configuring a pipeline controller and its steps. The controller and steps can be functions in your python code, or ClearML [tasks](../fundamentals/task.md). * [PipelineDecorator](../pipelines/pipelines_sdk_function_decorators.md) - A set diff --git a/docs/deploying_clearml/clearml_server_config.md b/docs/deploying_clearml/clearml_server_config.md index ee7b7ae9..4b3f8036 100644 --- a/docs/deploying_clearml/clearml_server_config.md +++ b/docs/deploying_clearml/clearml_server_config.md @@ -170,7 +170,7 @@ If the `secure.conf` file does not exist, create your own in ClearML Server's `/ an alternate folder you configured), and input the modified configuration ::: -The default secret for the system's apiserver component can be overridden by setting the following environment variable: +You can override the default secret for the system's `apiserver` component by setting the following environment variable: `CLEARML__SECURE__CREDENTIALS__APISERVER__USER_SECRET="my-new-secret"` :::note diff --git a/docs/getting_started/ds/best_practices.md b/docs/getting_started/ds/best_practices.md index 347b8764..f7858714 100644 --- a/docs/getting_started/ds/best_practices.md +++ b/docs/getting_started/ds/best_practices.md @@ -21,7 +21,7 @@ During early stages of model development, while code is still being modified hea - **Workstation with a GPU**, usually with a limited amount of memory for small batch-sizes. Use this workstation to train the model and ensure that you choose a model that makes sense, and the training procedure works. Can be used to provide initial models for testing. -The abovementioned setups might be folded into each other and that's great! If you have a GPU machine for each researcher, that's awesome! +These setups can be folded into each other and that's great! If you have a GPU machine for each researcher, that's awesome! The goal of this phase is to get a code, dataset, and environment set up, so you can start digging to find the best model! - [ClearML SDK](../../clearml_sdk/clearml_sdk.md) should be integrated into your code (check out [Getting Started](ds_first_steps.md)). diff --git a/docs/getting_started/ds/ds_first_steps.md b/docs/getting_started/ds/ds_first_steps.md index 62d2ae52..702461de 100644 --- a/docs/getting_started/ds/ds_first_steps.md +++ b/docs/getting_started/ds/ds_first_steps.md @@ -6,7 +6,7 @@ title: First Steps ## Install ClearML -First, [sign up for free](https://app.clear.ml) +First, [sign up for free](https://app.clear.ml). Install the `clearml` python package: ```bash diff --git a/docs/getting_started/video_tutorials/experiment_manager_hands-on.md b/docs/getting_started/video_tutorials/experiment_manager_hands-on.md index 1b892bfb..83839d0e 100644 --- a/docs/getting_started/video_tutorials/experiment_manager_hands-on.md +++ b/docs/getting_started/video_tutorials/experiment_manager_hands-on.md @@ -46,7 +46,7 @@ We can change the task’s name by clicking it here, and add a description or ge First of all, source code is captured. If you’re working in a git repository we’ll save your git information along with any uncommitted changes. If you’re running an unversioned script, `clearml` will save the script instead. -Together with the python packages your coded uses, this’ll allow you to recreate your experiment on any machine. +Together with the Python packages your code uses, this will allow you to recreate your experiment on any machine. Similarly, all of the output the code produces will also be captured. diff --git a/docs/webapp/applications/apps_aws_autoscaler.md b/docs/webapp/applications/apps_aws_autoscaler.md index 37c1288e..30bd3864 100644 --- a/docs/webapp/applications/apps_aws_autoscaler.md +++ b/docs/webapp/applications/apps_aws_autoscaler.md @@ -58,7 +58,7 @@ to open the app's instance launch form. * **Base Docker Image** (optional) - Available when `Use docker mode` is selected: Default Docker image in which the ClearML Agent will run. Provide an image stored in a Docker artifactory so instances can automatically fetch it * **Compute Resources** - * Resource Name - Assign a name to the resource type. This name will appear in the Autoscaler dashboard + * Resource Name - Assign a name to the resource type. This name will appear in the autoscaler dashboard * EC2 Instance Type - See [Instance Types](https://aws.amazon.com/ec2/instance-types) for full list of types * Run in CPU mode - Check box to run with CPU only * Use Spot Instance - Select to use a spot instance. Otherwise, a reserved instance is used. @@ -98,7 +98,7 @@ to open the app's instance launch form. instance. Read more [here](https://docs.aws.amazon.com/vpc/latest/userguide/VPC_SecurityGroups.html) * VPC Subnet ID - The subnet ID For the created instance. If more than one ID is provided, the instance will be started in the first available subnet. For more information, see [What is Amazon VPC?](https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html) * \+ Add Item - Define another resource type -* **IAM Instance Profile** (optional) - Set an IAM instance profile for all instances spun by the Autoscaler +* **IAM Instance Profile** (optional) - Set an IAM instance profile for all instances spun by the autoscaler * Arn - Amazon Resource Name specifying the instance profile * Name - Name identifying the instance profile * **Autoscaler Instance Name** (optional) - Name for the Autoscaler instance. This will appear in the instance list @@ -129,7 +129,7 @@ The Configuration Vault is available under the ClearML Enterprise plan. You can utilize the [configuration vault](../settings/webapp_settings_profile.md#configuration-vault) to set the following: * `aws_region` -* `aws_credentials_key_id` and `aws_secret_access_key` - AWS credentials for the Autoscaler +* `aws_credentials_key_id` and `aws_secret_access_key` - AWS credentials for the autoscaler * `extra_vm_bash_script` - A bash script to execute after launching the EC2 instance. This script will be appended to the one set in the `Init script` field of the instance launch form * `extra_clearml_conf` - ClearML configuration to use by the ClearML Agent when executing your experiments. This @@ -202,7 +202,7 @@ auto_scaler.v1.aws { #### Configure Instances Spawned by the Autoscaler To configure instances spawned by the autoscaler, do any of the following: * Add the configuration in the `auto_scaler.v1.aws.extra_clearml_conf` field of the configuration vault -* Run the Autoscaler using a [ClearML Service Account](../settings/webapp_settings_users.md#service-accounts). Add the +* Run the autoscaler using a [ClearML Service Account](../settings/webapp_settings_users.md#service-accounts). Add the configuration to the service account's configuration vault, and set the autoscaler to run under that account in the `Run with Service Account` field * Admins can add the configuration to a [ClearML Administrator Vault](../settings/webapp_settings_admin_vaults.md) diff --git a/docs/webapp/applications/apps_gcp_autoscaler.md b/docs/webapp/applications/apps_gcp_autoscaler.md index b3009e84..2d9d2dd6 100644 --- a/docs/webapp/applications/apps_gcp_autoscaler.md +++ b/docs/webapp/applications/apps_gcp_autoscaler.md @@ -58,7 +58,7 @@ to open the app's instance launch form. * **Base Docker Image** (optional) - Available when `Use docker mode` is selected. Default Docker image in which the ClearML Agent will run. Provide an image stored in a Docker artifactory so VM instances can automatically fetch it * **Compute Resources** - * Resource Name - Assign a name to the resource type. This name will appear in the Autoscaler dashboard + * Resource Name - Assign a name to the resource type. This name will appear in the autoscaler dashboard * GCP Machine Type - See list of [machine types](https://cloud.google.com/compute/docs/machine-types) * Run in CPU mode - Select to have the autoscaler utilize only CPU VM instances * GPU Type - See list of [supported GPUs by instance](https://cloud.google.com/compute/docs/gpus) @@ -106,7 +106,7 @@ to open the app's instance launch form. :::important Enterprise Feature You can utilize the [configuration vault](../settings/webapp_settings_profile.md#configuration-vault) to configure GCP -credentials for the Autoscaler in the following format: +credentials for the autoscaler in the following format: ``` auto_scaler.v1 {