From e4cde447aae0168e43b1b07db2af05cda5f89e7f Mon Sep 17 00:00:00 2001 From: pollfly <75068813+pollfly@users.noreply.github.com> Date: Tue, 12 Mar 2024 11:25:00 +0200 Subject: [PATCH] Small edits (#797) --- docs/apps/clearml_param_search.md | 2 +- .../data_management_examples/data_man_folder_sync.md | 4 ++-- .../data_management_examples/data_man_simple.md | 2 +- docs/getting_started/mlops/mlops_second_steps.md | 10 +++++++--- docs/webapp/webapp_exp_comparing.md | 2 +- 5 files changed, 12 insertions(+), 8 deletions(-) diff --git a/docs/apps/clearml_param_search.md b/docs/apps/clearml_param_search.md index d73381e0..f2123eb9 100644 --- a/docs/apps/clearml_param_search.md +++ b/docs/apps/clearml_param_search.md @@ -45,7 +45,7 @@ of the optimization results in table and graph forms. |`--args`| List of `=` strings to pass to the remote execution. Currently only argparse/click/hydra/fire arguments are supported. Example: `--args lr=0.003 batch_size=64`|Yes| |`--local`| If set, run the experiments locally. Notice that no new python environment will be created. The `--script` parameter must point to a local file entry point and all arguments must be passed with `--args`| Yes| |`--save-top-k-tasks-only`| Keep only the top \ performing tasks, and archive the rest of the experiments. Input `-1` to keep all tasks. Default: `10`.|Yes| -|`--time-limit-per-job`|Maximum execution time per single job in minutes. When time limit is exceeded, the job is aborted. Default: no time limit.|Yes| +|`--time-limit-per-job`|Maximum execution time per single job in minutes. When the time limit is exceeded, the job is aborted. Default: no time limit.|Yes| diff --git a/docs/clearml_data/data_management_examples/data_man_folder_sync.md b/docs/clearml_data/data_management_examples/data_man_folder_sync.md index caf4642d..645f8926 100644 --- a/docs/clearml_data/data_management_examples/data_man_folder_sync.md +++ b/docs/clearml_data/data_management_examples/data_man_folder_sync.md @@ -48,13 +48,13 @@ As can be seen, the `clearml-data sync` command creates the dataset, then upload ## Modifying Synced Folder -Now we'll modify the folder: +Modify the data folder: 1. Add another line to one of the files in the `data_samples` folder. 1. Add a file to the sample_data folder.
Run `echo "data data data" > data_samples/new_data.txt` (this will create the file `new_data.txt` and put it in the `data_samples` folder) -We'll repeat the process of creating a new dataset with the previous one as its parent, and syncing the folder. +Repeat the process of creating a new dataset with the previous one as its parent, and syncing the folder. ```bash clearml-data sync --project datasets --name second_ds --parents a1ddc8b0711b4178828f6c6e6e994b7c --folder data_samples diff --git a/docs/clearml_data/data_management_examples/data_man_simple.md b/docs/clearml_data/data_management_examples/data_man_simple.md index d71a9fbf..328d4dd8 100644 --- a/docs/clearml_data/data_management_examples/data_man_simple.md +++ b/docs/clearml_data/data_management_examples/data_man_simple.md @@ -126,7 +126,7 @@ You'll need to input the Dataset ID you received when created the dataset above 1 file added ``` -1. Remove a file. We'll need to specify the file's full path (within the dataset, not locally) to remove it. +1. Remove a file. You need to specify the file's full path (within the dataset, not locally) to remove it. ```bash clearml-data remove --files data_samples/dancing.jpg diff --git a/docs/getting_started/mlops/mlops_second_steps.md b/docs/getting_started/mlops/mlops_second_steps.md index 83e9c212..ca6a3c61 100644 --- a/docs/getting_started/mlops/mlops_second_steps.md +++ b/docs/getting_started/mlops/mlops_second_steps.md @@ -31,6 +31,8 @@ The second step is to preprocess the data. First access the data, then modify it and lastly create a new version of the data. ```python +from clearml import Task, Dataset + # create a task for the data processing part task = Task.init(project_name='data', task_name='create', task_type='data_processing') @@ -93,6 +95,8 @@ will first run the first and then run the second. It is important to remember that pipelines are Tasks by themselves and can also be automated by other pipelines (i.e. pipelines of pipelines). ```python +from clearml import PipelineController + pipe = PipelineController( project='data', name='pipeline demo', @@ -112,6 +116,6 @@ pipe.add_step( ) ``` -We can also pass the parameters from one step to the other (for example `Task.id`). -In addition to pipelines made up of Task steps, ClearML also supports pipelines consisting of function steps. See more in the -full pipeline documentation [here](../../pipelines/pipelines.md). +You can also pass the parameters from one step to the other (for example `Task.id`). +In addition to pipelines made up of Task steps, ClearML also supports pipelines consisting of function steps. For more +information, see the [full pipeline documentation](../../pipelines/pipelines.md). diff --git a/docs/webapp/webapp_exp_comparing.md b/docs/webapp/webapp_exp_comparing.md index 970fea0f..1e86e3f5 100644 --- a/docs/webapp/webapp_exp_comparing.md +++ b/docs/webapp/webapp_exp_comparing.md @@ -105,7 +105,7 @@ Switch on the **Show row extremes** toggle to highlight each variant's maximum a The **Hyperparameters** tab's **Parallel Coordinates** comparison shows experiments' hyperparameter impact on specified metrics: -1. Under **Performance Metrics**, select a metrics to compare for +1. Under **Performance Metrics**, select metrics to compare for 1. Select the values to use for each metric in the plot (can select multiple): * LAST - The final value, or the most recent value, for currently running experiments * MIN - Minimal value