mirror of
				https://github.com/clearml/clearml-docs
				synced 2025-06-26 18:17:44 +00:00 
			
		
		
		
	Small edits (#455)
This commit is contained in:
		
							parent
							
								
									18e3e7abe2
								
							
						
					
					
						commit
						61f822e613
					
				| @ -80,7 +80,7 @@ The following are the parameter type options and their corresponding fields: | ||||
|     - `"values": List[Any]`- A list of valid parameter values to sample from | ||||
| 
 | ||||
| For example: to specify a parameter search over uniform ranges of layer_1 and layer_2 sizes between 128 and 512  | ||||
| (in jumps of 128)  with varying batch sizes of 96, 128, and 160, use the following command: | ||||
| (in jumps of 128) with varying batch sizes of 96, 128, and 160, use the following command: | ||||
| 
 | ||||
| <div className="wb-normal"> | ||||
| 
 | ||||
|  | ||||
| @ -483,7 +483,7 @@ Self-hosted [ClearML Server](deploying_clearml/clearml_server.md) comes by defau | ||||
| By default, the server is open and does not require username and password, but it can be [password-protected](deploying_clearml/clearml_server_security.md#user-access-security). | ||||
| In case it is password-protected, the services agent will need to be configured with server credentials (associated with a user). | ||||
| 
 | ||||
| To do that, set these environment variables on  the ClearML Server machine with the appropriate credentials: | ||||
| To do that, set these environment variables on the ClearML Server machine with the appropriate credentials: | ||||
| ``` | ||||
| CLEARML_API_ACCESS_KEY | ||||
| CLEARML_API_SECRET_KEY | ||||
| @ -499,7 +499,7 @@ Build a Docker container that when launched executes a specific experiment, or a | ||||
|   ```bash | ||||
|   clearml-agent build --id <task-id> --docker --target <new-docker-name> --entry-point reuse_task | ||||
|   ``` | ||||
| - Build a Docker container that at launch will clone a Task  specified by Task ID, and will execute the newly cloned Task. | ||||
| - Build a Docker container that at launch will clone a Task specified by Task ID, and will execute the newly cloned Task. | ||||
|   ```bash | ||||
|   clearml-agent build --id <task-id> --docker --target <new-docker-name> --entry-point clone_task | ||||
|   ``` | ||||
|  | ||||
| @ -37,7 +37,7 @@ clearml-serving --id <service_id> model add --engine sklearn --endpoint "test_mo | ||||
| ``` | ||||
| 
 | ||||
| :::info Service ID | ||||
| Make sure that you have executed `clearml-servings`'s | ||||
| Make sure that you have executed `clearml-serving`'s | ||||
| [initial setup](clearml_serving.md#initial-setup), in which you create a Serving Service.  | ||||
| The Serving Service's ID is required to register a model, and to execute `clearml-serving`'s `metrics` and `config` commands  | ||||
| ::: | ||||
| @ -92,7 +92,7 @@ or with the `clearml-serving` CLI. | ||||
|    ``` | ||||
|      | ||||
|   You now have a new Model named `manual sklearn model` in the `serving examples` project. The CLI output prints  | ||||
|   the UID of the new model, which you will use it to register a new endpoint.  | ||||
|   the UID of the new model, which you will use to register a new endpoint.  | ||||
| 
 | ||||
|   In the [ClearML web UI](../webapp/webapp_overview.md), the new model is listed under the **Models** tab of its project.  | ||||
|   You can also download the model file itself directly from the web UI.  | ||||
| @ -105,7 +105,7 @@ or with the `clearml-serving` CLI. | ||||
| :::info Model Storage | ||||
| You can also provide a different storage destination for the model, such as S3/GS/Azure, by passing | ||||
| `--destination="s3://bucket/folder"`, `gs://bucket/folder`, `azure://bucket/folder`. There is no need to provide a unique  | ||||
| path tp the destination argument, the location of the model will be a unique path based on the serving service ID and the  | ||||
| path to the destination argument, the location of the model will be a unique path based on the serving service ID and the  | ||||
| model name | ||||
| ::: | ||||
| 
 | ||||
| @ -116,7 +116,7 @@ model name | ||||
| The ClearML Serving Service supports automatic model deployment and upgrades, which is connected with the model  | ||||
| repository and API. When the model auto-deploy is configured, new model versions will be automatically deployed when you  | ||||
| `publish` or `tag` a new model in the ClearML model repository. This automation interface allows for simpler CI/CD model  | ||||
| deployment process, as a single API automatically deploy (or remove) a model from the Serving Service. | ||||
| deployment process, as a single API automatically deploys (or removes) a model from the Serving Service. | ||||
| 
 | ||||
| #### Automatic Model Deployment Example | ||||
| 
 | ||||
| @ -142,7 +142,7 @@ deployment process, as a single API automatically deploy (or remove) a model fro | ||||
| 
 | ||||
| ### Canary Endpoint Setup | ||||
| 
 | ||||
| Canary endpoint deployment add a new endpoint where the actual request is sent to a preconfigured set of endpoints with  | ||||
| Canary endpoint deployment adds a new endpoint where the actual request is sent to a preconfigured set of endpoints with  | ||||
| pre-provided distribution. For example, let's create a new endpoint "test_model_sklearn_canary", you can provide a list  | ||||
| of endpoints and probabilities (weights). | ||||
| 
 | ||||
| @ -195,13 +195,13 @@ Example: | ||||
| ClearML serving instances send serving statistics (count/latency) automatically to Prometheus and Grafana can be used  | ||||
| to visualize and create live dashboards.  | ||||
| 
 | ||||
| The default docker-compose installation is preconfigured with Prometheus and Grafana, do notice that by default data/ate  | ||||
| The default docker-compose installation is preconfigured with Prometheus and Grafana. Notice that by default data/ate  | ||||
| of both containers is *not* persistent. To add persistence, we recommend adding a volume mount. | ||||
| 
 | ||||
| You can also add many custom metrics on the input/predictions of your models. Once a model endpoint is registered,  | ||||
| adding custom metric can be done using the CLI. | ||||
| 
 | ||||
| For example, assume the mock scikit-learn model is deployed on endpoint `test_model_sklearn`,  you can log the requests  | ||||
| For example, assume the mock scikit-learn model is deployed on endpoint `test_model_sklearn`, you can log the requests  | ||||
| inputs and outputs (see examples/sklearn/preprocess.py example): | ||||
| 
 | ||||
| ```bash | ||||
|  | ||||
| @ -55,7 +55,7 @@ help maintainers reproduce the problem: | ||||
| * **Describe the exact steps necessary to reproduce the problem** in as much detail as possible. Please do not just summarize what you did. Make sure to explain how you did it. | ||||
| * **Provide the specific environment setup.** Include the ``pip freeze`` output, specific environment variables, Python version, and other relevant information. | ||||
| * **Provide specific examples to demonstrate the steps.** Include links to files or GitHub projects, or copy / paste snippets which you use in those examples. | ||||
| * **If you are reporting any  ClearML crash,** include a crash report with a stack trace from the operating system. Make  | ||||
| * **If you are reporting any ClearML crash,** include a crash report with a stack trace from the operating system. Make  | ||||
|   sure to add the crash report in the issue and place it in a [code block](https://docs.github.com/en/github/writing-on-github/working-with-advanced-formatting/creating-and-highlighting-code-blocks),  | ||||
|   a [file attachment](https://help.github.com/articles/file-attachments-on-issues-and-pull-requests), or just put it in  | ||||
|   a [gist](https://gist.github.com) (and provide a link to that gist). | ||||
|  | ||||
| @ -73,7 +73,7 @@ The minimum requirements for ClearML Server are: | ||||
| 
 | ||||
| ## Restarting | ||||
| 
 | ||||
| **To restart  ClearML Server Docker deployment:** | ||||
| **To restart ClearML Server Docker deployment:** | ||||
| 
 | ||||
| * Stop and then restart the Docker containers by executing the following commands: | ||||
| 
 | ||||
|  | ||||
| @ -72,6 +72,6 @@ the models associated with a project are listed. | ||||
| ## SDK Interface | ||||
| 
 | ||||
| See [the Models SDK interface](../clearml_sdk/model_sdk.md) for an overview for using the most basic Pythonic methods of the model  | ||||
| classes. See a detailed list of all available methods  in the [Model](../references/sdk/model_model.md), [OutputModel](../references/sdk/model_outputmodel.md), and [InputModel](../references/sdk/model_inputmodel.md)  | ||||
| classes. See a detailed list of all available methods in the [Model](../references/sdk/model_model.md), [OutputModel](../references/sdk/model_outputmodel.md), and [InputModel](../references/sdk/model_inputmodel.md)  | ||||
| reference pages.  | ||||
| 
 | ||||
|  | ||||
| @ -32,7 +32,7 @@ pip install clearml | ||||
|    Please create new clearml credentials through the settings page in your `clearml-server` web app,  | ||||
|    or create a free account at https://app.clear.ml/settings/webapp-configuration | ||||
|      | ||||
|    In the settings > workspace  page, press "Create new credentials", then press "Copy to clipboard". | ||||
|    In the settings > workspace page, press "Create new credentials", then press "Copy to clipboard". | ||||
|    Paste copied configuration here:  | ||||
|    ``` | ||||
|        | ||||
|  | ||||
| @ -40,7 +40,7 @@ Check [this](../../fundamentals/hyperparameters.md) out for all Hyperparameter l | ||||
| 
 | ||||
| ## Log Artifacts | ||||
| 
 | ||||
| ClearML allows you to easily store the output products of an experiment -  Model snapshot / weights file, a preprocessing of your data, feature representation of data and more! | ||||
| ClearML allows you to easily store the output products of an experiment - Model snapshot / weights file, a preprocessing of your data, feature representation of data and more! | ||||
| 
 | ||||
| Essentially, artifacts are files (or python objects) uploaded from a script and are stored alongside the Task. | ||||
| These Artifacts can be easily accessed by the web UI or programmatically. | ||||
| @ -157,7 +157,7 @@ The experiment table is a powerful tool for creating dashboards and views of you | ||||
| 
 | ||||
| ### Creating Leaderboards | ||||
| Customize the [experiments table](../../webapp/webapp_exp_table.md) to fit your own needs, adding desired views of parameters, metrics and tags. | ||||
| It's possible to  filter and sort based on parameters and metrics, so creating custom views is simple and flexible. | ||||
| It's possible to filter and sort based on parameters and metrics, so creating custom views is simple and flexible. | ||||
| 
 | ||||
| Create a dashboard for a project, presenting the latest Models and their accuracy scores, for immediate insights. | ||||
| 
 | ||||
|  | ||||
| @ -45,7 +45,7 @@ The structure of your pipeline will be derived from looking at this `parents` ar | ||||
| 
 | ||||
| Now we do the same for the final step. However, remember the empty hyperparameters we saw before? We still have to overwrite these. We can use the `parameter_override` argument to do just that. | ||||
| 
 | ||||
| For example, we can tell the first step to use the global pipeline parameter raw data url like so. But we can also reference output artifacts from a previous step by using its name and we can of course also just overwrite a parameter with a normal value. Finally, we can even pass along the unique task ID of a previous step, so you can get the task object based on that ID and access anything and everything within that task. | ||||
| For example, we can tell the first step to use the global pipeline parameter raw data url like so. But we can also reference output artifacts from a previous step by using its name, and we can of course also just overwrite a parameter with a normal value. Finally, we can even pass along the unique task ID of a previous step, so you can get the task object based on that ID and access anything and everything within that task. | ||||
| 
 | ||||
| And that’s it! We now have our first pipeline! | ||||
| 
 | ||||
|  | ||||
| @ -29,7 +29,7 @@ it is commented out, make sure to uncomment the line. We will use the example sc | ||||
| 1. Search for and go to `docker_force_pull` in the document, and make sure that it is set to `true`, so that your docker  | ||||
|    image will be updated.  | ||||
|     | ||||
| 1. Run the `clearml-agent` in docker mode: `clearml-agent daemon --docker --queue default`. The agent  will use the default  | ||||
| 1. Run the `clearml-agent` in docker mode: `clearml-agent daemon --docker --queue default`. The agent will use the default  | ||||
|    Cuda/Nvidia Docker Image.  | ||||
| 
 | ||||
| 1. Enqueue any ClearML Task to the `default` queue, which the Agent is now listening to. The Agent pulls the Task, and then reproduces it,  | ||||
|  | ||||
| @ -6,7 +6,7 @@ The [cifar_ignite.py](https://github.com/allegroai/clearml/blob/master/examples/ | ||||
| script integrates ClearML into code that uses [PyTorch Ignite](https://github.com/pytorch/ignite).  | ||||
| 
 | ||||
| The example script does the following: | ||||
| * Trains a  neural network on the CIFAR10 dataset for image classification. | ||||
| * Trains a neural network on the CIFAR10 dataset for image classification. | ||||
| * Creates a [ClearML Task](../../../fundamentals/task.md) named `image classification CIFAR10`, which is associated with  | ||||
|   the `examples` project. | ||||
| * Calls the [`Task.connect`](../../../references/sdk/task.md#connect) method to track experiment configuration. | ||||
|  | ||||
| @ -68,7 +68,7 @@ The sections below describe in more detail what happens in the controller task a | ||||
|    Custom configuration values specific to this step execution are defined through the `parameter_override` parameter,  | ||||
|    where the first step’s artifact is fed into the second step. | ||||
| 
 | ||||
|    Special pre-execution and post-execution logic is added for this step  through the use of `pre_execute_callback`  | ||||
|    Special pre-execution and post-execution logic is added for this step through the use of `pre_execute_callback`  | ||||
|    and  `post_execute_callback` respectively.  | ||||
| 
 | ||||
|    ```python | ||||
|  | ||||
| @ -35,7 +35,7 @@ logged as required packages for the pipeline execution step. | ||||
|    ``` | ||||
|      | ||||
| 1. Set the default execution queue to be used. All the pipeline steps will be enqueued for execution in this queue  | ||||
|    (unless overridden by the `execution_queue` parameter of  the `add_function_step` method). | ||||
|    (unless overridden by the `execution_queue` parameter of the `add_function_step` method). | ||||
|      | ||||
|   ```python | ||||
|   pipe.set_default_execution_queue('default') | ||||
|  | ||||
| @ -18,8 +18,8 @@ to your needs, and enqueue it for execution directly from the ClearML UI. | ||||
| Configure the task execution by modifying the `args` dictionary: | ||||
| * `delete_threshold_days` - Tasks older than this number of days will be deleted. The default value is 30 days. | ||||
| * `cleanup_period_in_days` - Repeat the cleanup service at this interval, in days. The default value is 1.0 (run once a day). | ||||
| * `force_delete` - If `False` (default), delete only Draft tasks. If `True`,  allows deletion of  tasks in any status.  | ||||
| * `run_as_service` - If `True` (default),  the task will be enqueued for remote execution (default queue: "services"). Otherwise, the script will execute locally.  | ||||
| * `force_delete` - If `False` (default), delete only Draft tasks. If `True`, allows deletion of tasks in any status.  | ||||
| * `run_as_service` - If `True` (default), the task will be enqueued for remote execution (default queue: "services"). Otherwise, the script will execute locally.  | ||||
| 
 | ||||
| :::note Remote Execution | ||||
| If `run_as_service` is set to `True`, make sure a `clearml-agent` is assigned to the `services` queue. | ||||
| @ -48,7 +48,7 @@ This is followed by details from the cleanup. | ||||
| an `APIClient` object that establishes a session with the ClearML Server, and accomplishes the cleanup by calling: | ||||
| * [`Tasks.get_all`](../../references/api/tasks.md#post-tasksget_all) to get a list of Tasks to delete, providing the following parameters: | ||||
|     * `system_tags` - Get only Tasks tagged as `archived`. | ||||
|     * `status_changed` - Get Tasks whose last status change is older than then delete threshold (in seconds). | ||||
|     * `status_changed` - Get Tasks whose last status change is older than the delete threshold (in seconds). | ||||
| * [`Task.delete`](../../references/sdk/task.md#delete) - Delete a Task.   | ||||
| 
 | ||||
| ## Configuration | ||||
|  | ||||
| @ -124,7 +124,7 @@ Dataset.delete( | ||||
| ``` | ||||
| 
 | ||||
| This supports deleting sources located in AWS S3, GCP, and Azure Storage (not local storage). The `delete_sources`  | ||||
| parameter is ignored if `delete_all_versions` is `False`.  You can view the deletion process’ progress by passing  | ||||
| parameter is ignored if `delete_all_versions` is `False`. You can view the deletion process’ progress by passing  | ||||
| `show_progress=True` (`tqdm` required).   | ||||
| 
 | ||||
| ### Tagging Datasets | ||||
| @ -147,7 +147,7 @@ MyDataset.remove_tags(["dogs"]) | ||||
| Dataset versioning refers to the group of ClearML Enterprise SDK and WebApp (UI) features for creating,  | ||||
| modifying, and deleting Dataset versions. | ||||
| 
 | ||||
| ClearML Enterprise supports simple and advanced Dataset versioning paradigms.  A **simple version structure** consists of  | ||||
| ClearML Enterprise supports simple and advanced Dataset versioning paradigms. A **simple version structure** consists of  | ||||
| a single evolving version, with historic static snapshots. Continuously push your changes to your single dataset version,  | ||||
| and take a snapshot to record the content of your dataset at a specific point in time. | ||||
| 
 | ||||
|  | ||||
| @ -26,7 +26,7 @@ a SingleFrame: | ||||
|   * Metadata and data for the labeled area of an image | ||||
|    | ||||
|    | ||||
| See [Example 1](#example-1), which shows `masks` in `sources`, `mask` in `rois`,  and the key-value pairs used to relate  | ||||
| See [Example 1](#example-1), which shows `masks` in `sources`, `mask` in `rois`, and the key-value pairs used to relate  | ||||
| a mask to its source in a frame. | ||||
| 
 | ||||
| 
 | ||||
|  | ||||
| @ -188,7 +188,7 @@ This example demonstrates `sources` for video, `masks`, and `preview`. | ||||
| 
 | ||||
| This frame shows the `masks` section in `sources`, and the top-level `rois` array. | ||||
| 
 | ||||
| In `sources`,  the `masks` subsection contains the sources for the two masks associated with the raw data.  | ||||
| In `sources`, the `masks` subsection contains the sources for the two masks associated with the raw data.  | ||||
| 
 | ||||
| The raw mask data is located in: | ||||
| 
 | ||||
|  | ||||
| @ -47,7 +47,7 @@ The version information is presented in the following tabs: | ||||
| * [Info](#info)  | ||||
|     | ||||
| ## Frames | ||||
| The **Frames**  tab displays the contents of the selected dataset version. | ||||
| The **Frames** tab displays the contents of the selected dataset version. | ||||
| 
 | ||||
| View the version's frames as thumbnail previews or in a table. Use the view toggle to switch between thumbnail  | ||||
| view <img src="/docs/latest/icons/ico-grid-view.svg" alt="thumbnail view" className="icon size-md space-sm" /> and  | ||||
| @ -71,7 +71,7 @@ To view the details of a specific frame, click on its preview, which will open t | ||||
| ### Simple Frame Filtering | ||||
| Simple frame filtering returns frames containing at least one annotation with a specified label. | ||||
| 
 | ||||
| **To apply a simple frame filter,**  select a label from the **LABEL FILTER** list. | ||||
| **To apply a simple frame filter,** select a label from the **LABEL FILTER** list. | ||||
| 
 | ||||
| <details className="cml-expansion-panel screenshot"> | ||||
| <summary className="cml-expansion-panel-summary">Simple filter example</summary> | ||||
|  | ||||
| @ -2,11 +2,11 @@ | ||||
| title: The Dataviews Table | ||||
| --- | ||||
| 
 | ||||
| [Dataviews](../dataviews.md) appear in the same Project as the experiment that stored the Dataview in the ClearML Enterprise platform,  | ||||
| as well as the **DATAVIEWS** tab in the **All Projects** page. | ||||
| 
 | ||||
| The **Dataviews table** is a [customizable](#customizing-the-dataviews-table) list of Dataviews associated with a project. | ||||
| Use it to view, create, and edit Dataviews in the info panel.  | ||||
| Use it to view and create Dataviews, and access their info panels.  | ||||
| 
 | ||||
| The table lists independent Dataview objects. To see Dataviews logged by a task, go | ||||
| to the specific task's **DATAVIEWS** tab (see [Experiment Dataviews](webapp_exp_track_visual.md)). | ||||
| 
 | ||||
| View the Dataviews table in table view <img src="/docs/latest/icons/ico-table-view.svg" alt="Table view" className="icon size-md space-sm" />  | ||||
| or in details view <img src="/docs/latest/icons/ico-split-view.svg" alt="Details view" className="icon size-md space-sm" />, | ||||
| @ -84,7 +84,7 @@ The same information can be found in the bottom menu, in a tooltip that appears | ||||
| 
 | ||||
| ## Creating a Dataview  | ||||
| 
 | ||||
| Create a new Dataview by clicking the **+ NEW DATAVIEW** button at the top right of the table, which open a  | ||||
| Create a new Dataview by clicking the **+ NEW DATAVIEW** button at the top right of the table, which opens a  | ||||
| **NEW DATAVIEW** window.  | ||||
| 
 | ||||
|  | ||||
| @ -58,7 +58,7 @@ when creating a pipeline step. | ||||
| 
 | ||||
| ### Pipeline Step Caching | ||||
| The Pipeline controller also offers step caching, meaning, reusing outputs of previously executed pipeline steps, in the  | ||||
| case of  exact same step code, and the same step input values. By default, pipeline steps are not cached. Enable caching | ||||
| case of exact same step code, and the same step input values. By default, pipeline steps are not cached. Enable caching | ||||
| when creating a pipeline step. | ||||
| 
 | ||||
| When a step is cached, the step code is hashed, alongside the step’s parameters (as passed in runtime), into a single  | ||||
|  | ||||
| @ -38,7 +38,7 @@ def main(pickle_url, mock_parameter='mock'): | ||||
|   the following format: `{'section_name':['param_name']]}`. For example, the pipeline in the code above will store the  | ||||
|   `pickle_url` parameter in the `General` section and `mock_parameter` in the `Mock` section. By default, arguments will  | ||||
|   be stored in the `Args` section.   | ||||
| * `pool_frequency` - The pooling frequency (in minutes) for monitoring experiments / states. | ||||
| * `pool_frequency` - The polling frequency (in minutes) for monitoring experiments / states. | ||||
| * `add_pipeline_tags` - If `True`, add `pipe: <pipeline_task_id>` tag to all steps (Tasks) created by this pipeline  | ||||
|   (this is useful to create better visibility in projects with multiple pipelines, and for easy selection) (default:  | ||||
|   `False`). | ||||
| @ -111,11 +111,11 @@ def step_one(pickle_data_url: str, extra: int = 43): | ||||
|   Example, assuming we have two functions, `parse_data()` and `load_data()`: `[parse_data, load_data]` | ||||
| * `parents` – Optional list of parent steps in the pipeline. The current step in the pipeline will be sent for execution only after all the parent steps have been executed successfully. | ||||
| 
 | ||||
| Additionally, you can enable automatic logging of a step’s metrics / artifacts / models  to the pipeline task using the  | ||||
| Additionally, you can enable automatic logging of a step’s metrics / artifacts / models to the pipeline task using the  | ||||
| following arguments: | ||||
| * `monitor_metrics` (Optional) - Automatically log the step's reported metrics also on the pipeline Task. The expected  | ||||
|   format is one of the following: | ||||
|   * List of pairs metric (title, series) to log: [(step_metric_title, step_metric_series), ].  Example: `[('test', 'accuracy'), ]` | ||||
|   * List of pairs metric (title, series) to log: [(step_metric_title, step_metric_series), ]. Example: `[('test', 'accuracy'), ]` | ||||
|   * List of tuple pairs, to specify a different target metric to use on the pipeline Task: [((step_metric_title, step_metric_series), (target_metric_title, target_metric_series)), ]. | ||||
|   Example: `[[('test', 'accuracy'), ('model', 'accuracy')], ]` | ||||
| * `monitor_artifacts` (Optional) - Automatically log the step's artifacts on the pipeline Task.  | ||||
|  | ||||
| @ -187,7 +187,7 @@ def step_completed_callback( | ||||
| 
 | ||||
| #### Models, Artifacts, and Metrics  | ||||
| 
 | ||||
| You can enable automatic logging of a step’s metrics /artifacts / models  to the pipeline task using the following arguments: | ||||
| You can enable automatic logging of a step’s metrics /artifacts / models to the pipeline task using the following arguments: | ||||
| 
 | ||||
| * `monitor_metrics` (Optional) - Automatically log the step's reported metrics also on the pipeline Task. The expected  | ||||
|   format is one of the following: | ||||
|  | ||||
| @ -10,7 +10,7 @@ This release is not backwards compatible | ||||
| 
 | ||||
| **Breaking Changes** | ||||
| * `preprocess` and `postprocess` class functions get 3 arguments | ||||
| * Add support for per-request state storage, passing information between the pre/post processing functions | ||||
| * Add support for per-request state storage, passing information between the pre/post-processing functions | ||||
| 
 | ||||
| **Features & Bug Fixes** | ||||
| 
 | ||||
|  | ||||
| @ -239,7 +239,7 @@ This release is not backwards compatible - see notes below on upgrading | ||||
| - Add support for uploading artifacts with a list of files using `Task.upload_artifcats(name, [Path(), Path()])` | ||||
| - Add missing *clearml-task* parameters `--docker_args`, `--docker_bash_setup_script` and `--output-uri` | ||||
| - Change `CreateAndPopulate` will auto list packages imported but not installed locally | ||||
| - Add `clearml.task.populate.create_task_from_function()` to create a Task from a function,  wrapping function input arguments into hyper-parameter section as kwargs and storing function results as named artifacts | ||||
| - Add `clearml.task.populate.create_task_from_function()` to create a Task from a function, wrapping function input arguments into hyper-parameter section as kwargs and storing function results as named artifacts | ||||
| - Add support for Task serialization (e.g. for pickle) | ||||
| - Add `Task.get_configuration_object_as_dict()` | ||||
| - Add `docker_image` argument to `Task.set_base_docker()` (deprecate `docker_cmd`) | ||||
| @ -367,7 +367,7 @@ ClearML k8s glue default pod label was changed to `CLEARML=agent` (instead of `T | ||||
| **Bug Fixes** | ||||
| 
 | ||||
| - Fix experiment details UI failure opening hyperparameter sections beginning with `#` [ClearML Server GitHub issue #79](https://github.com/allegroai/clearml-server/issues/79) | ||||
| - Fix performance issues with UI comparison of  large experiments [Slack Channel](https://clearml.slack.com/archives/CTK20V944/p1621698235159800) | ||||
| - Fix performance issues with UI comparison of large experiments [Slack Channel](https://clearml.slack.com/archives/CTK20V944/p1621698235159800) | ||||
| - Fix filtering on hyperparameters [ClearML GitHub issue #385](https://github.com/allegroai/clearml/issues/385) [Slack Channel](https://clearml.slack.com/archives/CTK20V944/p1626600582284700) | ||||
| - Fix profile page user options toggle control area of effect | ||||
| - Fix browser resizing affecting plot zoom | ||||
|  | ||||
| @ -89,7 +89,7 @@ title: Version 1.6 | ||||
| * Fix listed models in UI pipeline run info panel doesn't link to model  | ||||
| * Fix "Load more" button disappears from UI experiment page 	 | ||||
| * Fix breadcrumb link to parent project does not navigate to the parent's project page | ||||
| * Fix spaces deleted while typing query in UI  search bars | ||||
| * Fix spaces deleted while typing query in UI search bars | ||||
| * Fix UI plots not loading in experiments  | ||||
| * Fix UI experiment debug sample full screen failing to display multiple metrics | ||||
| * Fix using search in UI tables removes custom columns | ||||
|  | ||||
| @ -148,7 +148,7 @@ configuration [here](#aws-iam-restricted-access-policy). | ||||
| 1. Complete creating the policy | ||||
| 1. Attach the created policy to an IAM user/group whose credentials will be used in the autoscaler app (you can create a  | ||||
|    new IAM user/group for this purpose) | ||||
| 1. Obtain a set of AWS IAM credentials for the user/group to which  you have attached the created policy in the previous step   | ||||
| 1. Obtain a set of AWS IAM credentials for the user/group to which you have attached the created policy in the previous step   | ||||
| 
 | ||||
| 
 | ||||
| ### AWS IAM Restricted Access Policy | ||||
|  | ||||
| @ -6,7 +6,7 @@ title: GCP Autoscaler | ||||
| The ClearML GCP Autoscaler App is available under the ClearML Pro plan | ||||
| ::: | ||||
| 
 | ||||
| The GCP Autoscaler Application optimizes GCP VM instance usage according to a user defined instance  budget: Define your  | ||||
| The GCP Autoscaler Application optimizes GCP VM instance usage according to a user defined instance budget: Define your  | ||||
| budget by specifying the type and amount of available compute resources. | ||||
| 
 | ||||
| Each resource type is associated with a ClearML [queue](../../fundamentals/agents_and_queues.md#what-is-a-queue) whose  | ||||
|  | ||||
| @ -16,7 +16,7 @@ ClearML provides the following applications: | ||||
| * [**GPU Compute**](apps_gpu_compute.md) - Launch cloud machines on demand and optimize their usage according to a  | ||||
|   defined budget--no previous setup necessary | ||||
| * [**AWS Autoscaler**](apps_aws_autoscaler.md) - Optimize AWS EC2 instance usage according to a defined instance budget | ||||
| * [**GCP Autoscaler**](apps_gcp_autoscaler.md) - Optimize  GCP instance usage according to a defined instance budget | ||||
| * [**GCP Autoscaler**](apps_gcp_autoscaler.md) - Optimize GCP instance usage according to a defined instance budget | ||||
| * [**Hyperparameter Optimization**](apps_hpo.md) - Find the parameter values that yield the best performing models | ||||
| * **Nvidia Clara** - Train models using Nvidia’s Clara framework | ||||
| * [**Project Dashboard**](apps_dashboard.md) - High-level project monitoring with Slack alerts | ||||
|  | ||||
| @ -132,7 +132,7 @@ The following table describes the actions that can be done from the experiments | ||||
| that allow each operation.   | ||||
| 
 | ||||
| Access these actions with the context menu in any of the following ways: | ||||
| * In the experiments table, right-click an experiment  or hover over an experiment and click <img src="/docs/latest/icons/ico-dots-v-menu.svg" alt="Dot menu" className="icon size-md space-sm" /> | ||||
| * In the experiments table, right-click an experiment or hover over an experiment and click <img src="/docs/latest/icons/ico-dots-v-menu.svg" alt="Dot menu" className="icon size-md space-sm" /> | ||||
| * In an experiment info panel, click the menu button <img src="/docs/latest/icons/ico-bars-menu.svg" alt="Bar menu" className="icon size-md space-sm" /> | ||||
| 
 | ||||
| | Action | Description | States Valid for the Action | State Transition | | ||||
|  | ||||
| @ -6,7 +6,7 @@ Use the Projects Page for project navigation and management. | ||||
| 
 | ||||
| Your projects are displayed like folders: click a folder to access its contents. The Projects Page shows the top-level  | ||||
| projects in your workspace. Projects that contain nested subprojects are identified by an extra nested project tab.  | ||||
| An exception is the **All Experiments** folder, which shows all projects’ and subprojects’  contents in a single, flat | ||||
| An exception is the **All Experiments** folder, which shows all projects’ and subprojects’ contents in a single, flat | ||||
| list. | ||||
| 
 | ||||
|  | ||||
|  | ||||
| @ -24,7 +24,7 @@ The worker table shows the currently available workers and their current executi | ||||
| 
 | ||||
| 
 | ||||
| Clicking on a worker will open the worker’s details panel and replace the graph with that worker’s resource utilization  | ||||
| information.  The resource metric being monitored can be selected through the menu at the graph’s top left corner: | ||||
| information. The resource metric being monitored can be selected through the menu at the graph’s top left corner: | ||||
| * CPU and GPU Usage | ||||
| * Memory Usage | ||||
| * Video Memory Usage | ||||
| @ -37,7 +37,7 @@ The worker’s details panel includes the following two tabs: | ||||
|     * Current Experiment - The experiment currently being executed by the worker | ||||
|     * Experiment Runtime - How long the currently executing experiment has been running | ||||
|     * Experiment iteration - The last reported training iteration for the experiment | ||||
| * **QUEUES** -  information about the queues that the worker is assigned to: | ||||
| * **QUEUES** -  Information about the queues that the worker is assigned to: | ||||
|     * Queue - The name of the Queue | ||||
|     * Next experiment - The next experiment available in this queue | ||||
|     * In Queue - The number of experiments currently enqueued | ||||
|  | ||||
		Loading…
	
		Reference in New Issue
	
	Block a user