mirror of
https://github.com/clearml/clearml-docs
synced 2025-02-12 15:36:12 +00:00
94 lines
4.5 KiB
Markdown
94 lines
4.5 KiB
Markdown
---
|
||
title: HuggingFace Transformers
|
||
---
|
||
|
||
HuggingFace's [Transformers](https://huggingface.co/docs/transformers/index) is a popular deep learning framework. You can
|
||
seamlessly integrate ClearML into your Transformer's PyTorch [Trainer](https://huggingface.co/docs/transformers/v4.34.1/en/main_classes/trainer)
|
||
code using the built-in [`ClearMLCallback`](https://huggingface.co/docs/transformers/v4.34.1/en/main_classes/callback#transformers.integrations.ClearMLCallback).
|
||
ClearML automatically logs Transformer's models, parameters, scalars, and more.
|
||
|
||
All you have to do is install and set up ClearML:
|
||
|
||
1. Install the `clearml` python package:
|
||
|
||
```commandline
|
||
pip install clearml
|
||
```
|
||
|
||
1. To keep track of your experiments and/or data, ClearML needs to communicate to a server. You have 2 server options:
|
||
* Sign up for free to the [ClearML Hosted Service](https://app.clear.ml/)
|
||
* Set up your own server, see [here](../deploying_clearml/clearml_server.md).
|
||
1. Connect the ClearML SDK to the server by creating credentials (go to the top right in the UI to **Settings > Workspace > Create new credentials**),
|
||
then execute the command below and follow the instructions:
|
||
|
||
```commandline
|
||
clearml-init
|
||
```
|
||
|
||
That’s it! In every training run from now on, the ClearML experiment
|
||
manager will capture:
|
||
* Source code and uncommitted changes
|
||
* Hyperparameters - PyTorch trainer [parameters](https://huggingface.co/docs/transformers/v4.34.1/en/main_classes/trainer#transformers.TrainingArguments)
|
||
and TensorFlow definitions
|
||
* Installed packages
|
||
* Model files (make sure the `CLEARML_LOG_MODEL` environment variable is set to `True`)
|
||
* Scalars (loss, learning rates)
|
||
* Console output
|
||
* General details such as machine details, runtime, creation date etc.
|
||
* And more
|
||
|
||
All of this is captured into a [ClearML Task](../fundamentals/task.md). By default, a task called `Trainer` is created
|
||
in the `HuggingFace Transformers` project. To change the task’s name or project, use the `CLEARML_PROJECT` and `CLEARML_TASK`
|
||
environment variables
|
||
|
||
:::tip project names
|
||
ClearML uses `/` as a delimiter for subprojects: using `example/sample` as a name will create the `sample`
|
||
task within the `example` project.
|
||
:::
|
||
|
||
In order to log the models created during training, set the `CLEARML_LOG_MODEL` environment variable to `True`.
|
||
|
||
You can see all the captured data in the task’s page of the ClearML [WebApp](../webapp/webapp_exp_track_visual.md).
|
||
|
||
![transformers scalars](../img/integrations_transformers_scalars.png)
|
||
|
||
Additionally, you can view all of your Transformers runs tracked by ClearML in the [Experiments Table](../webapp/webapp_model_table.md).
|
||
Add custom columns to the table, such as mAP values, so you can easily sort and see what is the best performing model.
|
||
You can also select multiple experiments and directly [compare](../webapp/webapp_exp_comparing.md) them.
|
||
|
||
|
||
## Remote Execution
|
||
ClearML logs all the information required to reproduce an experiment on a different machine (installed packages,
|
||
uncommitted changes etc.). The [ClearML Agent](../clearml_agent.md) listens to designated queues and when a task is
|
||
enqueued, the agent pulls it, recreates its execution environment, and runs it, reporting its scalars, plots, etc. to the
|
||
experiment manager.
|
||
|
||
Deploy a ClearML Agent onto any machine (e.g. a cloud VM, a local GPU machine, your own laptop) by simply running
|
||
the following command on it:
|
||
|
||
```commandline
|
||
clearml-agent daemon --queue <queues_to_listen_to> [--docker]
|
||
```
|
||
|
||
Use the ClearML [Autoscalers](../cloud_autoscaling/autoscaling_overview.md) to help you manage cloud workloads in the
|
||
cloud of your choice (AWS, GCP, Azure) and automatically deploy ClearML agents: the autoscaler automatically spins up
|
||
and shuts down instances as needed, according to a resource budget that you set.
|
||
|
||
|
||
### Cloning, Editing, and Enqueuing
|
||
|
||
![Cloning, editing, enqueuing gif](../img/gif/integrations_yolov5.gif)
|
||
|
||
Use ClearML’s web interface to edit task details, like configuration parameters or input models, then execute the task
|
||
with the new configuration on a remote machine:
|
||
* Clone the experiment
|
||
* Edit the hyperparameters and/or other details
|
||
* Enqueue the task
|
||
|
||
The ClearML Agent executing the task will use the new values to [override any hard coded values](../clearml_agent.md).
|
||
|
||
## Hyperparameter Optimization
|
||
Use ClearML’s [`HyperParameterOptimizer`](../references/sdk/hpo_optimization_hyperparameteroptimizer.md) class to find
|
||
the hyperparameter values that yield the best performing models. See [Hyperparameter Optimization](../fundamentals/hpo.md)
|
||
for more information.
|