HuggingFace's [Transformers](https://huggingface.co/docs/transformers/index) is a popular deep learning framework. You can
seamlessly integrate ClearML into your Transformer's PyTorch [Trainer](https://huggingface.co/docs/transformers/v4.34.1/en/main_classes/trainer)
code using the built-in [`ClearMLCallback`](https://huggingface.co/docs/transformers/v4.34.1/en/main_classes/callback#transformers.integrations.ClearMLCallback).
ClearML automatically logs Transformer's models, parameters, scalars, and more.
Use ClearML's web interface to reproduce tasks and edit their details, like hyperparameters or input models, then execute the tasks
with the new configuration on a remote machine.
When ClearML is integrated into a script, it captures and stores configurations, such as hyperparameters
and model settings. When executing a task, the ClearML Agent will, by default, override runtime configuration values
(such as hyperparameters and environment variables) with the values specified in the task.
However, for tasks using Transformers, the default behavior is different. By default, Transformers tasks ignore UI
overrides and use execution-time parameters (such as environment variables). This is done to prevent potential issues
with environment-specific settings when running tasks on different machines.
**To rerun a task with modified configuration:**
1. Clone the task
1. Edit the hyperparameters and/or other details.
1. In the **CONFIGURATION > HYPERPARAMETERS > Transformers** section, set both `_ignore_hparams_ui_overrides_` and `_ignore_model_config_ui_overrides_`
to `False` . This allows the task to use the new hyperparameter and model
configuration values respectively during execution.