mirror of
https://github.com/clearml/clearml-docs
synced 2025-02-01 06:56:56 +00:00
203 lines
5.5 KiB
Markdown
203 lines
5.5 KiB
Markdown
---
|
|
title: Storage
|
|
---
|
|
|
|
|
|
ClearML is able to interface with the most popular storage solutions in the market for storing model checkpoints, artifacts
|
|
and charts.
|
|
|
|
Supported storage mediums include:
|
|
|
|
![image](../../static/icons/ClearML_Supported_Storage--on-light.png)
|
|
|
|
:::note
|
|
Once uploading an object to a storage medium, each machine that uses the object must have access to it.
|
|
:::
|
|
|
|
## Configuring Storage
|
|
|
|
Configuration for storage is done by editing the [clearml.conf](../configs/clearml_conf.md).
|
|
|
|
The ClearML configuration file uses [HOCON](https://github.com/lightbend/config/blob/main/HOCON.md) format, which supports runtime environment variable access.
|
|
|
|
### Configuring AWS S3
|
|
|
|
Modify these parts of the clearml.conf file and add the key, secret, and region of the s3 bucket.
|
|
It's possible to also give access to specific s3 buckets.
|
|
```
|
|
aws {
|
|
s3 {
|
|
# S3 credentials, used for read/write access by various SDK elements
|
|
|
|
# default, used for any bucket not specified below
|
|
key: ""
|
|
secret: ""
|
|
region: ""
|
|
|
|
credentials: [
|
|
# specifies key/secret credentials to use when handling s3 urls (read or write)
|
|
{
|
|
bucket: "my-bucket-name"
|
|
key: ""
|
|
secret: ""
|
|
verify: "/path/to/ca/bundle.crt" OR false to not verify
|
|
},
|
|
|
|
]
|
|
}
|
|
boto3 {
|
|
pool_connections: 512
|
|
max_multipart_concurrency: 16
|
|
}
|
|
}
|
|
```
|
|
|
|
AWS's S3 access parameters can be specified by referencing the standard environment variables if already defined.
|
|
|
|
For example:
|
|
```
|
|
aws {
|
|
s3 {
|
|
# default, used for any bucket not specified below
|
|
key: ${AWS_ACCESS_KEY_ID}
|
|
secret: ${AWS_SECRET_ACCESS_KEY}
|
|
region: ${AWS_DEFAULT_REGION}
|
|
}
|
|
}
|
|
```
|
|
|
|
ClearML also supports [MinIO](https://github.com/minio/minio) by adding this configuration:
|
|
```
|
|
aws {
|
|
s3 {
|
|
# default, used for any bucket not specified below
|
|
key: ""
|
|
secret: ""
|
|
region: ""
|
|
|
|
credentials: [
|
|
{
|
|
# This will apply to all buckets in this host (unless key/value is specifically provided for a given bucket)
|
|
host: "my-minio-host:9000"
|
|
key: ""
|
|
secret: ""
|
|
multipart: false
|
|
secure: false
|
|
}
|
|
]
|
|
}
|
|
}
|
|
```
|
|
|
|
:::info non-AWS Endpoints
|
|
To force usage of a non-AWS endpoint (like the MinIO example above), port declaration is *always* needed, even if standard.
|
|
To enable TLS, pass `secure: true`.
|
|
:::
|
|
|
|
### Configuring Azure
|
|
To configure Azure blob storage specify the account name and key.
|
|
|
|
```
|
|
azure.storage {
|
|
containers: [
|
|
{
|
|
account_name: ""
|
|
account_key: ""
|
|
# container_name:
|
|
}
|
|
]
|
|
}
|
|
```
|
|
|
|
Azure's storage access parameters can be specified by referencing the standard environment variables if already defined.
|
|
|
|
For example:
|
|
```
|
|
azure.storage {
|
|
containers: [
|
|
{
|
|
account_name: ${AZURE_STORAGE_ACCOUNT}
|
|
account_key: ${AZURE_STORAGE_KEY}
|
|
# container_name:
|
|
}
|
|
]
|
|
}
|
|
```
|
|
|
|
### Configuring Google Storage
|
|
To configure Google Storage, specify the project and the path to the credentials json file.
|
|
It's also possible to specify credentials for a specific bucket.
|
|
|
|
```
|
|
google.storage {
|
|
# Default project and credentials file
|
|
# Will be used when no bucket configuration is found
|
|
project: "clearml"
|
|
credentials_json: "/path/to/credentials.json"
|
|
|
|
# Specific credentials per bucket and sub directory
|
|
credentials = [
|
|
{
|
|
bucket: ""
|
|
subdir: "path/in/bucket" # Not required
|
|
project: ""
|
|
credentials_json: "/path/to/credentials.json"
|
|
},
|
|
]
|
|
}
|
|
```
|
|
|
|
GCP's storage access parameters can be specified by referencing the standard environment variables if already defined.
|
|
|
|
```
|
|
google.storage {
|
|
credentials = [
|
|
{
|
|
bucket: ""
|
|
subdir: "path/in/bucket" # Not required
|
|
project: ""
|
|
credentials_json: ${GOOGLE_APPLICATION_CREDENTIALS}
|
|
},
|
|
]
|
|
}
|
|
```
|
|
|
|
## Storage Manager
|
|
|
|
ClearML offers the [StorageManager](../references/sdk/storage.md) class to manage downloading, uploading, and caching of
|
|
content directly from code.
|
|
|
|
See [Storage Examples](../guides/storage/examples_storagehelper.md).
|
|
|
|
|
|
## Caching
|
|
ClearML also manages a cache of all downloaded content so nothing is duplicated, and code won't need to download the same
|
|
piece twice!
|
|
|
|
Configure cache location by modifying the [clearml.conf](../configs/clearml_conf.md) file:
|
|
|
|
```
|
|
storage {
|
|
cache {
|
|
# Defaults to system temp folder / cache
|
|
default_base_dir: "~/.clearml/cache"
|
|
}
|
|
|
|
direct_access: [
|
|
# Objects matching are considered to be available for direct access, i.e. they will not be downloaded
|
|
# or cached, and any download request will return a direct reference.
|
|
# Objects are specified in glob format, available for url and content_type.
|
|
{ url: "file://*" } # file-urls are always directly referenced
|
|
]
|
|
}
|
|
```
|
|
|
|
### Direct Access
|
|
By default, all artifacts (Models / Artifacts / Datasets) are automatically downloaded to the cache before they're used.
|
|
|
|
Some storage mediums (NFS / Local storage) allows for direct access,
|
|
which means that the code would work with the object where it's originally stored and not downloaded to cache first.
|
|
|
|
To enable direct access, specify the urls to access directly.
|
|
|