From e39ab69cf988a4911f12e3c6d6a8ea089aaa9d62 Mon Sep 17 00:00:00 2001
From: allegroai <>
Date: Sun, 15 Dec 2019 00:11:14 +0200
Subject: [PATCH] Documentation
---
docs/faq.md | 37 +++++++++++++++++++++++++++++++++++--
1 file changed, 35 insertions(+), 2 deletions(-)
diff --git a/docs/faq.md b/docs/faq.md
index db0954b2..c3bb6636 100644
--- a/docs/faq.md
+++ b/docs/faq.md
@@ -26,6 +26,7 @@ Experiments
* [Sometimes I see experiments as running when in fact they are not. What's going on?](#experiment-running-but-stopped)
* [My code throws an exception, but my experiment status is not "Failed". What happened?](#exception-not-failed)
* [When I run my experiment, I get an SSL Connection error [CERTIFICATE_VERIFY_FAILED]. Do you have a solution?](#ssl-connection-error)
+* [How do I modify experiment names once they have been created?](#name-changing)
Graphs and Logs
@@ -173,13 +174,13 @@ TRAINS monitors your Python process. When the process exits in an orderly fashio
When the process crashes and terminates abnormally, the stop signal is sometimes missed. In such a case, you can safely right click the experiment in the Web-App and stop it.
-## My code throws an exception, but my experiment status is not "Failed". What happened?
+### My code throws an exception, but my experiment status is not "Failed". What happened?
This issue was resolved in v0.9.2. Upgrade TRAINS:
```pip install -U trains```
-## When I run my experiment, I get an SSL Connection error [CERTIFICATE_VERIFY_FAILED]. Do you have a solution?
+### When I run my experiment, I get an SSL Connection error [CERTIFICATE_VERIFY_FAILED]. Do you have a solution?
Your firewall may be preventing the connection. Try one of the following solutons:
@@ -199,6 +200,38 @@ Your firewall may be preventing the connection. Try one of the following soluton
1. Copy the new **trains.conf** file to ~/trains.conf (on Windows: C:\Users\your_username\trains.conf)
+### How do I modify experiment names once they have been created?
+
+An experiments' name is a user controlled property which can be accessed via the `Task.name` variable.
+This allows you to use meaningful naming schemes for to easily filter and compare different experiments.
+
+For example, to distinguish between different experiments you can append the task Id to the task name:
+
+```python
+task = Task.init('examples', 'train')
+task.name += ' {}'.format(task.id)
+```
+
+Or, even for post-execution:
+
+```python
+tasks = Task.get_tasks(project_name='examples', task_name='train')
+for t in tasks:
+ t.name += ' {}'.format(task.id)
+```
+
+Another example - To append a specific hyperparameter and its value to each task's name:
+
+```python
+tasks = Task.get_tasks(project_name='examples', task_name='my_automl_experiment')
+for t in tasks:
+ params = t.get_parameters()
+ if 'my_secret_parameter' in params:
+ t.name += ' my_secret_parameter={}'.format(params['my_secret_parameter'])
+```
+
+Use it also when creating automation pipelines with a naming convention, see our [random search automation example](https://github.com/allegroai/trains/blob/master/examples/automl/automl_random_search_example.py).
+
## Graphs and Logs
### The first log lines are missing from the experiment log tab. Where did they go?