mirror of
				https://github.com/clearml/clearml-docs
				synced 2025-06-26 18:17:44 +00:00 
			
		
		
		
	Small edits (#787)
This commit is contained in:
		
							parent
							
								
									f516d08c71
								
							
						
					
					
						commit
						1c45124714
					
				| @ -53,15 +53,15 @@ See all [storage capabilities](../../integrations/storage.md). | ||||
| 
 | ||||
| Upload a local file containing the preprocessed results of the data: | ||||
| ```python | ||||
| task.upload_artifact('/path/to/preprocess_data.csv', name='data') | ||||
| task.upload_artifact(name='data', artifact_object='/path/to/preprocess_data.csv') | ||||
| ``` | ||||
| 
 | ||||
| You can also upload an entire folder with all its content by passing the folder (the folder will be zipped and uploaded as a single zip file). | ||||
| ```python | ||||
| task.upload_artifact('/path/to/folder/', name='folder') | ||||
| task.upload_artifact(name='folder', artifact_object='/path/to/folder/') | ||||
| ``` | ||||
| 
 | ||||
| Lastly, you can upload an instance of an object; Numpy/Pandas/PIL Images are supported with npz/csv.gz/jpg formats accordingly. | ||||
| Lastly, you can upload an instance of an object; Numpy/Pandas/PIL Images are supported with `npz`/`csv.gz`/`jpg` formats accordingly. | ||||
| If the object type is unknown, ClearML pickles it and uploads the pickle file. | ||||
| 
 | ||||
| ```python | ||||
| @ -69,7 +69,7 @@ numpy_object = np.eye(100, 100) | ||||
| task.upload_artifact(name='features', artifact_object=numpy_object) | ||||
| ``` | ||||
| 
 | ||||
| Check out all [artifact logging](../../clearml_sdk/task_sdk.md#artifacts) options. | ||||
| For more artifact logging options, see [Artifacts](../../clearml_sdk/task_sdk.md#artifacts). | ||||
| 
 | ||||
| ### Using Artifacts | ||||
| 
 | ||||
| @ -137,26 +137,26 @@ This feature lets you easily get a full genealogy of every trained and used mode | ||||
| ## Log Metrics | ||||
| 
 | ||||
| Full metrics logging is the key to finding the best performing model! | ||||
| By default, everything that's reported to TensorBoard and Matplotlib is automatically captured and logged. | ||||
| By default, ClearML automatically captures and logs everything reported to TensorBoard and Matplotlib. | ||||
| 
 | ||||
| Since not all metrics are tracked that way, you can also manually report metrics using a [`Logger`](../../fundamentals/logger.md) object. | ||||
| 
 | ||||
| You can log everything, from time series data to confusion matrices to HTML, Audio and Video, to custom plotly graphs! Everything goes! | ||||
| You can log everything, from time series data and confusion matrices to HTML, Audio, and Video, to custom plotly graphs! Everything goes! | ||||
| 
 | ||||
|  | ||||
|  | ||||
| 
 | ||||
| Once everything is neatly logged and displayed, using the [comparison tool](../../webapp/webapp_exp_comparing.md) makes it easy to find the best configuration! | ||||
| Once everything is neatly logged and displayed, use the [comparison tool](../../webapp/webapp_exp_comparing.md) to find the best configuration! | ||||
| 
 | ||||
| 
 | ||||
| ## Track Experiments | ||||
| 
 | ||||
| The experiment table is a powerful tool for creating dashboards and views of your own projects, your team's projects, or the entire development. | ||||
| The experiments table is a powerful tool for creating dashboards and views of your own projects, your team's projects, or the entire development. | ||||
| 
 | ||||
|  | ||||
|  | ||||
| 
 | ||||
| 
 | ||||
| ### Creating Leaderboards | ||||
| Customize the [experiments table](../../webapp/webapp_exp_table.md) to fit your own needs, adding desired views of parameters, metrics and tags. | ||||
| Customize the [experiments table](../../webapp/webapp_exp_table.md) to fit your own needs, adding desired views of parameters, metrics, and tags. | ||||
| You can filter and sort based on parameters and metrics, so creating custom views is simple and flexible. | ||||
| 
 | ||||
| Create a dashboard for a project, presenting the latest Models and their accuracy scores, for immediate insights. | ||||
|  | ||||
| @ -10,8 +10,8 @@ Metadata can be customized as needed using: **meta** dictionaries: | ||||
| 
 | ||||
| ### Adding Frame Metadata | ||||
| 
 | ||||
| When instantiating a Frame, metadata that applies for the entire frame can be | ||||
| added as an argument. | ||||
| When instantiating a `SingleFrame`, metadata that applies to the entire frame can be | ||||
| added as an argument: | ||||
| 
 | ||||
| ```python | ||||
| from allegroai import SingleFrame | ||||
| @ -30,11 +30,13 @@ frame.metadata['dangerous'] = 'no' | ||||
| 
 | ||||
| ### Adding ROI Metadata | ||||
| 
 | ||||
| Metadata can be added to individual ROIs when adding an annotation to a `frame`, using the [`SingleFrame.add_annotation`](../references/hyperdataset/singleframe.md#add_annotation) | ||||
| method.  | ||||
| Metadata can be added to individual ROIs when adding an annotation to a `frame`, using [`SingleFrame.add_annotation()`](../references/hyperdataset/singleframe.md#add_annotation):  | ||||
| 
 | ||||
| ```python | ||||
| frame.add_annotation(box2d_xywh=(10, 10, 30, 20), labels=['tiger'], | ||||
|                      # insert metadata dictionary | ||||
|                      metadata={'dangerous':'yes'}) | ||||
| frame.add_annotation( | ||||
|     box2d_xywh=(10, 10, 30, 20),  | ||||
|     labels=['tiger'], | ||||
|     # insert metadata dictionary | ||||
|     metadata={'dangerous':'yes'} | ||||
| ) | ||||
| ``` | ||||
		Loading…
	
		Reference in New Issue
	
	Block a user