mirror of
https://github.com/clearml/clearml
synced 2025-06-26 18:16:07 +00:00
Add support for .get ing pipelines and enqueue-ing them
This commit is contained in:
@@ -5,20 +5,21 @@ from clearml import TaskTypes
|
||||
# Make the following function an independent pipeline component step
|
||||
# notice all package imports inside the function will be automatically logged as
|
||||
# required packages for the pipeline execution step
|
||||
@PipelineDecorator.component(return_values=['data_frame'], cache=True, task_type=TaskTypes.data_processing)
|
||||
@PipelineDecorator.component(return_values=["data_frame"], cache=True, task_type=TaskTypes.data_processing)
|
||||
def step_one(pickle_data_url: str, extra: int = 43):
|
||||
print('step_one')
|
||||
print("step_one")
|
||||
# make sure we have scikit-learn for this step, we need it to use to unpickle the object
|
||||
import sklearn # noqa
|
||||
import pickle
|
||||
import pandas as pd
|
||||
from clearml import StorageManager
|
||||
|
||||
local_iris_pkl = StorageManager.get_local_copy(remote_url=pickle_data_url)
|
||||
with open(local_iris_pkl, 'rb') as f:
|
||||
with open(local_iris_pkl, "rb") as f:
|
||||
iris = pickle.load(f)
|
||||
data_frame = pd.DataFrame(iris['data'], columns=iris['feature_names'])
|
||||
data_frame.columns += ['target']
|
||||
data_frame['target'] = iris['target']
|
||||
data_frame = pd.DataFrame(iris["data"], columns=iris["feature_names"])
|
||||
data_frame.columns += ["target"]
|
||||
data_frame["target"] = iris["target"]
|
||||
return data_frame
|
||||
|
||||
|
||||
@@ -28,18 +29,17 @@ def step_one(pickle_data_url: str, extra: int = 43):
|
||||
# Specifying `return_values` makes sure the function step can return an object to the pipeline logic
|
||||
# In this case, the returned tuple will be stored as an artifact named "X_train, X_test, y_train, y_test"
|
||||
@PipelineDecorator.component(
|
||||
return_values=['X_train, X_test, y_train, y_test'], cache=True, task_type=TaskTypes.data_processing
|
||||
return_values=["X_train", "X_test", "y_train", "y_test"], cache=True, task_type=TaskTypes.data_processing
|
||||
)
|
||||
def step_two(data_frame, test_size=0.2, random_state=42):
|
||||
print('step_two')
|
||||
print("step_two")
|
||||
# make sure we have pandas for this step, we need it to use the data_frame
|
||||
import pandas as pd # noqa
|
||||
from sklearn.model_selection import train_test_split
|
||||
y = data_frame['target']
|
||||
X = data_frame[(c for c in data_frame.columns if c != 'target')]
|
||||
X_train, X_test, y_train, y_test = train_test_split(
|
||||
X, y, test_size=test_size, random_state=random_state
|
||||
)
|
||||
|
||||
y = data_frame["target"]
|
||||
X = data_frame[(c for c in data_frame.columns if c != "target")]
|
||||
X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=test_size, random_state=random_state)
|
||||
|
||||
return X_train, X_test, y_train, y_test
|
||||
|
||||
@@ -49,37 +49,41 @@ def step_two(data_frame, test_size=0.2, random_state=42):
|
||||
# required packages for the pipeline execution step
|
||||
# Specifying `return_values` makes sure the function step can return an object to the pipeline logic
|
||||
# In this case, the returned object will be stored as an artifact named "model"
|
||||
@PipelineDecorator.component(return_values=['model'], cache=True, task_type=TaskTypes.training)
|
||||
@PipelineDecorator.component(return_values=["model"], cache=True, task_type=TaskTypes.training)
|
||||
def step_three(X_train, y_train):
|
||||
print('step_three')
|
||||
print("step_three")
|
||||
# make sure we have pandas for this step, we need it to use the data_frame
|
||||
import pandas as pd # noqa
|
||||
from sklearn.linear_model import LogisticRegression
|
||||
model = LogisticRegression(solver='liblinear', multi_class='auto')
|
||||
|
||||
model = LogisticRegression(solver="liblinear", multi_class="auto")
|
||||
model.fit(X_train, y_train)
|
||||
return model
|
||||
|
||||
|
||||
# Make the following function an independent pipeline component step
|
||||
# notice all package imports inside the function will be automatically logged as
|
||||
# required packages for the pipeline execution step
|
||||
# Specifying `return_values` makes sure the function step can return an object to the pipeline logic
|
||||
# In this case, the returned object will be stored as an artifact named "accuracy"
|
||||
@PipelineDecorator.component(return_values=['accuracy'], cache=True, task_type=TaskTypes.qc)
|
||||
@PipelineDecorator.component(return_values=["accuracy"], cache=True, task_type=TaskTypes.qc)
|
||||
def step_four(model, X_data, Y_data):
|
||||
from sklearn.linear_model import LogisticRegression # noqa
|
||||
from sklearn.metrics import accuracy_score
|
||||
|
||||
Y_pred = model.predict(X_data)
|
||||
return accuracy_score(Y_data, Y_pred, normalize=True)
|
||||
|
||||
|
||||
# The actual pipeline execution context
|
||||
# notice that all pipeline component function calls are actually executed remotely
|
||||
# Only when a return value is used, the pipeline logic will wait for the component execution to complete
|
||||
@PipelineDecorator.pipeline(name='custom pipeline logic', project='examples', version='0.0.5')
|
||||
def executing_pipeline(pickle_url, mock_parameter='mock'):
|
||||
print('pipeline args:', pickle_url, mock_parameter)
|
||||
@PipelineDecorator.pipeline(name="custom pipeline logic", project="examples", version="0.0.5")
|
||||
def executing_pipeline(pickle_url, mock_parameter="mock"):
|
||||
print("pipeline args:", pickle_url, mock_parameter)
|
||||
|
||||
# Use the pipeline argument to start the pipeline and pass it ot the first step
|
||||
print('launch step one')
|
||||
print("launch step one")
|
||||
data_frame = step_one(pickle_url)
|
||||
|
||||
# Use the returned data from the first step (`step_one`), and pass it to the next step (`step_two`)
|
||||
@@ -87,17 +91,17 @@ def executing_pipeline(pickle_url, mock_parameter='mock'):
|
||||
# the pipeline logic does not actually load the artifact itself.
|
||||
# When actually passing the `data_frame` object into a new step,
|
||||
# It waits for the creating step/function (`step_one`) to complete the execution
|
||||
print('launch step two')
|
||||
print("launch step two")
|
||||
X_train, X_test, y_train, y_test = step_two(data_frame)
|
||||
|
||||
print('launch step three')
|
||||
print("launch step three")
|
||||
model = step_three(X_train, y_train)
|
||||
|
||||
# Notice since we are "printing" the `model` object,
|
||||
# we actually deserialize the object from the third step, and thus wait for the third step to complete.
|
||||
print('returned model: {}'.format(model))
|
||||
print("returned model: {}".format(model))
|
||||
|
||||
print('launch step four')
|
||||
print("launch step four")
|
||||
accuracy = 100 * step_four(model, X_data=X_test, Y_data=y_test)
|
||||
|
||||
# Notice since we are "printing" the `accuracy` object,
|
||||
@@ -105,7 +109,7 @@ def executing_pipeline(pickle_url, mock_parameter='mock'):
|
||||
print(f"Accuracy={accuracy}%")
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
if __name__ == "__main__":
|
||||
# set the pipeline steps default execution queue (per specific step we can override it with the decorator)
|
||||
# PipelineDecorator.set_default_execution_queue('default')
|
||||
# Run the pipeline steps as subprocesses on the current machine, great for local executions
|
||||
@@ -114,7 +118,7 @@ if __name__ == '__main__':
|
||||
|
||||
# Start the pipeline execution logic.
|
||||
executing_pipeline(
|
||||
pickle_url='https://github.com/allegroai/events/raw/master/odsc20-east/generic/iris_dataset.pkl',
|
||||
pickle_url="https://github.com/allegroai/events/raw/master/odsc20-east/generic/iris_dataset.pkl",
|
||||
)
|
||||
|
||||
print('process completed')
|
||||
print("process completed")
|
||||
|
||||
Reference in New Issue
Block a user