Comment on page

Compute Flow

This page shows how you run a compute flow.
In this page, we provide the steps for publishing algorithm asset, run it on Ocean environment for C2D and retrieve the result logs, using
We assumed that you have completed the installation part with the preferred setup.
Here are the steps:
  1. 1.
    Alice publishes dataset
  2. 2.
    Alice publishes algorithm
  3. 3.
    Alice allows the algorithm for C2D for that data asset
  4. 4.
    Bob acquires datatokens for data and algorithm
  5. 5.
    Bob starts a compute job using a free C2D environment (no provider fees)
  6. 6.
    Bob monitors logs / algorithm output
Let's go through each step.

1. Alice publishes dataset

In the same python console:
# Publish data NFT, datatoken, and asset for dataset based on url
# offers multiple file object types. A simple url file is enough for here
from ocean_lib.structures.file_objects import UrlFile
DATA_url_file = UrlFile(
name = "Branin dataset"
(DATA_data_nft, DATA_datatoken, DATA_ddo) = ocean.assets.create_url_asset(name, DATA_url_file.url, {"from": alice}, with_compute=True, wait_for_aqua=True)
print(f"DATA_data_nft address = '{DATA_data_nft.address}'")
print(f"DATA_datatoken address = '{DATA_datatoken.address}'")
print(f"DATA_ddo did = '{DATA_ddo.did}'")
To customise the privacy and accessibility of your compute service, add the compute_values argument to create_url_asset to set values according to the DDO specs. The function assumes the documented defaults.

2. Alice publishes an algorithm

In the same Python console:
# Publish data NFT & datatoken for algorithm
ALGO_url = ""
name = "grp"
(ALGO_data_nft, ALGO_datatoken, ALGO_ddo) = ocean.assets.create_algo_asset(name, ALGO_url, {"from": alice}, wait_for_aqua=True)
print(f"ALGO_data_nft address = '{ALGO_data_nft.address}'")
print(f"ALGO_datatoken address = '{ALGO_datatoken.address}'")
print(f"ALGO_ddo did = '{ALGO_ddo.did}'")

3. Alice allows the algorithm for C2D for that data asset

In the same Python console:
compute_service =[1]
DATA_ddo = ocean.assets.update(DATA_ddo, {"from": alice})

4. Bob acquires datatokens for data and algorithm

In the same Python console:
# Alice mints DATA datatokens and ALGO datatokens to Bob.
# Alternatively, Bob might have bought these in a market.
from ocean_lib.ocean.util import to_wei, to_wei(5), {"from": alice}), to_wei(5), {"from": alice})
You can choose each method for getting access from consume flow approaches.

5. Bob starts a compute job using a free C2D environment

Only inputs needed: DATA_did, ALGO_did. Everything else can get computed as needed. For demo purposes, we will use the free C2D environment, which requires no provider fees.
In the same Python console:
# Convenience variables
DATA_did = DATA_ddo.did
ALGO_did = ALGO_ddo.did
# Operate on updated and indexed assets
DATA_ddo = ocean.assets.resolve(DATA_did)
ALGO_ddo = ocean.assets.resolve(ALGO_did)
compute_service =[1]
algo_service =[0]
free_c2d_env = ocean.compute.get_free_c2d_environment(compute_service.service_endpoint, DATA_ddo.chain_id)
from datetime import datetime, timedelta, timezone
from ocean_lib.models.compute_input import ComputeInput
DATA_compute_input = ComputeInput(DATA_ddo, compute_service)
ALGO_compute_input = ComputeInput(ALGO_ddo, algo_service)
# Pay for dataset and algo for 1 day
datasets, algorithm = ocean.assets.pay_for_compute_service(
tx_dict={"from": bob},
valid_until=int(( + timedelta(days=1)).timestamp()),
assert datasets, "pay for dataset unsuccessful"
assert algorithm, "pay for algorithm unsuccessful"
# Start compute job
job_id = ocean.compute.start(
print(f"Started compute job with id: {job_id}")

6. Bob monitors logs / algorithm output

In the same Python console, you can check the job status as many times as needed:
# Wait until job is done
import time
from decimal import Decimal
succeeded = False
for _ in range(0, 200):
status = ocean.compute.status(DATA_ddo, compute_service, job_id, bob)
if status.get("dateFinished") and Decimal(status["dateFinished"]) > 0:
succeeded = True
This will output the status of the current job. Here is a list of possible results: Operator Service Status description.
Once the returned status dictionary contains the dateFinished key, Bob can retrieve the job results using ocean.compute.result or, more specifically, just the output if the job was successful. For the purpose of this tutorial, let's choose the second option.
# Retrieve algorithm output and log files
output = ocean.compute.compute_job_result_logs(
DATA_ddo, compute_service, job_id, bob
import pickle
model = pickle.loads(output) # the gaussian model result
assert len(model) > 0, "unpickle result unsuccessful"
You can use the result however you like. For the purpose of this example, let's plot it.
Make sure you have matplotlib package installed in your virtual environment.
import numpy
from matplotlib import pyplot
X0_vec = numpy.linspace(-5., 10., 15)
X1_vec = numpy.linspace(0., 15., 15)
X0, X1 = numpy.meshgrid(X0_vec, X1_vec)
b, c, t = 0.12918450914398066, 1.5915494309189535, 0.039788735772973836
u = X1 - b * X0 ** 2 + c * X0 - 6
r = 10. * (1. - t) * numpy.cos(X0) + 10
Z = u ** 2 + r
fig, ax = pyplot.subplots(subplot_kw={"projection": "3d"})
ax.scatter(X0, X1, model, c="r", label="model")
pyplot.title("Data + model") # or pyplot.savefig("test.png") to save the plot as a .png file instead
You should see something like this:

Appendix. Tips & tricks

This README has a simple ML algorithm. However, Ocean C2D is not limited to usage in ML. The file has examples from vision and other fields.
In the "publish algorithm" step, to replace the sample algorithm with another one:
  • Use one of the standard Ocean algo_dockers images or publish a custom docker image.
  • Use the image name and tag in the container part of the algorithm metadata.
  • The image must have basic support for installing dependencies. E.g. "pip" for the case of Python. You can use other languages, of course.
  • More info is available on the algorithms page
The function to pay_for_compute_service automates order starting, order reusing and performs all the necessary Provider and on-chain requests. It modifies the contents of the given ComputeInput as follows:
  • If the dataset/algorithm contains a transfer_tx_id property, it will try to reuse that previous transfer id. If provider fees have expired but the order is still valid, then the order is reused on-chain.
  • If the dataset/algorithm does not contain a transfer_tx_id or the order has expired (based on the Provider's response), then one new order will be created.
This means you can reuse the same ComputeInput and you don't need to regenerate it everytime it is sent to pay_for_compute_service. This step makes sure you are not paying unnecessary or duplicated fees.
If you wish to upgrade the compute resources, you can use any (paid) C2D environment. Inspect the results of ocean.ocean_compute.get_c2d_environments(service.service_endpoint, DATA_ddo.chain_id) and ocean.retrieve_provider_fees_for_compute(datasets, algorithm_data, consumer_address, compute_environment, duration) for a preview of what you will pay. Don't forget to handle any minting, allowance or approvals on the desired token to ensure transactions pass.
Last modified 1mo ago