Skip to main content

Data Transfer Guide: EOS ↔ PFS

This guide explains how to transfer data from EOS to PFS or from PFS to EOS using Argo Workflows.

You can use these workflows to schedule or automate periodic data movement jobs between your object storage buckets and parallel filesystem volumes.

Prerequisites

Before running any workflow:

  • You must have an existing EOS bucket. To create an EOS, follow this link

  • You must have a PFS filesystem created and mounted to an instance. To create a PFS, follow this link

  • Ensure your access credentials:

    • bucket_access_key

    • bucket_secret_key

Bucket Overview

info

You can view the Bucket Name in the Dataset’s Overview tab.

Transfer Data: PFS → EOS

This workflow uploads files from your PFS to your EOS bucket.

Required Parameters

ParameterDescription
source_fs_nameName of the filesystem
source_pathPath inside the filesystem to upload
destination_bucket_nameEOS bucket name
destination_pathDestination folder inside the bucket
bucket_endpoint_urlEOS endpoint (default included)
bucket_access_keyYour EOS access key
bucket_secret_keyYour EOS secret key

Create Pipeline

Download fs-to-eos.yaml

Now create a pipeline using the fs-to-eos.yaml file. For more details about creating pipeline follow this link

note

Enter the correct parameters in the YAML file, such as source_fs_name, source_path, and others. You may also fill them in manually while creating a run for this pipeline.

Upload Yaml

Now create a run for the pipeline. While creating the run, make sure to provide the correct parameters; otherwise, they will be automatically fetched from the YAML file if already specified.

For more details about creating a run for the pipeline, follow this link

Parameters

Once the run completes successfully, you will see the following logs.

Logs

Transfer Data: EOS → PFS

This workflow downloads data from your EOS bucket into your PFS filesystem.

Required Parameters

ParameterDescription
source_bucket_nameEOS bucket name
source_pathFolder/object prefix in the bucket
destination_fs_nameTarget PFS filesystem name
destination_pathFolder inside filesystem to store data
bucket_endpoint_urlEOS endpoint
bucket_access_keyEOS access key
bucket_secret_keyEOS secret key

Create Pipeline

Download eos-to-fs.yaml

Now create a pipeline using the eos-to-fs.yaml file. For more details about creating pipeline, follow this link

note

Enter the correct parameters in the YAML file, such as source_bucket_name, source_path, and others. You may also fill them in manually while creating a run for this pipeline.

Upload Yaml

Now create a run for the pipeline. While creating the run, make sure to provide the correct parameters; otherwise, they will be automatically fetched from the YAML file if already specified.

For more details about creating a run for the pipeline, follow this link

Parameters

Once the run completes successfully, you will see the following logs.

Logs