Nodeum Docs
HomepageBlog
  • ✨What is Nodeum?
    • Data Management Software
  • 👣GETTING STARTED
    • Step by Step
  • 🏨ARCHITECTURE
    • Standalone
    • High Availability
    • Fully Scalable & Resilience
  • Install and Deploy Nodeum
    • Platform Support
    • Nodeum v1
      • Ansible based installation
    • Nodeum v2 - Data Mover
      • Ansible based Installation
        • Customize your Installation
      • Docker Based Deployment
    • SLURM Integration
    • Software License Application
  • Interfaces
    • ND Client
    • Console
      • Manual for Administrator
        • Login
        • Dashboard & Reports
        • Catalog
        • Data Mover Management
        • Advanced Task Management
        • Data Container
        • Primary Storage Configuration
        • Pool Management
        • TCO Calculator
        • Toolbox
        • System Settings
          • Information
          • Configuration
          • Date & Time
          • Backup
          • Services
          • Hostname and DNS Configuration
          • NAS Storage Configuration
          • Object Storage Configuration
          • Tape Library Configuration
          • User Management
          • Audits
      • Manual for End User
    • For Developers
      • RESTful API
      • Configuration through RestAPI Console
      • Software Developement Kits (SDK)
      • Nodeum API
        • API Terms of Use
        • release v1.x
        • release v2.x
  • DATA MOVER
    • Policy-Based Task orchestration
      • Pool Management
      • Scheduler
      • Data Integrity
      • Priority Management
      • Filtering (Basic or Advanced)
      • Hook service (callback)
    • Content traceability
    • Metadata Management
  • IDENTITY MANAGEMENT
    • Right - Authentication & Authorization
    • LDAP Plugin for JWT Token
  • Container Configuration
    • Prerequisites
    • About Container
    • Authorization and Authentication
    • Access your Container
  • HYBRID STORAGE MANAGEMENT
    • File System Managment
    • Object Storage Management
      • Customize your S3 connection
    • Tape Library Management
      • Tape Writing Format : LTFS
      • Tape Compression
      • Tape Rehydratation
      • Import a LTFS Tape
      • Task Maintenance for Tapes
  • ⏰Alert & Monitoring
    • Alerts
    • Monitoring
    • Log Management
  • 🏥Recover after Hardware Failure
    • Failover - Active/Passive
    • Failover Procedure - One Site
    • Backup & Restore
  • 🔐Security Guide
    • Advanced Network Configuration
    • Add a SSL Certificate on Web Console
    • Enable SSL Certificate Container accessible on S3 Protocol
  • Compatibility guide
    • Software requirement
    • Supported Storage
  • PRODUCT SPECIFICATIONS
    • Character Set Support
    • Limitations
    • Files Status
    • Task Status
Powered by GitBook
On this page
  • Type of Workflows
  • Execution Modes
  • Task Action
  • Extended Data Set Selector
  • Path Management
  • Result of Task execution

Was this helpful?

  1. DATA MOVER

Policy-Based Task orchestration

Last updated 26 days ago

Was this helpful?

Manages and coordinates data movement tasks based on predefined policies or rules. This allows the management of an unlimited number of tasks, including their execution, but also monitoring and reporting aspects.

Type of Workflows

Workflow Manager allows the creation of different data movement tasks:

  • Copy

  • Move

  • Erase

  • Scan

Different type of policies are available:

  • Data Migration

  • Data Archiving (Active Archive)

  • Data Protection (Offline Archive)

Currently, the two workflows Data Archiving and Data Protection are available in the , when the Data Migration is available in the .

Each task includes the definition of the source storage, with its dataset and the destination storage.

Execution Modes

It is possible to define multiple tasks, they are executed in parallel mode in a multi-threaded system.

Each task will execute the movement of a list of files (1 or n) and segment the processing into 3 main parts:

  • Preparation phase

  • Execution phase

  • Finalization phase

For each task execution process, specific jobs are created. The default segmentation is defined to optimize the data processing (example of a default setting: create a job of 1 GB or 100 files maximum).

Each job will be sent into a different queue system. A dispatching engine will coordinate the processing of each job in regard to the defined priority.

Then, different workers are created to do the copy of files and to do the reporting to Finalization process about the status of the copy.

The workflow manager can execute these 3 phases in two main different modes:

  • sequential processing

  • parallel processing

The Data Mover always takes into account the remaining space of the destination storage. This means that if the available capacity is insufficient to store the file being copied, the operation will either result in an error message such as "stopped by system" (in the case of NAS and file systems), or it will request a new tape (for tape storage). In any case, files are never split across multiple media.

Task Action

Each running task can be paused or stopped. If a task is paused, it can always be restarted. Pressing the stop button will stop the task. Once the stop button has been pressed, it may take a few seconds or minutes before the task is stopped. This is because the task always finishes the current job under processing.

Extended Data Set Selector

An unlimited data set selector brings granularity to the data you have to select. Movement can be executed between any type of storage from storage where hot data are stored to cold storage.

Path Management

For any type of data movement, the Workflow Manager always maintains the original path where the contents were originally stored.

Business applications are configured to access the same folders, even if data have been moved. It is why it is very important that when data are retrieved, they are recopied into their original folder.

Meanwhile, Nodeum offers different options to change this behavior and provides solutions to relocate the contents into another folder structure.

Result of Task execution

Each task finishes with a status, each task status determines the result of the processing that has been executed. This following describes each type of status.

table