powered by ASAP

What You Will Learn

This guide will explain how to do manage datastore values inside the Workflow Executor app.

For the following chapters the Workflow Executor App has to be in the ‚expert mode‘. How to toggle this mode can be found here.

1. What are ‚Datastore Values‘?

In order to make a Workflow run, certain values have to be provided for it (e.g. a path to a file). While the Workflow is being executed, it may create some values itself. All of these values will be shown in the datastore, as seen below:

Datastore View
Figure 1. Datastore view.

2. Datastores inside the Workflow Executor App

Runtime Datastore:

During the execution of a Workflow, its Datastore Values will be loaded into the ‚Runtime Datastore‘, where they will be kept and displayed while the Workflow is being executed. If a workflow aborts and has sub-workflows you can select each individual sub-workflow from a dropdown. The data shown then depends on what data the selected workflow used to read or write its values. As each Workflow has its own separate ‚Runtime Datastore‘, the Datastore Values inside the ‚Runtime Datastore‘ will change!

Workspace Datastore:

The Workspace Datastore is currently only active when a ‚Single Command‘ is about to be run, more about running Single Commands can be found here.
In general the Workspace Datastore holds and displays all the accumulated Datastore Values from already run Workflows or Single Commands.

3. Inspect Values

To inspect the datastore values, simply:

1 double click the desired value row
2 A new window with the value will be displayed
Inspection View
Figure 2. Datastore inspection window.

4. Edit Values

In some cases values lead to unexpected errors that need to be corrected.

4.1. Correct Values of a simple Workflow

A simple workflow is a workflow without any sub-workflows.

Figure 3. Workflow failing at runtime.
Figure 4. Dialog to correct runtime data.
Figure 5. Editor to edit keys with more complex data.
1 When an error occurs, a Correct Runtime Data button appears. Once clicked, a dialog opens to edit the current runtime data.
2 Here you should see the name of the current workflow and nothing else.
3 Here you should be able to see all editable keys.
4 For every key you should see text field where you can edit the value of the key.
5 Every text field contains an edit icon. Clicking this will open an editor for more complex values that are difficult to handle in a regular text field.
6 You can also add new keys with the Add new Variables section.
7 To save your changes you can click on the save button prompting you to confirm your actions.
Figure 6. Correcting data of a failed simple workflow.

4.2. Correct Values of a nested Workflow

A nested workflow uses other so called sub-workflows to do a part of their task. In some cases these can fail because of erroneous data.

You can correct of the main workflow in the same way you can with a simple workflow.

Figure 7. „Dialog to correct data at runtime.
1 A Tree view of the nested workflows. To see any nested workflows of a sub-workflow, you can click on the triangular shape on the left side of the workflow name to expose the nested workflows.
2 All keys visible to the currently selected workflow with their respective value.
The tree view only contains workflows that are a parent of or are an error causing workflow.

An example of workflows not being shown:

Figure 8. Failing workflow with tree view only showing error causing workflows.
1 This model does not fail.
2 This model also does not fail.
3 This model fails.
4 Tree view only shows failing workflow and its parent.

When a sub-workflow is the cause of the failure, you’ll need to select the exact workflow the error occurred in.

Figure 9. Steps to correct data at runtime.
1 Select the exact workflow that failed.
2 Edit the key that needs adjustments.
Changing a value only changes it in the currently selected workflow!

This is how editing data of a nested workflow can look like:

Figure 10. Correcting data of a failed nested workflow.

Sonatype Nexus

PAK features connectors and commands for Sonatype Nexus. This means the software can directly interact with Nexus repositories for storing and managing artifacts. Through these connectors, PAK can automate tasks like uploading binaries or retrieving dependencies, ensuring efficient artifact management within Nexus.


PAK has connectors and commands for Jenkins. This allows the software to directly communicate with Jenkins servers, enabling the automation of CI/CD (Continuous Integration/Continuous Deployment) tasks. Through these connectors, PAK can trigger builds, fetch build statuses, or manage job configurations, streamlining the CI/CD processes within Jenkins.

Git Hub

PAK possesses connectors and commands for GitHub. This means the software can interface directly with GitHub repositories, facilitating actions like code pushes, pull requests, or issue tracking. Through these connectors, PAK can automate various GitHub operations, enhancing code collaboration and repository management.

Atlassian Confluence

PAK is equipped with connectors and commands for Atlassian Confluence. This enables the software to directly interact with Confluence spaces and pages. Through these connectors, PAK can automate actions such as creating, updating, or retrieving documentation, ensuring efficient content management and collaboration within Confluence.


PAK features connectors and commands for Codebeamer. This allows the software to seamlessly integrate with Codebeamer’s ALM (Application Lifecycle Management) platform. Through these connectors, PAK can automate tasks like issue tracking, test management, or requirements tracing, enhancing the coordination and management of software development processes.

JFrog Artifactory

PAK has connectors and commands for JFrog Artifactory. This means the software can directly interface with Artifactory repositories, enabling actions like artifact storage, retrieval, and management. Through these connectors, PAK can automate tasks such as deploying artifacts or managing repository configurations, streamlining the integration and management of binary artifacts within Artifactory.

Amazon Web Services (AWS)

PAK has connectors and commands for Amazon Web Services (AWS). This means the software possesses specialized interfaces to directly interact with AWS services and execute actions on the AWS platform. Through these connectors, PAK can automate AWS-specific commands, such as launching EC2 instances, managing S3 buckets, or configuring Lambda functions. This allows for efficient integration, management, and automation of AWS resources and services directly from PAK.

Atlassian Jira

PAK features integration tools and capabilities for Atlassian Jira. These tools allow for a direct connection to Jira and the execution of specific actions. Using these integration tools, PAK can automate Jira actions such as adding comments or changing ticket priorities, ensuring seamless handling and coordination of Jira processes.


PAK has connectors and commands for Git. This means it has interfaces to directly communicate with Git and execute actions. Through these connectors, the software can automate Git commands such as retrieving changes or creating branches, enabling efficient integration and management of Git tasks.

Generic Human Tasks

PAK offers you a standard set of commands which require creative input from the user. Enables you to start with automating your workflows, that still need abit of human input.

Generic Commands

PAK offers a standard set of commands giving you the first steps to automate your workflows.

Nexus Maven Command Pool

Nexus is an artifact repository manager for storing binaries, libraries, and artifacts, supporting formats like Maven. Maven, a software project management tool, is based on the Project Object Model (POM) and allows developers to consistently define projects and dependencies. Our Command Pool offers commands for interactions between Maven and Nexus, such as artifact uploads or dependency retrieval.

Artifactory Maven Command Pool

Artifactory allows developers to store, retrieve, and manage binary files and artifacts, providing a
central source for all binaries used in a development process. Apache Maven is a software project
management and comprehension tool that enables developers to consistently describe a project and
its dependencies. Our Command Pool offers a collection of commands used to facilitate interactions
between Maven and Artifactory, such as uploading artifacts or retrieving dependencies.

Open API Command Interpreter

The OpenApi Command Interpreter allows you the automatic parsing of commands from an OpenApi defintion. No additional code needs to be written anymore, just add the address to the definition and our framework does the rest!

Kotlin Command Interpreter

The Kotlin Command Interpreter allows you the parsing and execution of commands within a Kotlin environment to automate various tasks or processes.

Bpmn Interpreter

Workflows come in many shapes and forms. The BPMN (Business Process Model and Notation) Interpreter enables the parsing of worklows defined in the BPMN format into the PAK intern model.

Human Task Interpreter

The Human Task Interpreter allows you the parsing and running of commands within a HTML and Javascript environment. Use this to build commands which need the creative input of a workflow user!

Java Command Interpreter

The Java Command Interpreter allows you the parsing and execution of commands within a Java
environment to automate various tasks or processes.


The heart of the PAK-Framework. Contains the means to run workflows with the PAK engine, but also the possibility to enrich the frameworks interfaces with your own implementations and solutions.

RocksDB Persistence

Data that is generated by a workflow run needs to be saved for short or longer terms. Our solution to the Persistence Interface of the PAK-Framework is to use the high-performance, key-value based RocksDB developed by Facebook.

PAK online

PAK Online is a web based application and provides an Open API based REST API. It enables you to upload workflows and run them periodically or on REST demand.

Command Line App

Run tasks and workflows on the console or as part of a CI/CD Pipeline with our Command Line Interface.

Workflow Editor

With our specially developed editor, a wide variety of workflows can be easily modeled in the wide known BPMN process format.

Workflow Executor

The Workflow Executor is the application to run your workflows. It features a multilingual UI and easy managment of your favorite workflows.


We offer a community website where you can exchange ideas and support each other. For our Pro packages we also offer full support via email.