Suche

1. What you will learn

In this short summary you will get an idea on how data in PAK is persisted.

2. Basics

The IPersistenceService-interface is the main component involved in storing and persisting your data. While an Engine is running, the Persistence Service functions as the datastore, making it the direct entry point for all data created or consumed by any Command. The Persistence Service should essentially function as a key-value database.

3. Signature

Before giving a short example on how the interface should be used, this section describes the signature of the IPersistenceService interface, as well as an overview on which methods have to be implemented.

3.1. Transaction Model

The interface offers the ability to implement the commonly known ACID-principle by declaring different methods based on transactions.

void begin() throws PersistenceException;

Begins a new transaction in the current Persistence; Should always be thread dependent

void commit() throws PersistenceException;

Finishes the current transaction by committing all changes of the current thread

void rollback() throws PersistenceException;

Undoes all changes made in the current transaction and finishes it.

While you can choose not to follow the transaction model, be advised that your implementation should ensure thread-safety

3.2. Data Processing

The interface offers two operations relating to data processing. find for looking up values stored in the datastore and store in order to populate it.

Object find(final String key);

Tries retrieving the object for the specified key from the Persistence, returns null if no value was found.

Map<String, Object> find(final Collection<String> key);

Tries retrieving all objects for the specified keys from the Persistence, if no value was found for a key, there will be no entry in the resulting map.

void store(final String key, final Object value);

Stores the given object for the specified key in the Persistence.

void store(final Map<String, ?> keyValues);

Stores all given entries in the Persistence.

You must implement the find and store method for single values. By default, the respective operations which take a map as an input only call your single element operations for every element of the map.

If the need for data deletion arises, implement this functionality for the store operation using IPersistenceService.InitialValue.class

4. Usage

This section briefly outlines a few examples on how an implementing Persistence should be used.

4.1. Accessing Data

private final IPersistenceService persistence = ...;

public void storeExample() {
	this.persistence.begin(); (1)
	this.persistence.store("testKey", "Hello World"); (2)
	this.persistence.commit(); (3)
}

public Object retrieveExample() {
    this.persistence.begin(); (1)
    try {
    	return this.persistence().find("testKey"); (2)
    } finally {
    	this.persistence.commit(); (3)
    }
}
1 Before we perform actions on our Persistence we should start a new transaction, so possible mistakes may be rolled back
2 After we initialized our transaction we can either store or retrieve a value to/from the Persistence
3 Finally, when we are finished editing, we commit the transaction, thus closing it and saving the state

4.2. Data Deletion

The deletion of data is not a built-in specification as the deleting of data from the datastore should hardly be necessary. However, if you need the functionality the IPersistenceService offers a marker class for that reason:

/** marker class */
final class InitialValue {
    private InitialValue() {
        // private constructor this is only used via InitialValue.class
    }
}

Let’s suppose we want to delete the object behind the key superSecretVar. We simply call store using our marker class like

persistence.store("superSecretVar", IPersistenceService.InitialValue.class);

4.3. Data Transformation

All data that is written to the Persistence Service will be in the JSON format by default. However, should your implementation be dependent on another data format, do not solve this in your Persistence, but rather use the designated interface IDataTransformer.

5. Default Implementations

There are two default implementations of the IPersistenceService offered by PAK.

  • The SimplePersistenceService: Functions as an in-memory Persistence which can be used analogous to a map. It is mostly used for prototyping and testing (Dependency: de.asap.pak.core:simple)

  • The RocksDBPersistence: Functions as a bridge between PAK and the RocksDB implementation. (Dependency: de.asap.pak.extra:pak-rocksdb)

Sonatype Nexus

PAK features connectors and commands for Sonatype Nexus. This means the software can directly interact with Nexus repositories for storing and managing artifacts. Through these connectors, PAK can automate tasks like uploading binaries or retrieving dependencies, ensuring efficient artifact management within Nexus.

Jenkins

PAK has connectors and commands for Jenkins. This allows the software to directly communicate with Jenkins servers, enabling the automation of CI/CD (Continuous Integration/Continuous Deployment) tasks. Through these connectors, PAK can trigger builds, fetch build statuses, or manage job configurations, streamlining the CI/CD processes within Jenkins.

Git Hub

PAK possesses connectors and commands for GitHub. This means the software can interface directly with GitHub repositories, facilitating actions like code pushes, pull requests, or issue tracking. Through these connectors, PAK can automate various GitHub operations, enhancing code collaboration and repository management.

Atlassian Confluence

PAK is equipped with connectors and commands for Atlassian Confluence. This enables the software to directly interact with Confluence spaces and pages. Through these connectors, PAK can automate actions such as creating, updating, or retrieving documentation, ensuring efficient content management and collaboration within Confluence.

Codebeamer

PAK features connectors and commands for Codebeamer. This allows the software to seamlessly integrate with Codebeamer’s ALM (Application Lifecycle Management) platform. Through these connectors, PAK can automate tasks like issue tracking, test management, or requirements tracing, enhancing the coordination and management of software development processes.

JFrog Artifactory

PAK has connectors and commands for JFrog Artifactory. This means the software can directly interface with Artifactory repositories, enabling actions like artifact storage, retrieval, and management. Through these connectors, PAK can automate tasks such as deploying artifacts or managing repository configurations, streamlining the integration and management of binary artifacts within Artifactory.

Amazon Web Services (AWS)

PAK has connectors and commands for Amazon Web Services (AWS). This means the software possesses specialized interfaces to directly interact with AWS services and execute actions on the AWS platform. Through these connectors, PAK can automate AWS-specific commands, such as launching EC2 instances, managing S3 buckets, or configuring Lambda functions. This allows for efficient integration, management, and automation of AWS resources and services directly from PAK.

Atlassian Jira

PAK features integration tools and capabilities for Atlassian Jira. These tools allow for a direct connection to Jira and the execution of specific actions. Using these integration tools, PAK can automate Jira actions such as adding comments or changing ticket priorities, ensuring seamless handling and coordination of Jira processes.

Git

PAK has connectors and commands for Git. This means it has interfaces to directly communicate with Git and execute actions. Through these connectors, the software can automate Git commands such as retrieving changes or creating branches, enabling efficient integration and management of Git tasks.

Generic Human Tasks

PAK offers you a standard set of commands which require creative input from the user. Enables you to start with automating your workflows, that still need abit of human input.

Generic Commands

PAK offers a standard set of commands giving you the first steps to automate your workflows.

Nexus Maven Command Pool

Nexus is an artifact repository manager for storing binaries, libraries, and artifacts, supporting formats like Maven. Maven, a software project management tool, is based on the Project Object Model (POM) and allows developers to consistently define projects and dependencies. Our Command Pool offers commands for interactions between Maven and Nexus, such as artifact uploads or dependency retrieval.

Artifactory Maven Command Pool

Artifactory allows developers to store, retrieve, and manage binary files and artifacts, providing a
central source for all binaries used in a development process. Apache Maven is a software project
management and comprehension tool that enables developers to consistently describe a project and
its dependencies. Our Command Pool offers a collection of commands used to facilitate interactions
between Maven and Artifactory, such as uploading artifacts or retrieving dependencies.

Open API Command Interpreter

The OpenApi Command Interpreter allows you the automatic parsing of commands from an OpenApi defintion. No additional code needs to be written anymore, just add the address to the definition and our framework does the rest!

Kotlin Command Interpreter

The Kotlin Command Interpreter allows you the parsing and execution of commands within a Kotlin environment to automate various tasks or processes.

Bpmn Interpreter

Workflows come in many shapes and forms. The BPMN (Business Process Model and Notation) Interpreter enables the parsing of worklows defined in the BPMN format into the PAK intern model.

Human Task Interpreter

The Human Task Interpreter allows you the parsing and running of commands within a HTML and Javascript environment. Use this to build commands which need the creative input of a workflow user!

Java Command Interpreter

The Java Command Interpreter allows you the parsing and execution of commands within a Java
environment to automate various tasks or processes.

Core

The heart of the PAK-Framework. Contains the means to run workflows with the PAK engine, but also the possibility to enrich the frameworks interfaces with your own implementations and solutions.

RocksDB Persistence

Data that is generated by a workflow run needs to be saved for short or longer terms. Our solution to the Persistence Interface of the PAK-Framework is to use the high-performance, key-value based RocksDB developed by Facebook.

PAK online

PAK Online is a web based application and provides an Open API based REST API. It enables you to upload workflows and run them periodically or on REST demand.

Command Line App

Run tasks and workflows on the console or as part of a CI/CD Pipeline with our Command Line Interface.

Workflow Editor

With our specially developed editor, a wide variety of workflows can be easily modeled in the wide known BPMN process format.

Workflow Executor

The Workflow Executor is the application to run your workflows. It features a multilingual UI and easy managment of your favorite workflows.

Support

We offer a community website where you can exchange ideas and support each other. For our Pro packages we also offer full support via email.