Workflow is a distributed event driven workflow framework that runs robust, durable, and scalable sequential business logic on your services.
Workflow uses a RoleScheduler to distribute the work across your instances through a role assignment process (similar to a leadership election process, but with more than a single role of leader).
Workflow expects to be run on multiple instances but can also be run on single instances. Using the above-mentioned RoleScheduler, Workflow is able to make sure each process only runs once at any given time regardless if you are running 40 instances of your service or 1 instance.
To start using workflow you will need to add the workflow module to your project. You can do this by running:
go get github.com/luno/workflow
Adapters enable Workflow to be tech stack agnostic by placing an interface / protocol between Workflow and the tech stack. Workflow uses adapters to understand how to use that specific tech stack.
For example, the Kafka adapter enables workflow to produce messages to a topic as well as consume them from a topic using a set of predefined methods that wrap the kafka client. Reflex is an event streaming framework that works very differently to Kafka and the adapter pattern allows for the differences to be contained and localised in the adapter and not spill into the main implementation.
The EventStreamer adapter interface defines what is needed to be satisfied in order for an event streaming platform or framework to be used by Workflow.
All implementations of the EventStreamer interface should be tested using adaptertest.TestEventStreamer
The RecordStore adapter interface defines what is needed to satisfied in order for a storage solution to be used by Workflow.
All implementations of the RecordStore interface should be tested using adaptertest.RunRecordStoreTest
The RoleScheduler adapter interface defines what is needed to satisfied in order for a role scheduling solution to be used by Workflow.
All implementations of the RoleScheduler interface should be tested using adaptertest.RunRoleSchedulerTest
There are more adapters available but only the above 3 are core requirements to use Workflow. To start, use the in-memory implementations as that is the simplest way to experiment and get used to Workflow. For testing other adapter types be sure to look at adaptertest which are tests written for adapters to ensure that they meet the specification.
Adapters, except for the in-memory implementations, don’t come with the core Workflow module such as kafkastreamer
, reflexstreamer
, sqlstore
,
sqltimeout
, rinkrolescheduler
, webui
and many more. If you wish to use these you need to add them individually
based on your needs or build out your own adapter.
go get github.com/luno/workflow/adapters/kafkastreamer
go get github.com/luno/workflow/adapters/reflexstreamer
go get github.com/luno/workflow/adapters/sqlstore
go get github.com/luno/workflow/adapters/sqltimeout
go get github.com/luno/workflow/adapters/rinkrolescheduler
go get github.com/luno/workflow/adapters/webui
Connectors allow Workflow to consume events from an event streaming platform or framework and either trigger a workflow run or provide a callback to the workflow run. This means that Connectors can act as a way for Workflow to connect with the rest of the system.
Connectors are implemented as adapters as they would share a lot of the same code as implementations of an EventStreamer and can be seen as a subsection of an adapter.
An example can be found here.
package usage
import (
"context"
"github.com/luno/workflow"
)
type Step int
func (s Step) String() string {
switch s {
case StepOne:
return "One"
case StepTwo:
return "Two"
case StepThree:
return "Three"
default:
return "Unknown"
}
}
const (
StepUnknown Step = 0
StepOne Step = 1
StepTwo Step = 2
StepThree Step = 3
)
type MyType struct {
Field string
}
func Workflow() *workflow.Workflow[MyType, Step] {
b := workflow.NewBuilder[MyType, Step]("my workflow name")
b.AddStep(StepOne, func(ctx context.Context, r *workflow.Run[MyType, Step]) (Step, error) {
r.Object.Field = "Hello,"
return StepTwo, nil
}, StepTwo)
b.AddStep(StepTwo, func(ctx context.Context, r *workflow.Run[MyType, Step]) (Step, error) {
r.Object.Field += " world!"
return StepThree, nil
}, StepThree)
return b.Build(...)
}
---
title: The above defined workflow creates the below Directed Acyclic Graph
---
stateDiagram-v2
direction LR
[*]-->One
One-->Two
Two-->Three
Three-->[*]
wf := usage.Workflow()
ctx := context.Background()
wf.Run(ctx)
Stop: To stop all processes and wait for them to shut down correctly call
wf.Stop()
foreignID := "82347982374982374"
runID, err := wf.Trigger(ctx, foreignID, StepOne)
if err != nil {
...
}
Awaiting results: If appropriate and desired you can wait for the workflow to complete. Using context timeout (cancellation) is advised.
foreignID := "82347982374982374"
runID, err := wf.Trigger(ctx, foreignID, StepOne)
if err != nil {
...
}
ctx, cancel := context.WithTimeout(ctx, 10 * time.Second)
defer cancel()
record, err := wf.Await(ctx, foreignID, runID, StepThree)
if err != nil {
...
}
Head on over to ./_examples to get familiar with callbacks, timeouts, testing, connectors and more about the syntax in depth 😊
When a Workflow is triggered it creates an individual workflow instance called a Run. This is represented as workflow.Run in Workflow. Each run has a lifecycle which is a finite set of states - commonly referred to as Finite State Machine. Each workflow Run has the following of states (called RunState in Workflow):
Run State | Value (int) | Description |
---|---|---|
Unknown | 0 | Has no meaning. Protects against default zero value. |
Initiated | 1 | State assinged at creation of Run and is yet to be processed. |
Running | 2 | Has begun to be processed and is currently still being processed by a step in the workflow. |
Paused | 3 | Temporary stoppage that can be resumed or cancelled. Will prevent any new triggers of the same Foreign ID. |
Completed | 4 | Finished all the steps configured at time of execution. |
Cancelled | 5 | Did not complete all the steps and was terminated before completion. |
Data Deleted | 6 | Run Object has been modified to remove data or has been entirely removed. Likely for PII scrubbing reasons. |
Requested Data Deleted | 7 | Request state for the workflow to apply the default or custom provided delete operation to the Run Object. |
A Run can only exist in one state at any given time and the RunState allows for control over the Run.
---
title: Diagram the run states of a workflow
---
stateDiagram-v2
direction LR
Initiated-->Running
Running-->Completed
Running-->Paused
Paused-->Running
Running --> Cancelled
Paused --> Cancelled
state Finished {
Completed --> RequestedDataDeleted
Cancelled --> RequestedDataDeleted
DataDeleted-->RequestedDataDeleted
RequestedDataDeleted-->DataDeleted
}
Hooks allow for you to write some functionality for Runs that enter a specific RunState. For example when
using PauseAfterErrCount
the usage of the OnPause hook can be used to send a notification to a team to notify
them that a specific Run has errored to the threshold and now has been paused and should be investigated. Another
example is handling a known sentinel error in a Workflow Run and cancelling the Run by calling (where r is *Run)
r.Cancel(ctx) or if a Workflow Run is manually cancelled from a UI then a notifgication can be sent to the team for visibility.
Hooks run in an event consumer. This means that it will retry until a nil error has been returned and is durable across deploys and interruptions. At-least-once delivery is guaranteed, and it is advised to use the RunID as an idempotency key to ensure that the operation is idempotent.
Hook | Parameter(s) | Return(s) | Description | Is Event Driven? |
---|---|---|---|---|
OnPause | workflow.RunStateChangeHookFunc | error | Fired when a Run enters RunStatePaused | Yes |
OnCancelled | workflow.RunStateChangeHookFunc | error | Fired when a Run enters RunStateCancelled | Yes |
OnCompleted | workflow.RunStateChangeHookFunc | error | Fired when a Run enters RunStateCompleted | Yes |
This package provides several options to configure the behavior of the workflow process. You can use these options to customize the instance count, polling frequency, error handling, lag settings, and more. Each option is defined as a function that takes a pointer to an options
struct and modifies it accordingly. Below is a description of each available option:
ParallelCount
func ParallelCount(instances int) Option
instances
: The total number of parallel instances to create.b.AddStep(
StepOne,
...,
StepTwo,
).WithOptions(
workflow.ParallelCount(5)
)
PollingFrequency
func PollingFrequency(d time.Duration) Option
d
: The polling frequency as a time.Duration
.b.AddStep(
StepOne,
...,
StepTwo,
).WithOptions(
workflow.PollingFrequency(10 * time.Second)
)
ErrBackOff
func ErrBackOff(d time.Duration) Option
d
: The backoff duration as a time.Duration
.b.AddStep(
StepOne,
...,
StepTwo,
).WithOptions(
workflow.ErrBackOff(5 * time.Minute)
)
LagAlert
func LagAlert(d time.Duration) Option
d
: The duration of the lag alert as a time.Duration
.b.AddStep(
StepOne,
...,
StepTwo,
).WithOptions(
workflow.LagAlert(15 * time.Minute),
)
ConsumeLag
func ConsumeLag(d time.Duration) Option
d
: The lag duration as a time.Duration
.b.AddStep(
StepOne,
...,
StepTwo,
).WithOptions(
workflow.ConsumeLag(10 * time.Minute),
)
PauseAfterErrCount
func PauseAfterErrCount(count int) Option
RunStatePaused
. This mechanism acts similarly to a Dead Letter Queue, preventing further processing of problematic records and allowing for investigation and retry.count
: The maximum number of errors before pausing.b.AddStep(
StepOne,
...,
StepTwo,
).WithOptions(
workflow.PauseAfterErrCount(3),
)
Term | Description |
---|---|
Builder | A struct type that facilitates the construction of workflows. It provides methods for adding steps, callbacks, timeouts, and connecting workflows. |
Callback | A method in the workflow API that can be used to trigger a callback function for a specified status. It passes data from a reader to the specified callback function. |
Consumer | A component that consumes events from an event stream. In this context, it refers to the background consumer goroutines launched by the workflow. |
EventStreamer | An interface representing a stream for workflow events. It includes methods for producing and consuming events. |
Graph | A representation of the workflow’s structure, showing the relationships between different statuses and transitions. |
Hooks | An event driven process that take place on a Workflow’s Run’s lifecycle defined in a finite number of states called RunState. |
Producer | A component that produces events to an event stream. It is responsible for sending events to the stream. |
Record | Is the “wire format” and representation of a Run that can be stored and retrieved. The RecordStore is used for storing and retrieving records. |
RecordStore | An interface representing a store for Record(s). It defines the methods needed for storing and retrieving records. The RecordStore’s underlying technology must support transactions in order to prevent dual-writes. |
RoleScheduler | An interface representing a scheduler for roles in the workflow. It is responsible for coordinating the execution of different roles. |
Run | A Run is the representation of the instance that is created and processed by the Workflow. Each time Trigger is called a new “Run” is created. |
RunState | RunState defines the finite number of states that a Run can be in. This is used to control and monitor the lifecycle of Runs. |
Topic | A method that generates a topic for producing events in the event streamer based on the workflow name and status. |
Trigger | A method in the workflow API that initiates a workflow for a specified foreignID and starting status. It returns a Run ID and allows for additional configuration options. |