A lightweight yet powerful TypeScript library for building and executing DAG (Directed Acyclic Graph) pipelines of tasks. Provides flexible task orchestration, persistent storage, workflow management, and error handling for complex task execution scenarios.
- Installation
- Quick Start
- Core Concepts
- Creating Custom Tasks
- Building Task Graphs
- Using Workflows
- Storage and Caching
- Error Handling
- Advanced Patterns
- API Reference
- Examples
- Testing
- License
npm install @workglow/task-graph
# or
bun add @workglow/task-graph
# or
yarn add @workglow/task-graphHere's a simple example that demonstrates the core concepts:
import { Task, TaskGraph, Dataflow, Workflow } from "@workglow/task-graph";
import { DataPortSchema } from "@workglow/util";
// 1. Define a custom task
class MultiplyBy2Task extends Task<{ value: number }, { result: number }> {
static readonly type = "MultiplyBy2Task";
static readonly category = "Math";
static readonly title = "Multiply by 2";
static readonly description = "Multiplies a number by 2";
static inputSchema() {
return {
type: "object",
properties: {
value: {
type: "number",
description: "Input number",
},
},
required: ["value"],
additionalProperties: false,
} as const satisfies DataPortSchema;
}
static outputSchema() {
return {
type: "object",
properties: {
result: {
type: "number",
description: "Multiplied result",
},
},
required: ["result"],
additionalProperties: false,
} as const satisfies DataPortSchema;
}
async execute(input: { value: number }) {
return { result: input.value * 2 };
}
}
// 2. Use the Task
// 2.1 Use it directly
const task = new MultiplyBy2Task({ value: 15 });
const result = await task.run();
console.log(result); // { result: 30 }
// 2.2 Use it with TaskGraph
const graph = new TaskGraph();
graph.addTask(new MultiplyBy2Task({ value: 15 }, { id: "multiply1" }));
graph.addTask(new MultiplyBy2Task({}, { id: "multiply2" }));
graph.addDataflow(new Dataflow("multiply1", "result", "multiply2", "value"));
const results = await graph.run();
console.log(results); // [{ id: "multiply1", data: { result: 60 } }]
// 2.3 With Workflow
const wf = new Workflow();
wf.addTask(new MultiplyBy2Task({ value: 15 }));
wf.addTask(new MultiplyBy2Task()); // auto-connects to previous task
const result = await wf.run();
console.log(result); // { result: 60 }
// 2.3.1 Adding to Workflow
import { CreateWorkflow } from "@workglow/task-graph";
declare module "@workglow/task-graph" {
interface Workflow {
multiplyBy2: CreateWorkflow<{ value: number }>;
}
}
Workflow.prototype.multiplyBy2 = CreateWorkflow(MultiplyBy2Task);
const wf = new Workflow();
wf.multiplyBy2({ value: 15 });
wf.multiplyBy2(); // input is output from previous task
const result = await wf.run();
console.log(result); // { result: 60 }
// 2.3 Create a helper function
export const MultiplyBy2 = (input: { value: number }) => {
return new MultiplyBy2Task().run(input);
};
const first = await MultiplyBy2({ value: 15 });
const second = await MultiplyBy2({ value: first.result });
console.log(second); // { result: 60 }Tasks are the fundamental units of work. Each task:
- Defines input/output schemas using JSON Schema (from
@workglow/util), TypeBox, or Zod - Implements
execute()for main logic orexecuteReactive()for UI updates - Has a unique type identifier and category
- Can be cached based on inputs
- Emits lifecycle events
TaskGraph is the low-level API for building directed acyclic graphs of tasks:
- Manages tasks and their dependencies
- Handles execution order and parallelization
- Provides detailed control over data flow
- Returns results as an array of task outputs
Data flows between tasks through Dataflow objects that specify:
- Source task and output port
- Target task and input port
- Data transformation and validation
- Error propagation
- Edges in the graph
Workflow is the high-level API that provides:
- Builder pattern for easier task composition
- Automatic task connection based on compatible input/output types
- Pipeline operations (
pipe,parallel) - Simplified result handling
- Event management
You can define schemas using plain JSON Schema, TypeBox, or Zod. Here are examples of each approach:
import { Task, IExecuteContext } from "@workglow/task-graph";
import { DataPortSchema, FromSchema } from "@workglow/util";
const MyInputSchema = {
type: "object",
properties: {
text: {
type: "string",
description: "Text to process",
},
multiplier: {
type: "number",
description: "Repeat multiplier",
default: 1,
},
},
required: ["text"],
additionalProperties: false,
} as const satisfies DataPortSchema;
type MyInput = FromSchema<typeof MyInputSchema>;
// Equivalent to:
// type MyInput = {
// text: string;
// multiplier?: number;
// };
const MyOutputSchema = {
type: "object",
properties: {
processed: {
type: "string",
description: "Processed text",
},
length: {
type: "number",
description: "Text length",
},
},
required: ["processed", "length"],
additionalProperties: false,
} as const satisfies DataPortSchema;
type MyOutput = FromSchema<typeof MyOutputSchema>;
class TextProcessorTask extends Task<MyInput, MyOutput> {
static readonly type = "TextProcessorTask";
static readonly title = "Text Processor";
static readonly description = "Processes text";
static readonly category = "Text Processing";
static readonly cacheable = true;
static inputSchema() {
return MyInputSchema;
}
static outputSchema() {
return MyOutputSchema;
}
async execute(input: MyInput, context: IExecuteContext): Promise<MyOutput> {
const { text, multiplier = 1 } = input;
const { signal, updateProgress } = context;
if (signal?.aborted) {
throw new TaskAbortedError("Task was cancelled");
}
await updateProgress(0.5, "Processing text...");
await new Promise((resolve) => setTimeout(resolve, 100));
const processed = text.repeat(multiplier);
return {
processed,
length: processed.length,
};
}
}TypeBox schemas are JSON Schema compatible and can be used directly:
import { Task, IExecuteContext } from "@workglow/task-graph";
import { Type } from "@sinclair/typebox";
import { DataPortSchema, FromSchema } from "@workglow/util";
const MyInputSchema = Type.Object({
text: Type.String({ description: "Text to process" }),
multiplier: Type.Optional(Type.Number({ description: "Repeat multiplier", default: 1 })),
}) satisfies DataPortSchema;
type MyInput = FromSchema<typeof MyInputSchema>;
const MyOutputSchema = Type.Object({
processed: Type.String({ description: "Processed text" }),
length: Type.Number({ description: "Text length" }),
}) satisfies DataPortSchema;
type MyOutput = FromSchema<typeof MyOutputSchema>;
class TextProcessorTask extends Task<MyInput, MyOutput> {
static readonly type = "TextProcessorTask";
static readonly title = "Text Processor";
static readonly description = "Processes text";
static readonly category = "Text Processing";
static readonly cacheable = true;
static inputSchema() {
return MyInputSchema;
}
static outputSchema() {
return MyOutputSchema;
}
async execute(input: MyInput, context: IExecuteContext): Promise<MyOutput> {
const { text, multiplier = 1 } = input;
const { signal, updateProgress } = context;
if (signal?.aborted) {
throw new TaskAbortedError("Task was cancelled");
}
await updateProgress(0.5, "Processing text...");
await new Promise((resolve) => setTimeout(resolve, 100));
const processed = text.repeat(multiplier);
return {
processed,
length: processed.length,
};
}
// Override validation to use TypeBox's native validation -- only if you needed as the default will work in most cases.
async validateInput(input: Partial<MyInput>): Promise<boolean> {
// Use TypeBox's Value.Check for validation
if (!Value.Check(MyInputSchema, input)) {
const errors = [...Value.Errors(MyInputSchema, input)];
const errorMessages = errors.map((error) => {
const path = error.path || "";
return `${error.message}${path ? ` (${path})` : ""}`;
});
throw new TaskInvalidInputError(
`Input ${JSON.stringify(input)} does not match schema: ${errorMessages.join(", ")}`
);
}
return true;
}
}Zod 4 has built-in JSON Schema support using the .toJSONSchema() method:
import { Task, IExecuteContext } from "@workglow/task-graph";
import { z } from "zod";
import { DataPortSchema } from "@workglow/util";
const MyInputSchemaZod = z.object({
text: z.string().describe("Text to process"),
multiplier: z.number().default(1).optional().describe("Repeat multiplier"),
});
const MyInputSchema = MyInputSchemaZod.toJSONSchema() as DataPortSchema;
// Infer TypeScript types using Zod's built-in type inference
type MyInput = z.infer<typeof MyInputSchemaZod>;
const MyOutputSchemaZod = z.object({
processed: z.string().describe("Processed text"),
length: z.number().describe("Text length"),
});
const MyOutputSchema = MyOutputSchemaZod.toJSONSchema() as DataPortSchema;
type MyOutput = z.infer<typeof MyOutputSchemaZod>;
class TextProcessorTask extends Task<MyInput, MyOutput> {
static readonly type = "TextProcessorTask";
static readonly title = "Text Processor";
static readonly description = "Processes text";
static readonly category = "Text Processing";
static readonly cacheable = true;
static inputSchema() {
return MyInputSchema;
}
static outputSchema() {
return MyOutputSchema;
}
async execute(input: MyInput, context: IExecuteContext): Promise<MyOutput> {
const { text, multiplier = 1 } = input;
const { signal, updateProgress } = context;
if (signal?.aborted) {
throw new TaskAbortedError("Task was cancelled");
}
await updateProgress(0.5, "Processing text...");
await new Promise((resolve) => setTimeout(resolve, 100));
const processed = text.repeat(multiplier);
return {
processed,
length: processed.length,
};
}
// Override validation to use Zod's native validation -- only if you needed as the default will work in most cases.
async validateInput(input: Partial<MyInput>): Promise<boolean> {
try {
// Use Zod's .parse() for validation (throws on error)
MyInputSchemaZod.parse(input);
return true;
} catch (error) {
if (error instanceof z.ZodError) {
const errorMessages = error.errors.map((err) => {
const path = err.path.join(".");
return `${err.message}${path ? ` (${path})` : ""}`;
});
throw new TaskInvalidInputError(
`Input ${JSON.stringify(input)} does not match schema: ${errorMessages.join(", ")}`
);
}
throw error;
}
}
}Note: When using native validation, you still need to return a JSON Schema from inputSchema() and outputSchema() for compatibility with the task graph system. The native validation only affects runtime validation, not schema compatibility checking.
import { DataPortSchema } from "@workglow/util";
class FileProcessorTask extends Task<{ filePath: string }, { content: string }> {
static readonly type = "FileProcessorTask";
static inputSchema() {
return {
type: "object",
properties: {
filePath: {
type: "string",
description: "Path to file",
},
},
required: ["filePath"],
additionalProperties: false,
} as const satisfies DataPortSchema;
}
static outputSchema() {
return {
type: "object",
properties: {
content: {
type: "string",
description: "File content",
},
},
required: ["content"],
additionalProperties: false,
} as const satisfies DataPortSchema;
}
async execute(input: { filePath: string }, { signal, updateProgress }: IExecuteContext) {
try {
await updateProgress(0.1, "Starting file read...");
if (signal?.aborted) {
throw new TaskAbortedError("File read cancelled");
}
// Simulate file reading with progress
await updateProgress(0.5, "Reading file...");
const content = await this.readFile(input.filePath);
await updateProgress(1.0, "File read complete");
return { content };
} catch (error) {
if (error instanceof TaskAbortedError) {
throw error; // Re-throw abort errors
}
throw new TaskError(`Failed to read file: ${error.message}`);
}
}
private async readFile(path: string): Promise<string> {
// Implementation would go here
return "file content";
}
}import { TaskGraph, Dataflow } from "@workglow/task-graph";
// Create tasks
const task1 = new TextProcessorTask({ text: "Hello" }, { id: "processor1" });
const task2 = new TextProcessorTask({ text: "World" }, { id: "processor2" });
const task3 = new TextProcessorTask({ text: "" }, { id: "combiner" });
// Build graph
const graph = new TaskGraph();
graph.addTasks([task1, task2, task3]);
// Define data flows
graph.addDataflow(new Dataflow("processor1", "processed", "combiner", "text"));
graph.addDataflow(new Dataflow("processor2", "processed", "combiner", "text"));
// Execute
const results = await graph.run();import { DataPortSchema } from "@workglow/util";
class AddTask extends Task<{ a: number; b: number }, { sum: number }> {
static readonly type = "AddTask";
static inputSchema() {
return {
type: "object",
properties: {
a: { type: "number" },
b: { type: "number" },
},
required: ["a", "b"],
additionalProperties: false,
} as const satisfies DataPortSchema;
}
static outputSchema() {
return {
type: "object",
properties: {
sum: { type: "number" },
},
required: ["sum"],
additionalProperties: false,
} as const satisfies DataPortSchema;
}
async execute(input: { a: number; b: number }) {
return { sum: input.a + input.b };
}
}
// Create a computational pipeline
const doubleTask = new MultiplyBy2Task({ value: 5 }, { id: "double" });
const doubleTask2 = new MultiplyBy2Task({ value: 5 }, { id: "double2" });
const addTask = new AddTask({}, { id: "add" });
const graph = new TaskGraph();
graph.addTasks([doubleTask, doubleTask2, addTask]);
// Connect outputs to inputs
graph.addDataflow(new Dataflow("double", "result", "add", "a"));
graph.addDataflow(new Dataflow("double2", "result", "add", "b"));
const results = await graph.run();
// Results: double=10, double2=10, add=20// Task that might fail
class RiskyTask extends Task<{ shouldFail: boolean }, { success: boolean }> {
static readonly type = "RiskyTask";
async execute(input: { shouldFail: boolean }) {
if (input.shouldFail) {
throw new TaskError("Task failed as requested");
}
return { success: true };
}
}
// Task that handles errors
class ErrorHandlerTask extends Task<{ fallback: string }, { result: string }> {
static readonly type = "ErrorHandlerTask";
async execute(input: { fallback: string }) {
return { result: input.fallback };
}
}
const graph = new TaskGraph();
const riskyTask = new RiskyTask({ shouldFail: true }, { id: "risky" });
const handlerTask = new ErrorHandlerTask({ fallback: "default" }, { id: "handler" });
graph.addTasks([riskyTask, handlerTask]);
// Connect error output to handler
graph.addDataflow(new Dataflow("risky", "[error]", "handler", "error"));
try {
const results = await graph.run();
} catch (error) {
console.log("Graph execution failed:", error.message);
}import { Workflow } from "@workglow/task-graph";
const workflow = new Workflow();
// Add tasks to workflow
workflow.addTask(new TextProcessorTask({ text: "Hello, World!" }));
// Run workflow
const result = await workflow.run();
console.log(result); // { processed: "Hello, World!", length: 13 }// Create a processing pipeline
const workflow = new Workflow();
// Method 1: Using workflow.pipe()
workflow.pipe(
new TextProcessorTask({ text: "Start" }),
new TextProcessorTask({ multiplier: 2 }),
new TextProcessorTask({ multiplier: 3 })
);
const result = await workflow.run();
// Method 2: Using the pipe helper
import { pipe } from "@workglow/task-graph";
const pipeline = pipe([
new TextProcessorTask({ text: "Start" }),
new TextProcessorTask({ multiplier: 2 }),
new TextProcessorTask({ multiplier: 3 }),
]);
const result2 = await pipeline.run();import { parallel } from "@workglow/task-graph";
// Method 1: Using workflow.parallel()
const workflow = new Workflow();
workflow.parallel([
new TextProcessorTask({ text: "Task 1" }),
new TextProcessorTask({ text: "Task 2" }),
new TextProcessorTask({ text: "Task 3" }),
]);
const results = await workflow.run();
// Results will be an array of outputs
// Method 2: Using the parallel helper
const parallelWorkflow = parallel([
new TextProcessorTask({ text: "Task A" }),
new TextProcessorTask({ text: "Task B" }),
]);
const results2 = await parallelWorkflow.run();// Workflow automatically connects compatible input/output types
const workflow = new Workflow();
// These will auto-connect because output "result" matches input "value"
workflow.addTask(new MultiplyBy2Task({ value: 5 })); // Outputs: { result: number }
workflow.addTask(new MultiplyBy2Task({})); // Inputs: { value: number }
workflow.addTask(new MultiplyBy2Task({})); // Inputs: { value: number }
const result = await workflow.run();
// Result: 5 * 2 * 2 * 2 = 40// Register tasks with the workflow system
declare module "@workglow/task-graph" {
interface Workflow {
myTextProcessor: CreateWorkflow<MyInput, MyOutput>;
}
}
Workflow.prototype.myTextProcessor = Workflow.createWorkflow(TextProcessorTask);
// Now you can use it fluently
const workflow = new Workflow();
workflow.myTextProcessor({ text: "Hello" }).myTextProcessor({ multiplier: 3 });
const result = await workflow.run();Output caching lets repeat executions with identical inputs return instantly without redoing work.
import { Task, TaskGraph, Workflow } from "@workglow/task-graph";
import { DataPortSchema } from "@workglow/util";
import { InMemoryTaskOutputRepository } from "@workglow/test";
// A cacheable task that simulates expensive work
class ExpensiveTask extends Task<{ n: number }, { result: number }> {
static readonly type = "ExpensiveTask";
static readonly cacheable = true;
static inputSchema() {
return {
type: "object",
properties: {
n: { type: "number" },
},
required: ["n"],
additionalProperties: false,
} as const satisfies DataPortSchema;
}
static outputSchema() {
return {
type: "object",
properties: {
result: { type: "number" },
},
required: ["result"],
additionalProperties: false,
} as const satisfies DataPortSchema;
}
async execute(input: { n: number }) {
// Simulate 500ms of CPU/IO work
await new Promise((r) => setTimeout(r, 500));
return { result: input.n * 2 };
}
}
// Create an output cache
const outputCache = new InMemoryTaskOutputRepository();
// Example 1: TaskGraph caching (second run is near-instant)
const graph = new TaskGraph({ outputCache });
graph.addTask(new ExpensiveTask({ n: 42 }, { id: "exp" }));
let t = Date.now();
await graph.run();
const firstRunMs = Date.now() - t;
t = Date.now();
await graph.run(); // identical inputs -> served from cache
const secondRunMs = Date.now() - t;
console.log({ firstRunMs, secondRunMs });
// e.g. { firstRunMs: ~500, secondRunMs: ~1-5 }
// Example 2: Direct Task caching across instances
const missTask = new ExpensiveTask({ n: 43 }, { outputCache });
t = Date.now();
await missTask.run(); // cache miss -> compute and store
const missMs = Date.now() - t;
const hitTask = new ExpensiveTask({ n: 43 }, { outputCache });
t = Date.now();
await hitTask.run(); // cache hit -> instant
const hitMs = Date.now() - t;
console.log({ missMs, hitMs });
// e.g. { missMs: ~500, hitMs: ~1-5 }
// Example 3: Workflow with the same cache
const workflow = new Workflow(outputCache);
workflow.addTask(new ExpensiveTask({ n: 10 }));
t = Date.now();
await workflow.run(); // compute
const wfFirstMs = Date.now() - t;
t = Date.now();
await workflow.run(); // cached
const wfSecondMs = Date.now() - t;
console.log({ wfFirstMs, wfSecondMs });import { FsFolderTaskGraphRepository } from "@workglow/test";
// Create repository
const repository = new FsFolderTaskGraphRepository("./task-graphs");
// Save task graph
const graph = new TaskGraph();
graph.addTask(new MultiplyBy2Task({ value: 10 }));
await repository.saveTaskGraph("my-graph", graph);
// Load task graph
const loadedGraph = await repository.getTaskGraph("my-graph");
const results = await loadedGraph.run();// In-memory (for testing)
import { InMemoryTaskOutputRepository, InMemoryTaskGraphRepository } from "@workglow/test";
// File system
import { FsFolderTaskOutputRepository, FsFolderTaskGraphRepository } from "@workglow/test";
// SQLite
import { SqliteTaskOutputRepository, SqliteTaskGraphRepository } from "@workglow/test";
// IndexedDB (browser)
import { IndexedDbTaskOutputRepository, IndexedDbTaskGraphRepository } from "@workglow/test";class RobustTask extends Task<{ input: string }, { output: string }> {
async execute(input: { input: string }, { signal }: IExecuteContext) {
try {
// Check for cancellation
if (signal?.aborted) {
throw new TaskAbortedError("Task cancelled");
}
// Your logic here
const result = await this.processInput(input.input);
return { output: result };
} catch (error) {
if (error instanceof TaskAbortedError) {
throw error; // Re-throw cancellation
}
// Convert to TaskError with context
throw new TaskError(`Processing failed: ${error.message}`);
}
}
}try {
const results = await graph.run();
} catch (error) {
if (error instanceof TaskAbortedError) {
console.log("Execution was cancelled");
} else if (error instanceof TaskFailedError) {
console.log("A task failed:", error.message);
console.log("Failed task:", error.taskId);
} else if (error instanceof TaskError) {
console.log("Task error:", error.message);
}
}const workflow = new Workflow();
workflow.events.on("error", (error) => {
console.error("Workflow error:", error);
});
workflow.events.on("start", () => {
console.log("Workflow started");
});
workflow.events.on("complete", () => {
console.log("Workflow completed");
});
workflow.addTask(new TextProcessorTask({ text: "Hello" }));
await workflow.run();const workflow = new Workflow();
workflow.addTask(new LongRunningTask());
// Start execution
const resultPromise = workflow.run();
// Abort after 1 second
setTimeout(() => {
workflow.abort();
}, 1000);
try {
await resultPromise;
} catch (error) {
if (error instanceof TaskAbortedError) {
console.log("Execution was aborted");
}
}class RemoteProcessingTask extends JobQueueTask<{ data: string }, { result: string }> {
static readonly type = "RemoteProcessingTask";
async createJob() {
return new Job({
input: this.runInputData,
execute: async (input) => {
// This runs in a job queue (can be distributed)
const processed = await this.callRemoteAPI(input.data);
return { result: processed };
},
});
}
private async callRemoteAPI(data: string): Promise<string> {
// Simulate API call
return `Processed: ${data}`;
}
}class CompositeTask extends GraphAsTask<{ input: string }, { output: string }> {
static readonly type = "CompositeTask";
constructor(input: { input: string }, config: any = {}) {
super(input, config);
// Build internal graph
const subTask1 = new TextProcessorTask({ text: input.input });
const subTask2 = new TextProcessorTask({ multiplier: 2 });
this.subGraph.addTasks([subTask1, subTask2]);
this.subGraph.addDataflow(new Dataflow(subTask1.id, "processed", subTask2.id, "text"));
}
}class TaskFactory extends Task<{ count: number }, { results: any[] }> {
async execute(input: { count: number }, context: IExecuteContext) {
const results = [];
for (let i = 0; i < input.count; i++) {
// Create tasks dynamically
const dynamicTask = new MultiplyBy2Task({ value: i });
// Register with execution context
context.own(dynamicTask);
const result = await dynamicTask.run();
results.push(result);
}
return { results };
}
}In this project, task graphs have connections between tasks called dataflows. These dataflows have different kinds of compatibility checks:
Static rules help decide if an edge should be connected at all. A connection is statically compatible if:
- The source and target are the same exact type
- The source connects to the equivalent of "any" (target accepts anything)
- The source type is acceptable to the target (e.g., a string to something that accepts
oneOf[string[], string])
Assuming the connection is allowed at design time (passes static check), runtime rules determine if they are compatible during execution.
Currently, there is one runtime compatibility check:
- If both input and output schemas have
formatannotations attached,- The format annotation matches the pattern
/\w+(:\w+)?/; the first part is the "name". If alone, it matches any other with the same "name". If there is a second part, then that narrows the type.
- The format annotation matches the pattern
- Format checks apply to all types (strings, arrays, etc.), not just strings.
- A schema with format can connect to a schema with no format (source has format, target doesn't).
- A schema with no format cannot connect to a schema with format (source doesn't have format, target does).
Example:
In the AI package, format: 'model' and format: 'model:EmbeddingTask' are used on string types.
An input with property model and format: 'model' connects to a target with property model and format: 'model:EmbeddingTask'—this compatibility is called "runtime".
It first passes the static check as compatible and then notices a difference in format at runtime.
Format is also used on array types, e.g., format: 'Float64Array' on arrays containing Float64 numbers.
Note: Only connections that pass the runtime check will pass data at runtime.
Task<Input, Output, Config>: Base class for all tasksTaskGraph: Low-level graph execution engineWorkflow<Input, Output>: High-level workflow builderDataflow: Represents data flow between tasksTaskRunner: Handles individual task execution
run(overrides?): Execute the task with optional input overridesrunReactive(overrides?): Execute in reactive modeabort(): Cancel executionsetInput(input): Set input valuesvalidateInput(input): Validate input against schema
addTask(task)/addTasks(tasks): Add tasks to graphaddDataflow(dataflow)/addDataflows(dataflows): Add data flowsrun(input?, config?): Execute the graphgetTask(id): Get task by IDgetDataflow(id): Get dataflow by ID
addTask(task): Add task to workflowpipe(...tasks): Create pipelineparallel(tasks, strategy?): Create parallel executionrun(input?): Execute workflowabort(): Cancel executionreset(): Reset workflow state
TaskOutputRepository: Interface for task output cachingTaskGraphRepository: Interface for task graph persistence
TaskError: Base error classTaskAbortedError: Task was cancelledTaskFailedError: Task execution failedTaskInvalidInputError: Invalid input provided
// Define processing tasks
class LoadDataTask extends Task<{ source: string }, { data: any[] }> {
static readonly type = "LoadDataTask";
async execute(input: { source: string }) {
const data = await this.loadFromSource(input.source);
return { data };
}
private async loadFromSource(source: string): Promise<any[]> {
// Implementation
return [];
}
}
class TransformDataTask extends Task<{ data: any[] }, { transformed: any[] }> {
static readonly type = "TransformDataTask";
async execute(input: { data: any[] }) {
const transformed = input.data.map((item) => ({
...item,
processed: true,
timestamp: new Date(),
}));
return { transformed };
}
}
class SaveDataTask extends Task<{ data: any[] }, { saved: boolean }> {
static readonly type = "SaveDataTask";
async execute(input: { data: any[] }) {
await this.saveToDestination(input.data);
return { saved: true };
}
private async saveToDestination(data: any[]): Promise<void> {
// Implementation
}
}
// Build pipeline
const pipeline = pipe([
new LoadDataTask({ source: "database" }),
new TransformDataTask(),
new SaveDataTask(),
]);
const result = await pipeline.run();// Process multiple data sources in parallel
const workflow = new Workflow();
workflow.parallel([
new LoadDataTask({ source: "api-1" }),
new LoadDataTask({ source: "api-2" }),
new LoadDataTask({ source: "api-3" }),
]);
// Merge results
workflow.addTask(new MergeDataTask());
const result = await workflow.run();class RetryableTask extends Task<{ url: string; retries: number }, { data: any }> {
async execute(input: { url: string; retries: number }) {
for (let i = 0; i < input.retries; i++) {
try {
const data = await fetch(input.url).then((r) => r.json());
return { data };
} catch (error) {
if (i === input.retries - 1) {
throw new TaskError(`Failed after ${input.retries} retries: ${error.message}`);
}
await new Promise((resolve) => setTimeout(resolve, 1000 * Math.pow(2, i)));
}
}
throw new TaskError("Unexpected error");
}
}
const workflow = new Workflow();
workflow.addTask(new RetryableTask({ url: "https://api.example.com", retries: 3 }));
try {
const result = await workflow.run();
} catch (error) {
console.log("All retries failed:", error.message);
}The package includes comprehensive test utilities:
# Run all tests
bun test
# Run specific test file
bun test src/test/task-graph/TaskGraph.test.ts
# Run tests with coverage
bun test --coverageimport { describe, test, expect } from "vitest";
describe("MyCustomTask", () => {
test("should process input correctly", async () => {
const task = new MyCustomTask({ input: "test" });
const result = await task.run();
expect(result.output).toBe("expected-result");
});
test("should handle errors gracefully", async () => {
const task = new MyCustomTask({ input: "invalid" });
await await expect(task.run()).rejects.toThrow(TaskError);
});
test("should respect cancellation", async () => {
const task = new LongRunningTask();
const resultPromise = task.run();
task.abort();
await expect(resultPromise).rejects.toThrow(TaskAbortedError);
});
});Apache 2.0 - See LICENSE for details.