Workflows are a way to chain multiple models together to create a more complex pipeline. This allows you to create a single endpoint that can take an input and pass it through multiple models in sequence. This is useful for creating more complex models that require multiple steps, or for creating a single endpoint that can handle multiple tasks.
Use this file to discover all available pages before exploring further.
Workflows let you chain multiple models together into a single endpoint, creating complex pipelines that run as one API call. Instead of orchestrating individual model requests yourself, you define the steps and fal handles the execution, passing outputs from one model as inputs to the next.Unlike standard model calls that return a single result, workflows emit streaming events as each step progresses, giving you access to intermediate results along the way. This makes them ideal for multi-step generation tasks where you want real-time feedback. For more on consuming streaming responses, see streaming inference.
Workflow APIs work the same way as other model endpoints, you can simply send a request and get a response back. However, it is common for workflows to contain multiple steps and produce intermediate results, as each step contains their own response that could be relevant in your use-case.Therefore, workflows benefit from the streaming feature, which allows you to get partial results as they are being generated.
The workflow API will trigger a few events during its execution, these events can be used to monitor the progress of the workflow and get intermediate results. Below are the events that you can expect from a workflow stream:
The error event is triggered when an error occurs during the execution of a step. The error object contains the error.status with the HTTP status code, an error message as well as error.body with the underlying error serialized.
{ "type": "error", "node_id": "stable_diffusion_xl", "message": "Error while fetching the result of the request d778bdf4-0275-47c2-9f23-16c27041cbeb", "error": { "status": 422, "body": { "detail": [ { "loc": ["body", "num_images"], "msg": "ensure this value is less than or equal to 8", "type": "value_error.number.not_le", "ctx": { "limit_value": 8 } } ] } }}
A cool and simple example of the power of workflows is workflows/fal-ai/sdxl-sticker, which consists of three steps:
1
Generates an image using fal-ai/fast-sdxl.
2
Remove the background of the image using fal-ai/imageutils/rembg.
3
Converts the image to a sticker using fal-ai/face-to-sticker.
What could be a tedious process of running and coordinating three different models is now a single endpoint that you can call with a single request.
Javascript
python
python (async)
Swift
import { fal } from "@fal-ai/client";const stream = await fal.stream("workflows/fal-ai/sdxl-sticker", {input: { prompt: "a face of a cute puppy, in the style of pixar animation",},});for await (const event of stream) {console.log("partial", event);}const result = await stream.done();console.log("final result", result);
import fal_clientstream = fal_client.stream( "workflows/fal-ai/sdxl-sticker", arguments={ "prompt": "a face of a cute puppy, in the style of pixar animation", },)for event in stream: print(event)
import asyncioimport fal_clientasync def main(): stream = await fal_client.stream_async( "workflows/fal-ai/sdxl-sticker", arguments={ "prompt": "a face of a cute puppy, in the style of pixar animation", }, ) async for event in stream: print(event)if __name__ == "__main__": asyncio.run(main())
Coming soonThe Swift client does not support streaming yet.