Skip to content

Execution Planning

Imagine that we’ve refactored our generateDadJoke program from our Getting Started guide. Now, instead of handling all errors internally, the code can fail with domain-specific issues like network interruptions or provider outages:

import type {
import AiResponse
AiResponse
,
import Completions
Completions
} from "@effect/ai"
import {
import OpenAiCompletions
OpenAiCompletions
} from "@effect/ai-openai"
import {
import Data
Data
,
import Effect

@since2.0.0

@since2.0.0

@since2.0.0

Effect
} from "effect"
class
class NetworkError
NetworkError
extends
import Data
Data
.
const TaggedError: <"NetworkError">(tag: "NetworkError") => new <A>(args: Equals<A, {}> extends true ? void : { readonly [P in keyof A as P extends "_tag" ? never : P]: A[P]; }) => YieldableError & {
...;
} & Readonly<...>

@since2.0.0

TaggedError
("NetworkError") {}
class
class ProviderOutage
ProviderOutage
extends
import Data
Data
.
const TaggedError: <"ProviderOutage">(tag: "ProviderOutage") => new <A>(args: Equals<A, {}> extends true ? void : { readonly [P in keyof A as P extends "_tag" ? never : P]: A[P]; }) => YieldableError & {
...;
} & Readonly<...>

@since2.0.0

TaggedError
("ProviderOutage") {}
declare const
const generateDadJoke: Effect.Effect<AiResponse.AiResponse, NetworkError | ProviderOutage, Completions.Completions>
generateDadJoke
:
import Effect

@since2.0.0

@since2.0.0

@since2.0.0

Effect
.
interface Effect<out A, out E = never, out R = never>

The Effect interface defines a value that describes a workflow or job, which can succeed or fail.

Details

The Effect interface represents a computation that can model a workflow involving various types of operations, such as synchronous, asynchronous, concurrent, and parallel interactions. It operates within a context of type R, and the result can either be a success with a value of type A or a failure with an error of type E. The Effect is designed to handle complex interactions with external resources, offering advanced features such as fiber-based concurrency, scheduling, interruption handling, and scalability. This makes it suitable for tasks that require fine-grained control over concurrency and error management.

To execute an Effect value, you need a Runtime, which provides the environment necessary to run and manage the computation.

@since2.0.0

@since2.0.0

Effect
<
import AiResponse
AiResponse
.
class AiResponse

@since1.0.0

AiResponse
,
class NetworkError
NetworkError
|
class ProviderOutage
ProviderOutage
,
import Completions
Completions
.
class Completions

@since1.0.0

@since1.0.0

Completions
>
const
const main: Effect.Effect<void, NetworkError | ProviderOutage, OpenAiClient | AiModels>
main
=
import Effect

@since2.0.0

@since2.0.0

@since2.0.0

Effect
.
const gen: <YieldWrap<Effect.Effect<AiPlan<in Error, in out Provides, in out Requires>.Provider<Completions.Completions | Tokenizer>, never, OpenAiClient | AiModels>> | YieldWrap<...>, void>(f: (resume: Effect.Adapter) => Generator<...>) => Effect.Effect<...> (+1 overload)

Provides a way to write effectful code using generator functions, simplifying control flow and error handling.

When to Use

Effect.gen allows you to write code that looks and behaves like synchronous code, but it can handle asynchronous tasks, errors, and complex control flow (like loops and conditions). It helps make asynchronous code more readable and easier to manage.

The generator functions work similarly to async/await but with more explicit control over the execution of effects. You can yield* values from effects and return the final result at the end.

Example

import { Effect } from "effect"
const addServiceCharge = (amount: number) => amount + 1
const applyDiscount = (
total: number,
discountRate: number
): Effect.Effect<number, Error> =>
discountRate === 0
? Effect.fail(new Error("Discount rate cannot be zero"))
: Effect.succeed(total - (total * discountRate) / 100)
const fetchTransactionAmount = Effect.promise(() => Promise.resolve(100))
const fetchDiscountRate = Effect.promise(() => Promise.resolve(5))
export const program = Effect.gen(function* () {
const transactionAmount = yield* fetchTransactionAmount
const discountRate = yield* fetchDiscountRate
const discountedAmount = yield* applyDiscount(
transactionAmount,
discountRate
)
const finalAmount = addServiceCharge(discountedAmount)
return `Final amount to charge: ${finalAmount}`
})

@since2.0.0

gen
(function*() {
const
const gpt4o: AiPlan.Provider<Completions.Completions | Tokenizer>
gpt4o
= yield*
import OpenAiCompletions
OpenAiCompletions
.
const model: (model: (string & {}) | OpenAiCompletions.Model, config?: Omit<OpenAiCompletions.Config.Service, "model">) => AiModel<Completions.Completions | Tokenizer, OpenAiClient>

@since1.0.0

model
("gpt-4o")
const
const response: AiResponse.AiResponse
response
= yield*
const gpt4o: AiPlan.Provider<Completions.Completions | Tokenizer>
gpt4o
.
AiPlan<in Error, in out Provides, in out Requires>.Provider<Completions | Tokenizer>.provide: <AiResponse.AiResponse, NetworkError | ProviderOutage, Completions.Completions>(effect: Effect.Effect<AiResponse.AiResponse, NetworkError | ProviderOutage, Completions.Completions>) => Effect.Effect<...>
provide
(
const generateDadJoke: Effect.Effect<AiResponse.AiResponse, NetworkError | ProviderOutage, Completions.Completions>
generateDadJoke
)
})

This is fine, but what if we instead want to:

  • Retry the program a fixed number of times on NetworkErrors
  • Add some backoff delay between retries
  • Fallback to a different model provider if OpenAi is down

How can we accomplish such logic?

The Effect AI integrations provide a robust method for creating structured execution plans for your LLM interactions through the AiPlan data type. Rather than making a single model call and hoping it succeeds, AiPlan lets you describe how to handle retries, fallbacks, and recoverable errors in a clear, declarative way.

This is especially useful when:

  • You want to fall back to a secondary model if the primary one is unavailable
  • You want to retry on transient errors (e.g. network failures)
  • You want to control timing between retry attempts

Think of an AiPlan as the blueprint for for an LLM interaction, with logic for when to keep trying to interact with one provider and when to switch providers.

interface AiPlan<Errors, Provides, Requires> {}

An AiPlan has three generic type parameters:

  • Errors: Any errors that can be handled during execution of the plan
  • Provides: The services that this plan can provide (e.g. Completions, Embeddings)
  • Requires: The services that this plan requires (e.g. OpenAiClient, AnthropicClient)

If you’ve used AiModel before (via OpenAiCompletions.model() or similar), you’ll find AiPlan familiar. In fact, an AiModel is in fact an AiPlan with just a single step.

This means you can start by writing your code with plain AiModels, and as your needs become more complex (e.g. adding retries, switching providers), you can upgrade to AiPlan without changing how the rest of your code works.

The primary entry point to building up an AiPlan is the AiPlan.fromModel constructor.

This method defines the primary model that you would like to use for an LLM interaction, as well as the rules for retrying it under specific conditions.

Use this when you want to:

  • Retry a model multiple times
  • Customize backoff timing between attempts
  • Decide whether to retry based on the error type

Example (Creating an AiPlan from an AiModel)

import {
import AiPlan
AiPlan
} from "@effect/ai"
13 collapsed lines
import type {
import AiResponse
AiResponse
,
import Completions
Completions
} from "@effect/ai"
import {
import OpenAiCompletions
OpenAiCompletions
} from "@effect/ai-openai"
import {
import Data
Data
,
import Effect

@since2.0.0

@since2.0.0

@since2.0.0

Effect
,
import Schedule
Schedule
} from "effect"
class
class NetworkError
NetworkError
extends
import Data
Data
.
const TaggedError: <"NetworkError">(tag: "NetworkError") => new <A>(args: Equals<A, {}> extends true ? void : { readonly [P in keyof A as P extends "_tag" ? never : P]: A[P]; }) => YieldableError & {
...;
} & Readonly<...>

@since2.0.0

TaggedError
("NetworkError") {}
class
class ProviderOutage
ProviderOutage
extends
import Data
Data
.
const TaggedError: <"ProviderOutage">(tag: "ProviderOutage") => new <A>(args: Equals<A, {}> extends true ? void : { readonly [P in keyof A as P extends "_tag" ? never : P]: A[P]; }) => YieldableError & {
...;
} & Readonly<...>

@since2.0.0

TaggedError
("ProviderOutage") {}
declare const
const generateDadJoke: Effect.Effect<AiResponse.AiResponse, NetworkError | ProviderOutage, Completions.Completions>
generateDadJoke
:
import Effect

@since2.0.0

@since2.0.0

@since2.0.0

Effect
.
interface Effect<out A, out E = never, out R = never>

The Effect interface defines a value that describes a workflow or job, which can succeed or fail.

Details

The Effect interface represents a computation that can model a workflow involving various types of operations, such as synchronous, asynchronous, concurrent, and parallel interactions. It operates within a context of type R, and the result can either be a success with a value of type A or a failure with an error of type E. The Effect is designed to handle complex interactions with external resources, offering advanced features such as fiber-based concurrency, scheduling, interruption handling, and scalability. This makes it suitable for tasks that require fine-grained control over concurrency and error management.

To execute an Effect value, you need a Runtime, which provides the environment necessary to run and manage the computation.

@since2.0.0

@since2.0.0

Effect
<
import AiResponse
AiResponse
.
class AiResponse

@since1.0.0

AiResponse
,
class NetworkError
NetworkError
|
class ProviderOutage
ProviderOutage
,
import Completions
Completions
.
class Completions

@since1.0.0

@since1.0.0

Completions
>
const
const DadJokePlan: AiPlan.AiPlan<NetworkError | ProviderOutage, Completions.Completions | Tokenizer, OpenAiClient>
DadJokePlan
=
import AiPlan
AiPlan
.
const fromModel: <Completions.Completions | Tokenizer, OpenAiClient, NetworkError | ProviderOutage, Duration, unknown, never, never>(model: AiModel<...>, options?: {
...;
} | undefined) => AiPlan.AiPlan<...>

@since1.0.0

fromModel
(
import OpenAiCompletions
OpenAiCompletions
.
const model: (model: (string & {}) | OpenAiCompletions.Model, config?: Omit<OpenAiCompletions.Config.Service, "model">) => AiModel<Completions.Completions | Tokenizer, OpenAiClient>

@since1.0.0

model
("gpt-4o"), {
attempts?: number | undefined
attempts
: 3,
schedule?: Schedule.Schedule<Duration, unknown, never> | undefined
schedule
:
import Schedule
Schedule
.
const exponential: (base: DurationInput, factor?: number) => Schedule.Schedule<Duration>

Creates a schedule that recurs indefinitely with exponentially increasing delays.

Details

This schedule starts with an initial delay of base and increases the delay exponentially on each repetition using the formula base * factor^n, where n is the number of times the schedule has executed so far. If no factor is provided, it defaults to 2, causing the delay to double after each execution.

@since2.0.0

exponential
("100 millis", 1.5),
while?: ((error: NetworkError | ProviderOutage) => boolean | Effect.Effect<boolean, never, never>) | undefined
while
: (
error: NetworkError | ProviderOutage
error
:
class NetworkError
NetworkError
|
class ProviderOutage
ProviderOutage
) =>
error: NetworkError | ProviderOutage
error
.
_tag: "NetworkError" | "ProviderOutage"
_tag
=== "NetworkError"
})
const
const main: Effect.Effect<void, NetworkError | ProviderOutage, OpenAiClient | AiModels>
main
=
import Effect

@since2.0.0

@since2.0.0

@since2.0.0

Effect
.
const gen: <YieldWrap<Effect.Effect<AiPlan.AiPlan<in Error, in out Provides, in out Requires>.Provider<Completions.Completions | Tokenizer>, never, OpenAiClient | AiModels>> | YieldWrap<...>, void>(f: (resume: Effect.Adapter) => Generator<...>) => Effect.Effect<...> (+1 overload)

Provides a way to write effectful code using generator functions, simplifying control flow and error handling.

When to Use

Effect.gen allows you to write code that looks and behaves like synchronous code, but it can handle asynchronous tasks, errors, and complex control flow (like loops and conditions). It helps make asynchronous code more readable and easier to manage.

The generator functions work similarly to async/await but with more explicit control over the execution of effects. You can yield* values from effects and return the final result at the end.

Example

import { Effect } from "effect"
const addServiceCharge = (amount: number) => amount + 1
const applyDiscount = (
total: number,
discountRate: number
): Effect.Effect<number, Error> =>
discountRate === 0
? Effect.fail(new Error("Discount rate cannot be zero"))
: Effect.succeed(total - (total * discountRate) / 100)
const fetchTransactionAmount = Effect.promise(() => Promise.resolve(100))
const fetchDiscountRate = Effect.promise(() => Promise.resolve(5))
export const program = Effect.gen(function* () {
const transactionAmount = yield* fetchTransactionAmount
const discountRate = yield* fetchDiscountRate
const discountedAmount = yield* applyDiscount(
transactionAmount,
discountRate
)
const finalAmount = addServiceCharge(discountedAmount)
return `Final amount to charge: ${finalAmount}`
})

@since2.0.0

gen
(function*() {
const
const plan: AiPlan.AiPlan.Provider<Completions.Completions | Tokenizer>
plan
= yield*
const DadJokePlan: AiPlan.AiPlan<NetworkError | ProviderOutage, Completions.Completions | Tokenizer, OpenAiClient>
DadJokePlan
const
const response: AiResponse.AiResponse
response
= yield*
const plan: AiPlan.AiPlan.Provider<Completions.Completions | Tokenizer>
plan
.
AiPlan<in Error, in out Provides, in out Requires>.Provider<Completions | Tokenizer>.provide: <AiResponse.AiResponse, NetworkError | ProviderOutage, Completions.Completions>(effect: Effect.Effect<AiResponse.AiResponse, NetworkError | ProviderOutage, Completions.Completions>) => Effect.Effect<...>
provide
(
const generateDadJoke: Effect.Effect<AiResponse.AiResponse, NetworkError | ProviderOutage, Completions.Completions>
generateDadJoke
)
})

This plan will:

  • Attempt to use OpenAi’s "gpt-4o" model up to 3 times
  • Wait with an exponential backoff between attempts (starting at 100ms)
  • Only re-attempt the call to OpenAi if the error is a NetworkError

To make your LLM interactions resilient to provider outages, you can define a fallback model to use via AiPlan.withFallback. This will allow the plan to automatically fallback to another model if the previous step in the execution plan fails.

Use this when:

  • You want to make your LLM interactions resilient to provider outages
  • You want to potentially have multiple fallback models

Example (Adding a Fallback to Anthropic from OpenAi)

import {
import AiPlan
AiPlan
} from "@effect/ai"
14 collapsed lines
import type {
import AiResponse
AiResponse
,
import Completions
Completions
} from "@effect/ai"
import {
import AnthropicCompletions
AnthropicCompletions
} from "@effect/ai-anthropic"
import {
import OpenAiCompletions
OpenAiCompletions
} from "@effect/ai-openai"
import {
import Data
Data
,
import Effect

@since2.0.0

@since2.0.0

@since2.0.0

Effect
,
import Schedule
Schedule
} from "effect"
class
class NetworkError
NetworkError
extends
import Data
Data
.
const TaggedError: <"NetworkError">(tag: "NetworkError") => new <A>(args: Equals<A, {}> extends true ? void : { readonly [P in keyof A as P extends "_tag" ? never : P]: A[P]; }) => YieldableError & {
...;
} & Readonly<...>

@since2.0.0

TaggedError
("NetworkError") {}
class
class ProviderOutage
ProviderOutage
extends
import Data
Data
.
const TaggedError: <"ProviderOutage">(tag: "ProviderOutage") => new <A>(args: Equals<A, {}> extends true ? void : { readonly [P in keyof A as P extends "_tag" ? never : P]: A[P]; }) => YieldableError & {
...;
} & Readonly<...>

@since2.0.0

TaggedError
("ProviderOutage") {}
declare const
const generateDadJoke: Effect.Effect<AiResponse.AiResponse, NetworkError | ProviderOutage, Completions.Completions>
generateDadJoke
:
import Effect

@since2.0.0

@since2.0.0

@since2.0.0

Effect
.
interface Effect<out A, out E = never, out R = never>

The Effect interface defines a value that describes a workflow or job, which can succeed or fail.

Details

The Effect interface represents a computation that can model a workflow involving various types of operations, such as synchronous, asynchronous, concurrent, and parallel interactions. It operates within a context of type R, and the result can either be a success with a value of type A or a failure with an error of type E. The Effect is designed to handle complex interactions with external resources, offering advanced features such as fiber-based concurrency, scheduling, interruption handling, and scalability. This makes it suitable for tasks that require fine-grained control over concurrency and error management.

To execute an Effect value, you need a Runtime, which provides the environment necessary to run and manage the computation.

@since2.0.0

@since2.0.0

Effect
<
import AiResponse
AiResponse
.
class AiResponse

@since1.0.0

AiResponse
,
class NetworkError
NetworkError
|
class ProviderOutage
ProviderOutage
,
import Completions
Completions
.
class Completions

@since1.0.0

@since1.0.0

Completions
>
const
const DadJokePlan: AiPlan.AiPlan<NetworkError | ProviderOutage, Completions.Completions | Tokenizer, OpenAiClient | AnthropicClient>
DadJokePlan
=
import AiPlan
AiPlan
.
const fromModel: <Completions.Completions | Tokenizer, OpenAiClient, NetworkError | ProviderOutage, Duration, unknown, never, never>(model: AiModel<...>, options?: {
...;
} | undefined) => AiPlan.AiPlan<...>

@since1.0.0

fromModel
(
import OpenAiCompletions
OpenAiCompletions
.
const model: (model: (string & {}) | OpenAiCompletions.Model, config?: Omit<OpenAiCompletions.Config.Service, "model">) => AiModel<Completions.Completions | Tokenizer, OpenAiClient>

@since1.0.0

model
("gpt-4o"), {
4 collapsed lines
attempts?: number | undefined
attempts
: 3,
schedule?: Schedule.Schedule<Duration, unknown, never> | undefined
schedule
:
import Schedule
Schedule
.
const exponential: (base: DurationInput, factor?: number) => Schedule.Schedule<Duration>

Creates a schedule that recurs indefinitely with exponentially increasing delays.

Details

This schedule starts with an initial delay of base and increases the delay exponentially on each repetition using the formula base * factor^n, where n is the number of times the schedule has executed so far. If no factor is provided, it defaults to 2, causing the delay to double after each execution.

@since2.0.0

exponential
("100 millis", 1.5),
while?: ((error: NetworkError | ProviderOutage) => boolean | Effect.Effect<boolean, never, never>) | undefined
while
: (
error: NetworkError | ProviderOutage
error
:
class NetworkError
NetworkError
|
class ProviderOutage
ProviderOutage
) =>
error: NetworkError | ProviderOutage
error
.
_tag: "NetworkError" | "ProviderOutage"
_tag
=== "NetworkError"
}).
Pipeable.pipe<AiPlan.AiPlan<NetworkError | ProviderOutage, Completions.Completions | Tokenizer, OpenAiClient>, AiPlan.AiPlan<...>>(this: AiPlan.AiPlan<...>, ab: (_: AiPlan.AiPlan<...>) => AiPlan.AiPlan<...>): AiPlan.AiPlan<...> (+21 overloads)
pipe
(
import AiPlan
AiPlan
.
const withFallback: <Completions.Completions | Tokenizer, Completions.Completions | Tokenizer, AnthropicClient, Duration, NetworkError | ProviderOutage, unknown, never, never>(options: {
...;
}) => <E, Requires>(self: AiPlan.AiPlan<...>) => AiPlan.AiPlan<...> (+1 overload)

@since1.0.0

withFallback
({
model: AiModel<Completions.Completions | Tokenizer, AnthropicClient>
model
:
import AnthropicCompletions
AnthropicCompletions
.
const model: (model: (string & {}) | AnthropicCompletions.Model, config?: Omit<AnthropicCompletions.Config.Service, "model">) => AiModel<Completions.Completions | Tokenizer, AnthropicClient>

@since1.0.0

model
("claude-3-7-sonnet-latest"),
attempts?: number | undefined
attempts
: 2,
schedule?: Schedule.Schedule<Duration, unknown, never> | undefined
schedule
:
import Schedule
Schedule
.
const exponential: (base: DurationInput, factor?: number) => Schedule.Schedule<Duration>

Creates a schedule that recurs indefinitely with exponentially increasing delays.

Details

This schedule starts with an initial delay of base and increases the delay exponentially on each repetition using the formula base * factor^n, where n is the number of times the schedule has executed so far. If no factor is provided, it defaults to 2, causing the delay to double after each execution.

@since2.0.0

exponential
("100 millis", 1.5),
while?: ((error: NetworkError | ProviderOutage) => boolean | Effect.Effect<boolean, never, never>) | undefined
while
: (
error: NetworkError | ProviderOutage
error
:
class NetworkError
NetworkError
|
class ProviderOutage
ProviderOutage
) =>
error: NetworkError | ProviderOutage
error
.
_tag: "NetworkError" | "ProviderOutage"
_tag
=== "ProviderOutage"
})
)
const
const main: Effect.Effect<void, NetworkError | ProviderOutage, OpenAiClient | AiModels | AnthropicClient>
main
=
import Effect

@since2.0.0

@since2.0.0

@since2.0.0

Effect
.
const gen: <YieldWrap<Effect.Effect<AiPlan.AiPlan<in Error, in out Provides, in out Requires>.Provider<Completions.Completions | Tokenizer>, never, OpenAiClient | AiModels | AnthropicClient>> | YieldWrap<...>, void>(f: (resume: Effect.Adapter) => Generator<...>) => Effect.Effect<...> (+1 overload)

Provides a way to write effectful code using generator functions, simplifying control flow and error handling.

When to Use

Effect.gen allows you to write code that looks and behaves like synchronous code, but it can handle asynchronous tasks, errors, and complex control flow (like loops and conditions). It helps make asynchronous code more readable and easier to manage.

The generator functions work similarly to async/await but with more explicit control over the execution of effects. You can yield* values from effects and return the final result at the end.

Example

import { Effect } from "effect"
const addServiceCharge = (amount: number) => amount + 1
const applyDiscount = (
total: number,
discountRate: number
): Effect.Effect<number, Error> =>
discountRate === 0
? Effect.fail(new Error("Discount rate cannot be zero"))
: Effect.succeed(total - (total * discountRate) / 100)
const fetchTransactionAmount = Effect.promise(() => Promise.resolve(100))
const fetchDiscountRate = Effect.promise(() => Promise.resolve(5))
export const program = Effect.gen(function* () {
const transactionAmount = yield* fetchTransactionAmount
const discountRate = yield* fetchDiscountRate
const discountedAmount = yield* applyDiscount(
transactionAmount,
discountRate
)
const finalAmount = addServiceCharge(discountedAmount)
return `Final amount to charge: ${finalAmount}`
})

@since2.0.0

gen
(function*() {
const
const plan: AiPlan.AiPlan.Provider<Completions.Completions | Tokenizer>
plan
= yield*
const DadJokePlan: AiPlan.AiPlan<NetworkError | ProviderOutage, Completions.Completions | Tokenizer, OpenAiClient | AnthropicClient>
DadJokePlan
const
const response: AiResponse.AiResponse
response
= yield*
const plan: AiPlan.AiPlan.Provider<Completions.Completions | Tokenizer>
plan
.
AiPlan<in Error, in out Provides, in out Requires>.Provider<Completions | Tokenizer>.provide: <AiResponse.AiResponse, NetworkError | ProviderOutage, Completions.Completions>(effect: Effect.Effect<AiResponse.AiResponse, NetworkError | ProviderOutage, Completions.Completions>) => Effect.Effect<...>
provide
(
const generateDadJoke: Effect.Effect<AiResponse.AiResponse, NetworkError | ProviderOutage, Completions.Completions>
generateDadJoke
)
})

This plan will:

  • Fallback to Anthropic if the OpenAi step of the plan fails
  • Attempt to use Anthropic’s "claude-3-7-sonnet" model up to 2 times
  • Wait with an exponential backoff between attempts (starting at 100ms)
  • Only run and / or re-attempt the fallback if the error is a ProviderOutage

The following is the complete program with the desired AiPlan fully implemented:

import { AiPlan } from "@effect/ai"
import type { AiResponse, Completions } from "@effect/ai"
import { AnthropicClient, AnthropicCompletions } from "@effect/ai-anthropic"
import { OpenAiClient, OpenAiCompletions } from "@effect/ai-openai"
import { NodeHttpClient } from "@effect/platform-node"
import { Data, Effect, Schedule } from "effect"
class NetworkError extends Data.TaggedError("NetworkError") {}
class ProviderOutage extends Data.TaggedError("ProviderOutage") {}
declare const generateDadJoke: Effect.Effect<
AiResponse.AiResponse,
NetworkError | ProviderOutage,
Completions.Completions
>
const DadJokePlan = AiPlan.fromModel(OpenAiCompletions.model("gpt-4o"), {
attempts: 3,
schedule: Schedule.exponential("100 millis", 1.5),
while: (error: NetworkError | ProviderOutage) =>
error._tag === "NetworkError"
}).pipe(
AiPlan.withFallback({
model: AnthropicCompletions.model("claude-3-7-sonnet-latest"),
attempts: 2,
schedule: Schedule.exponential("100 millis", 1.5),
while: (error: NetworkError | ProviderOutage) =>
error._tag === "ProviderOutage"
})
)
const main = Effect.gen(function*() {
const plan = yield* DadJokePlan
const response = yield* plan.provide(generateDadJoke)
})
const Anthropic = AnthropicClient.layerConfig({
apiKey: Config.redacted("ANTHROPIC_API_KEY")
}).pipe(Layer.provide(NodeHttpClient.layerUndici))
const OpenAi = OpenAiClient.layerConfig({
apiKey: Config.redacted("OPENAI_API_KEY")
}).pipe(Layer.provide(NodeHttpClient.layerUndici))
main.pipe(
Effect.provide([Anthropic, OpenAi]),
Effect.runPromise
)
  • You can chain multiple AiPlan.withFallback calls together for more complex failover logic
  • Just like AiModel, AiPlan is re-usable and can be abstracted into shared logic if desired
  • Great for teams needing multi-provider resilience or predictable behavior under failure