import type { LanguageModelV2StreamPart } from '@ai-sdk/provider-v5'; import { MastraBase } from '../../base.js'; import type { ChunkType, CreateStream, OnResult } from '../types.js'; /** * Safely enqueue a chunk into a ReadableStreamDefaultController. * Returns true if the enqueue succeeded, false if the controller was already closed/errored. * * Prefer this over checking desiredSize before enqueue, because desiredSize === 0 * indicates backpressure (queue full, stream still open) — not closure. * Guarding on desiredSize would silently drop chunks under normal backpressure. */ export declare function safeEnqueue(controller: ReadableStreamDefaultController, chunk: T): boolean; /** * Safely close a ReadableStreamDefaultController. * Returns true if the close succeeded, false if the controller was already closed/errored. */ export declare function safeClose(controller: ReadableStreamDefaultController): boolean; /** * Safely signal an error on a ReadableStreamDefaultController. * Returns true if the error succeeded, false if the controller was already closed/errored. */ export declare function safeError(controller: ReadableStreamDefaultController, error: unknown): boolean; export declare abstract class MastraModelInput extends MastraBase { abstract transform({ runId, stream, controller, }: { runId: string; stream: ReadableStream>; controller: ReadableStreamDefaultController; }): Promise; initialize({ runId, createStream, onResult }: { createStream: CreateStream; runId: string; onResult: OnResult; }): ReadableStream; } //# sourceMappingURL=input.d.ts.map