Class responsible for parsing the output of AutoGPT. It extends the BaseOutputParser class.

Hierarchy

Constructors

Methods

  • Method not implemented in the class and will throw an error if called. It is likely meant to be overridden in subclasses to provide specific format instructions.

    Returns string

    Throws an error.

  • Calls the parser with a given input and optional configuration options. If the input is a string, it creates a generation with the input as text and calls parseResult. If the input is a BaseMessage, it creates a generation with the input as a message and the content of the input as text, and then calls parseResult.

    Parameters

    • input: string | BaseMessage

      The input to the parser, which can be a string or a BaseMessage.

    • Optional options: BaseCallbackConfig

      Optional configuration options.

    Returns Promise<AutoGPTAction>

    A promise of the parsed output.

  • Asynchronous method that takes a string as input and attempts to parse it into an AutoGPTAction object. If the input string cannot be parsed directly, the method tries to preprocess the string using the preprocessJsonInput function and parse it again. If parsing fails again, it returns an AutoGPTAction object with an error message.

    Parameters

    • text: string

      The string to be parsed.

    Returns Promise<AutoGPTAction>

    A Promise that resolves to an AutoGPTAction object.

  • Parses the result of an LLM call. This method is meant to be implemented by subclasses to define how the output from the LLM should be parsed.

    Parameters

    Returns Promise<AutoGPTAction>

    A promise of the parsed output.

  • Create a new runnable sequence that runs each individual runnable in series, piping the output of one runnable into another runnable or runnable-like.

    Type Parameters

    • NewRunOutput

    Parameters

    • coerceable: RunnableLike<AutoGPTAction, NewRunOutput>

      A runnable, function, or object whose values are functions or runnables.

    Returns RunnableSequence<string | BaseMessage, Exclude<NewRunOutput, Error>>

    A new runnable sequence.

  • Stream all output from a runnable, as reported to the callback system. This includes all inner runs of LLMs, Retrievers, Tools, etc. Output is streamed as Log objects, which include a list of jsonpatch ops that describe how the state of the run has changed in each step, and the final state of the run. The jsonpatch ops can be applied in order to construct state.

    Parameters

    • input: string | BaseMessage
    • Optional options: Partial<BaseCallbackConfig>
    • Optional streamOptions: Omit<LogStreamCallbackHandlerInput, "autoClose">

    Returns AsyncGenerator<RunLogPatch, any, unknown>

  • Default implementation of transform, which buffers input and then calls stream. Subclasses should override this method if they can start producing output while input is still being generated.

    Parameters

    Returns AsyncGenerator<AutoGPTAction, any, unknown>

Generated using TypeDoc