Interface ChatCompletionOptions

interface ChatCompletionOptions {
    abortSignal?: AbortSignal;
    nPredict?: number;
    sampling?: SamplingConfig;
    stopTokens?: number[];
    stream?: boolean;
    useCache?: boolean;
    onNewToken(
        token: number,
        piece: Uint8Array,
        currentText: string,
        optionals: { abortSignal: () => any },
    ): any;
}

Properties

abortSignal?: AbortSignal

Optional abort signal to stop the generation. This can also be used to stop during prompt processing (with a bit of delay.)

nPredict?: number
sampling?: SamplingConfig
stopTokens?: number[]

List of custom token IDs for stopping the generation. Note: To convert from text to token ID, use lookupToken()

stream?: boolean

If true, return an AsyncIterable instead of a string

useCache?: boolean

Equivalent to cache_prompt option in llama.cpp server. Useful for chat, because it skip evaluating the history part of the conversation.

Methods

  • Parameters

    • token: number
    • piece: Uint8Array
    • currentText: string
    • optionals: { abortSignal: () => any }
      • abortSignal: () => any

        DEPRECATED, use ChatCompletionOptions["abortSignal"] instead

    Returns any