graniteLLMClient
Workspace API / services/granite/graniteLLMClient
services/granite/graniteLLMClient
Classes
LLMTimeoutError
Defined in: services/granite/graniteLLMClient.ts:107
Custom error for LLM request timeouts
Extends
Error
Constructors
Constructor
new LLMTimeoutError(
message,timeoutMs):LLMTimeoutError
Defined in: services/granite/graniteLLMClient.ts:108
Parameters
message
string
timeoutMs
number
Returns
Overrides
Error.constructor
Properties
timeoutMs
readonlytimeoutMs:number
Defined in: services/granite/graniteLLMClient.ts:110
LLMInputTooLargeError
Defined in: services/granite/graniteLLMClient.ts:120
Custom error for LLM input exceeding limits
Extends
Error
Constructors
Constructor
new LLMInputTooLargeError(
message,inputTokens,maxContextLength):LLMInputTooLargeError
Defined in: services/granite/graniteLLMClient.ts:121
Parameters
message
string
inputTokens
number
maxContextLength
number
Returns
Overrides
Error.constructor
Properties
inputTokens
readonlyinputTokens:number
Defined in: services/granite/graniteLLMClient.ts:123
maxContextLength
readonlymaxContextLength:number
Defined in: services/granite/graniteLLMClient.ts:124
GraniteLLMClient
Defined in: services/granite/graniteLLMClient.ts:191
Client for Granite LLM via vLLM OpenAI-compatible API
Constructors
Constructor
new GraniteLLMClient():
GraniteLLMClient
Defined in: services/granite/graniteLLMClient.ts:200
Returns
Methods
generateText()
generateText(
prompt,options):Promise<string>
Defined in: services/granite/graniteLLMClient.ts:382
Generate text using Granite LLM (non-streaming)
Parameters
prompt
string
options
GraniteLLMGenerateOptions = {}
Returns
Promise<string>
generateTextStream()
generateTextStream(
prompt,options):AsyncGenerator<string,void,unknown>
Defined in: services/granite/graniteLLMClient.ts:542
Generate text with streaming Note: Retries only happen before stream starts, not mid-stream
Parameters
prompt
string
options
GraniteLLMGenerateOptions = {}
Returns
AsyncGenerator<string, void, unknown>
Interfaces
GraniteLLMGenerateOptions
Defined in: services/granite/graniteLLMClient.ts:155
Properties
maxTokens?
optionalmaxTokens:number
Defined in: services/granite/graniteLLMClient.ts:156
temperature?
optionaltemperature:number
Defined in: services/granite/graniteLLMClient.ts:157
topP?
optionaltopP:number
Defined in: services/granite/graniteLLMClient.ts:158
stream?
optionalstream:boolean
Defined in: services/granite/graniteLLMClient.ts:159
systemInstruction?
optionalsystemInstruction:string
Defined in: services/granite/graniteLLMClient.ts:160
timeoutMs?
optionaltimeoutMs:number
Defined in: services/granite/graniteLLMClient.ts:161
preCalculatedInputTokens?
optionalpreCalculatedInputTokens:number
Defined in: services/granite/graniteLLMClient.ts:167
Pre-calculated input token count (optional) If provided, the client will use this value instead of re-estimating. Useful when the caller has already calculated tokens (e.g., RAG service).
Variables
GRANITE_LLM_MODEL_ID
constGRANITE_LLM_MODEL_ID:string=GRANITE_CONFIG.model
Defined in: services/granite/graniteLLMClient.ts:135
GRANITE_LLM_ENDPOINT
constGRANITE_LLM_ENDPOINT:string=GRANITE_CONFIG.endpoint
Defined in: services/granite/graniteLLMClient.ts:136
Functions
getGraniteLLMClient()
getGraniteLLMClient():
GraniteLLMClient
Defined in: services/granite/graniteLLMClient.ts:799