Class: DeepInfra
Unified language model interface
Extends
Constructors
new DeepInfra()
new DeepInfra(
init
?):DeepInfra
Parameters
• init?: Partial
<OpenAI
>
Returns
Overrides
Defined in
packages/llamaindex/src/llm/deepinfra.ts:9
Properties
additionalChatOptions?
optional
additionalChatOptions:OpenAIAdditionalChatOptions
Inherited from
Defined in
packages/llamaindex/src/llm/openai.ts:176
additionalSessionOptions?
optional
additionalSessionOptions:Omit
<Partial
<ClientOptions
>,"apiKey"
|"timeout"
|"maxRetries"
>
Inherited from
OpenAI
.additionalSessionOptions
Defined in
packages/llamaindex/src/llm/openai.ts:183
apiKey?
optional
apiKey:string
=undefined
Inherited from
Defined in
packages/llamaindex/src/llm/openai.ts:179
maxRetries
maxRetries:
number
Inherited from
Defined in
packages/llamaindex/src/llm/openai.ts:180
maxTokens?
optional
maxTokens:number
Inherited from
Defined in
packages/llamaindex/src/llm/openai.ts:175
model
model:
ChatModel
|string
&object
Inherited from
Defined in
packages/llamaindex/src/llm/openai.ts:169
session
session:
OpenAISession
Inherited from
Defined in
packages/llamaindex/src/llm/openai.ts:182
temperature
temperature:
number
Inherited from
Defined in
packages/llamaindex/src/llm/openai.ts:173
timeout?
optional
timeout:number
Inherited from
Defined in
packages/llamaindex/src/llm/openai.ts:181
topP
topP:
number
Inherited from
Defined in
packages/llamaindex/src/llm/openai.ts:174
Accessors
metadata
get
metadata():LLMMetadata
Returns
Inherited from
Defined in
packages/llamaindex/src/llm/openai.ts:238
supportToolCall
get
supportToolCall():boolean
Returns
boolean
Inherited from
Defined in
packages/llamaindex/src/llm/openai.ts:234
Methods
chat()
chat(params)
chat(
params
):Promise
<AsyncIterable
<ChatResponseChunk
<ToolCallLLMMessageOptions
>>>
Get a chat response from the LLM
Parameters
• params: LLMChatParamsStreaming
<OpenAIAdditionalChatOptions
, ToolCallLLMMessageOptions
>
Returns
Promise
<AsyncIterable
<ChatResponseChunk
<ToolCallLLMMessageOptions
>>>
Inherited from
Defined in
packages/llamaindex/src/llm/openai.ts:315
chat(params)
chat(
params
):Promise
<ChatResponse
<ToolCallLLMMessageOptions
>>
Get a chat response from the LLM
Parameters
• params: LLMChatParamsNonStreaming
<OpenAIAdditionalChatOptions
, ToolCallLLMMessageOptions
>
Returns
Promise
<ChatResponse
<ToolCallLLMMessageOptions
>>
Inherited from
Defined in
packages/llamaindex/src/llm/openai.ts:321