Skip to main content

Documentation Index

Fetch the complete documentation index at: https://docs.nscale.com/llms.txt

Use this file to discover all available pages before exploring further.

Aliases: inferencing, infer

Subcommands


list-models

Returns a list of all models available on the serverless platform.
nscale inferencing list-models [flags]

Flags

FlagDescription
--jsonOutput in JSON format

Example

nscale inferencing list-models --json

list-endpoints

Returns a list of all model endpoints available for use by the specified organization.
nscale inferencing list-endpoints [flags]

Flags

FlagDescription
--org stringOrganization ID
--jsonOutput in JSON format

Example

nscale inferencing list-endpoints --org <org-id>

chat

Send a chat completion request to the inference API using a configuration file. Supports both batch and interactive modes.
nscale inferencing chat [flags]

Flags

FlagDescription
--config stringPath to a chat configuration file (JSON)
--messages stringPath to a JSON+LD file containing additional messages
--uiLaunch interactive chat TUI

Examples

# Send a chat completion using a config file
nscale inferencing chat --config chat-config.json

# Launch the interactive chat TUI
nscale inferencing chat --config chat-config.json --ui

# Include additional messages from a file
nscale inferencing chat --config chat-config.json --messages extra-messages.json

Models

Learn about available models on the Nscale platform.

Chat Use Case

End-to-end guide for chat inferencing.