SambaStack v0.5.14 Release
Release Date: April 1, 2026 This release introduces simplified checkpoint discovery, new model support (MiniMax-M2.5, Agentic RAG bundle), enhanced installation verification tools, and multiple API enhancements for improved OpenAI compatibility.For full deployment details, bundle configurations, and context length options for all models and bundles mentioned below, see Supported models and bundles.
New Features and Enhancements
Checkpoint Path Discovery via Model CRs
Checkpoint paths are now discoverable through the Model Custom Resource (CR), eliminating the need for customers to manually locate checkpoint paths in configuration files. The following Kubernetes command can be used to view Model CRs, which now contain checkpoint paths:- Model CRs now include checkpoint path information for all supported models.
- Supports multiple checkpoints for different model configurations.
- Backwards compatible with existing bundle configurations - checkpoint paths in bundle CRs override Model CR paths if specified.
- Works with on-prem and air-gapped deployments.
Improved Discovery of Available Models and Bundles
All available models and bundles are now applied automatically by the SambaStack helm chart. No additional configuration is required.- Models can be discovered using
kubectl -n <namespace> get models. - Bundles can be discovered using
kubectl -n <namespace> get bundles.
Agentic RAG Bundle
Added theus-agentic-rag-1-1 bundle, a multi-model bundle optimized for retrieval-augmented generation (RAG) workflows. It contains the following model configs:
gpt-oss-120b- Seq Length: 32K, BS: 4
- Seq Length: 64K, BS: 2
- Seq Length: 128K, BS: 2
Llama-4-Maverick-17B-128E-Instruct- Seq Length: 8K, BS: 1
- Seq Length: 16K, BS: 1
Meta-Llama-3.3-70B(Target) /Meta-Llama-3.2-1B(Draft)- Seq Length: 4K, BS: 1, 4, 8, 16, 32
- Seq Length: 8K, BS: 1, 4, 8
- Seq Length: 16K, BS: 1, 4
- Seq Length: 32K, BS: 1, 4
- Seq Length: 64K, BS: 1
- Seq Length: 128K, BS: 1
Meta-Llama-3.1-8B-Instruct- Seq Length: 4K, BS: 1, 4, 16, 32
- Seq Length: 8K, BS: 1, 4, 16, 32
- Seq Length: 16K, BS: 1, 4, 8
E5-Mistral-7B-Instruct- Seq Length: 4K, BS: 1, 4, 8, 16, 32
MiniMax-M2.5 Model Support
Added support for MiniMax-M2.5 model on SambaStack.- Checkpoint accessible via your artifact reader service account.
- Customers can include MiniMax-M2.5 in bundles that pass bundle validation and deploy successfully.
- Includes reasoning support.
Pre/Post-Install Verification Scripts
New verification scripts help customers validate their SambaStack environment before and after installation.- Pre-install script: Validates all hardware, connectivity, and software prerequisites.
- Post-install script: Confirms all SambaStack components are installed and running correctly.
- Clear pass/fail reporting with actionable guidance on failures.
- Scripts maintained and validated against the current SambaStack release.
- Distributed via the sambastack-tools public GitHub repository with README instructions.
Per-Model Queue Depth Configuration
Added support for configuring different queue depths based on context length, enabling optimized uptime for high-context-length requests.- Queue depths can now be configured per context length group using
contextGroupsin the bundle configuration. - Queue depth controls how many concurrent requests can be queued for a model configuration.
- Lower queue depths for higher context lengths help prevent memory exhaustion and improve overall service stability.
- SambaStack now validates queue depth configuration at request time. Misconfigured models with missing queue depth definitions will surface a clear error instead of failing silently.
- The empty string
""incontextLengthsmatches requests to the base model name without a context length suffix (e.g.,DeepSeek-R1-0528). Requests with explicit suffixes like-8kor-128kmatch their correspondingcontextLengthsvalues.
Context length suffixes (8k, 16k, 32k, etc.) are case-sensitive. Use lowercase
k in all configurations. This applies to all models supported by SambaNova.contextGroups field is a sub-component of a model grouping within a service tier. Example configuration:
Helm Chart Configuration Update
Thesubstitutions field has moved from bundles to global in the SambaStack Helm chart.
This is a breaking change that affects air-gapped and NFS customers. Update your Helm values file before upgrading.
API Improvements and Fixes
Enhancements to improve OpenAI compatibility across thechat/completions endpoint and a new, non-standard feature to track usage in streaming chunks.
Text Object Support in User Message Content
- Expanded support for text objects in content arrays, matching OpenAI
ChatCompletionsContentPartTextspecification. - Enabled for: DeepSeek-R1-0528, Llama-3.3-Swallow-70B-Instruct-v0.4, and MiniMax-M2.5.
Log Probabilities
- Added
logprobsfield that, when set totrue, returns the log probabilities for each generated token. - Added
top_logprobsfield that, when set to an integern, returns the topnlog probabilities for each generation.
Usage in Streaming Chunks
- Added a non-standard feature to allow users to obtain partial usage statistics in chunks returned in streaming responses.
- This feature is enabled by setting
STREAM_USAGE_IN_CHUNKS: truein the replica group section of your custom bundle deployment.
Tool Choices
Supported only for models that support function calling.tool_choice: noneensures that the model will not see available tools.
Invalid Message Role Validation
- The
chat/completionsendpoint now rejects invalid message roles. Onlyuser,assistant,system, andtoolare accepted.
Whisper Audio Error Response
- The Whisper transcription endpoint now returns descriptive error messages when audio file processing fails, instead of a bare HTTP 400 status code.
Bug Fixes
Function Calling Routing Fix
- Fixed an issue where function calling routing did not apply the model name prefix check correctly, causing some models to skip tool routing.
Air-Gap Inventory Fix
- Fixed the air-gap inventory to include the correct
cloudnative-pgimage configuration, preventing missing image errors during offline installation.
Known Issues
SambaRack Manager Does Not Support 2 PDU Configurations
SambaRack Manager does not currently support configurations with 2 PDUs. Customers using 2 PDU setups should contact SambaNova Support for guidance on alternative configurations.SambaStack v0.4.8 Release
Release Date: March 10, 2026 This release introduces air-gapped deployment support, custom checkpoint management with NFS storage, swappable model configurations, and multiple API enhancements for improved OpenAI compatibility.New Features and Enhancements
SambaStack Air-gapped Support
Added support for air-gapped mode of operation, enabling secure, isolated deployments.- Install, upgrade, and setup for air-gapped configurations is performed in conjunction with SambaNova support.
- Ongoing administration (Auth, User Management, Custom DB) is designed for self-service and follows the same workflows as on-prem deployments.
Install, setup, port forwarding to access Keycloak UI, and upgrade steps are not documented for air-gapped deployments due to varying customer network configurations. Please work with SambaNova support for these workflows.
Custom Checkpoints with NFS Storage
Added the ability to reference custom checkpoints from customer-provided NFS storage in deployments.Swappable Models in Bundles
Added configurable model swapping behavior to optimize high-bandwidth memory (HBM) utilization.- By default, all models in bundles can be swapped out of HBM and replaced with other models in DDR memory.
- Use the
swappable: <boolean>field in the bundle YAML definition to enable or disable this behavior. - Default value is
true. When set tofalse, the model remains in HBM and cannot be swapped out, ensuring zero switching time for requests to that model.
API Improvements and Fixes
Enhancements to improve OpenAI compatibility across thechat/completions endpoint.
Text Object Support in User Message Content
- Added support for text objects in content arrays, matching OpenAI
ChatCompletionsContentPartTextspecification. - Enabled for: gpt-oss-120b, DeepSeek-V3.1, DeepSeek-V3.1-Terminus, DeepSeek-V3.2, DeepSeek-V3-0324, Qwen3-32B, Qwen3-235B.
Response Format Text Option
- Fixed an issue where
response_format=textwould throw an error. - The endpoint now supports all OpenAI formats:
text,json_object,json_schema.
Extended Temperature Range
- Expanded
temperaturerange from 0.0–1.0 to 0.0–2.0, matching OpenAI specification.
Tool Calling Number Type Fix
- Tools with number-type arguments were always returned as floats.
- Now integers are preserved as integers, matching JSON Schema number specification.
Parallel Tool Calls Support
- Added
parallel_tool_callsparameter support. - When set to
false, the model will make at most one tool call per response, matching OpenAI specification.
Streaming Token Usage Reporting
- Added support for token usage reporting in each chunk of stream.
Known Issues
- Parallel Tool Calls with Constrained Decoding.
- The following models return
nullforlogprobseven whenlogprobs=trueortop_logprobsis set. The parameters are accepted without error but have no effect:- Llama-4-Maverick-17B-128E-Instruct
- Whisper-Large-v3
SambaStack Initial Release
Release Date: September 19, 2025 This release introduces the comprehensive SambaStack documentation suite.New Features and Enhancements
SambaStack Guide
Added the SambaStack Guide, providing step-by-step instructions for deploying, configuring, and managing SambaStack.- Setup, installation, and environment configuration.
- User and authentication management (Keycloak, OIDC).
- Monitoring, logging, and artifact management.
- Bundle and model deployment workflows.
- Common command reference.
SambaStack Models
Added the SambaStack models and bundles page to help customers understand which models are available on SambaStack and how to configure them.- Lists all supported models (e.g., Llama 3.3, Llama 4 Maverick, DeepSeek).
- Shows context length, batch size options, and supported features.
- Instructions for using the Model list API to check availability in your environment.
