Stay updated on the latest developments of the SambaNova Cloud!
October 29, 2024
We’re thrilled to bring you the latest updates to SambaNova Cloud, featuring expanded model support, new multimodal capabilities, and usability improvements!
-
Llama 3.2 11B & 90B Models
Expanded Llama 3.2 models now include 11B and 90B versions, with multi-modality support for text and image inputs, enabling more versatile AI applications and use cases. -
Function Calling
The Function-Calling API enables dynamic, agentic workflows by allowing the model to suggest and select function calls based on user input. This feature facilitates flexible agentic workflows that adapt to varied needs. -
Multimodality in API and Playground
Interact with multimodal models directly through the Inference API (OpenAI compatible) and Playground for seamless text and image processing. -
Python and Gradio Code Samples for Faster Development
New Python and Gradio code samples make it easier to build and deploy applications on SambaNova Cloud. These examples simplify integrating AI models, enabling faster prototyping and reducing setup time. -
User Experience Improvements
- A ‘How to Use API’ guide provides a quick start with example cURL code for both text and image inputs.
- Streamlined access to updated code snippets for easier discoverability
- A new “Clear Chat” option making experimentation in the Playground even smoother.
- New UI components to give a smoother user experience with added tool tips.
-
Updated AI Starter Kits
Two new additions to our AI Starter Kits:- Multi Modal Retriever: Chart, Image, and Figure Understanding – Unlock insights from complex PDFs and images with advanced retrieval and answer generation that combines both visual and textual data.
- Llama 3.1 Instruct-o1: Enhanced Reasoning with Llama 3.1 405B – Experience advanced thinking capabilities with Llama 3.1 Instruct-o1, hosted on Hugging Face Spaces.
October 10, 2024
- Llama 3.1 8B model: max sequence length increased from 8k to 16k.
- Llama 3.1 70B model: max sequence length increased from 8k to 64k.
- Automatic Routing Based on Sequence Length:
You no longer need to change the model name to specify different sequence lengths. The system will automatically route requests based on sequence length. For example, there is no need to useMeta-Llama-3.1-70B-Instruct-8k
for the 8k sequence length anymore. While we still support the existing method for backward compatibility, we recommend switching to the new method for the best experience. - Improved performance for Llama 3.2 1B and 3B models.
October 1st, 2024
- Released Llama 3.2 1B and 3B models
- Available to all tiers at the fastest inference speed
September 10th, 2024
- Public launch of the SambaNova Cloud portal, API and the community
- Access to Llama 3.1 8B, 70B and 405B at full precision and 10x faster inference compared to GPUs
- Launched with two tiers - free and enterprise (paid)