The post Together AI Launches DeepSeek-V3.1: A Versatile Hybrid Model appeared on BitcoinEthereumNews.com. Terrill Dicki Aug 25, 2025 23:56 Together AI introduces DeepSeek-V3.1, a hybrid model offering fast responses and deep reasoning modes, ensuring efficiency and reliability for various applications. Together AI has unveiled DeepSeek-V3.1, an advanced hybrid model designed to cater to both fast response requirements and complex reasoning tasks. The model, now available for deployment on Together AI’s platform, is particularly noted for its dual-mode functionality, allowing users to select between non-thinking and thinking modes to optimize performance based on task complexity. Features and Capabilities DeepSeek-V3.1 is crafted to provide enhanced efficiency and reliability, according to Together AI. It supports serverless deployment with a 99.9% SLA, ensuring robust performance across a variety of use cases. The model’s thinking mode offers comparable quality to its predecessor, DeepSeek-R1, but with a significant improvement in speed, making it suitable for production environments. The model is built on a substantial training dataset, with 630 billion tokens for 32K context and 209 billion tokens for 128K context, enhancing its capability to handle extended conversations and large codebases. This ensures that the model is well-equipped for tasks that require detailed analysis and multi-step reasoning. Real-World Applications DeepSeek-V3.1 excels in various applications, including code and search agent tasks. In non-thinking mode, it efficiently handles routine tasks such as API endpoint generation and simple queries. In contrast, the thinking mode is ideal for complex problem-solving, such as debugging distributed systems and designing zero-downtime database migrations. For document processing, the model offers non-thinking capabilities for entity extraction and basic parsing, while thinking mode supports comprehensive analysis of compliance workflows and regulatory cross-referencing. Performance Metrics Benchmark tests reveal the model’s strengths in both modes. For instance, in the MMLU-Redux benchmark, the thinking mode achieved a 93.7% success rate, surpassing the non-thinking mode by… The post Together AI Launches DeepSeek-V3.1: A Versatile Hybrid Model appeared on BitcoinEthereumNews.com. Terrill Dicki Aug 25, 2025 23:56 Together AI introduces DeepSeek-V3.1, a hybrid model offering fast responses and deep reasoning modes, ensuring efficiency and reliability for various applications. Together AI has unveiled DeepSeek-V3.1, an advanced hybrid model designed to cater to both fast response requirements and complex reasoning tasks. The model, now available for deployment on Together AI’s platform, is particularly noted for its dual-mode functionality, allowing users to select between non-thinking and thinking modes to optimize performance based on task complexity. Features and Capabilities DeepSeek-V3.1 is crafted to provide enhanced efficiency and reliability, according to Together AI. It supports serverless deployment with a 99.9% SLA, ensuring robust performance across a variety of use cases. The model’s thinking mode offers comparable quality to its predecessor, DeepSeek-R1, but with a significant improvement in speed, making it suitable for production environments. The model is built on a substantial training dataset, with 630 billion tokens for 32K context and 209 billion tokens for 128K context, enhancing its capability to handle extended conversations and large codebases. This ensures that the model is well-equipped for tasks that require detailed analysis and multi-step reasoning. Real-World Applications DeepSeek-V3.1 excels in various applications, including code and search agent tasks. In non-thinking mode, it efficiently handles routine tasks such as API endpoint generation and simple queries. In contrast, the thinking mode is ideal for complex problem-solving, such as debugging distributed systems and designing zero-downtime database migrations. For document processing, the model offers non-thinking capabilities for entity extraction and basic parsing, while thinking mode supports comprehensive analysis of compliance workflows and regulatory cross-referencing. Performance Metrics Benchmark tests reveal the model’s strengths in both modes. For instance, in the MMLU-Redux benchmark, the thinking mode achieved a 93.7% success rate, surpassing the non-thinking mode by…

Together AI Launches DeepSeek-V3.1: A Versatile Hybrid Model



Terrill Dicki
Aug 25, 2025 23:56

Together AI introduces DeepSeek-V3.1, a hybrid model offering fast responses and deep reasoning modes, ensuring efficiency and reliability for various applications.





Together AI has unveiled DeepSeek-V3.1, an advanced hybrid model designed to cater to both fast response requirements and complex reasoning tasks. The model, now available for deployment on Together AI’s platform, is particularly noted for its dual-mode functionality, allowing users to select between non-thinking and thinking modes to optimize performance based on task complexity.

Features and Capabilities

DeepSeek-V3.1 is crafted to provide enhanced efficiency and reliability, according to Together AI. It supports serverless deployment with a 99.9% SLA, ensuring robust performance across a variety of use cases. The model’s thinking mode offers comparable quality to its predecessor, DeepSeek-R1, but with a significant improvement in speed, making it suitable for production environments.

The model is built on a substantial training dataset, with 630 billion tokens for 32K context and 209 billion tokens for 128K context, enhancing its capability to handle extended conversations and large codebases. This ensures that the model is well-equipped for tasks that require detailed analysis and multi-step reasoning.

Real-World Applications

DeepSeek-V3.1 excels in various applications, including code and search agent tasks. In non-thinking mode, it efficiently handles routine tasks such as API endpoint generation and simple queries. In contrast, the thinking mode is ideal for complex problem-solving, such as debugging distributed systems and designing zero-downtime database migrations.

For document processing, the model offers non-thinking capabilities for entity extraction and basic parsing, while thinking mode supports comprehensive analysis of compliance workflows and regulatory cross-referencing.

Performance Metrics

Benchmark tests reveal the model’s strengths in both modes. For instance, in the MMLU-Redux benchmark, the thinking mode achieved a 93.7% success rate, surpassing the non-thinking mode by 1.9%. Similarly, the GPQA-Diamond benchmark showed a 5.2% improvement in thinking mode. These metrics underscore the model’s ability to enhance performance across various tasks.

Deployment and Integration

DeepSeek-V3.1 is available through Together AI’s serverless API and dedicated endpoints, offering technical specifications with 671 billion total parameters and an MIT license for extensive application. The infrastructure is designed for reliability, featuring North American data centers and SOC 2 compliance.

Developers can swiftly integrate the model into their applications using the provided Python SDK, enabling seamless incorporation of DeepSeek-V3.1’s capabilities into existing systems. Together AI’s infrastructure supports large mixture-of-experts models, ensuring both thinking and non-thinking modes operate efficiently under production workloads.

With the launch of DeepSeek-V3.1, Together AI aims to provide a versatile solution for businesses seeking to enhance their AI-driven applications with both rapid response and deep analytical capabilities.

Image source: Shutterstock


Source: https://blockchain.news/news/together-ai-launches-deepseek-v3-1-versatile-hybrid-model

Market Opportunity
Mode Network Logo
Mode Network Price(MODE)
$0.0003645
$0.0003645$0.0003645
-7.46%
USD
Mode Network (MODE) Live Price Chart
Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact [email protected] for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.

You May Also Like

The Channel Factories We’ve Been Waiting For

The Channel Factories We’ve Been Waiting For

The post The Channel Factories We’ve Been Waiting For appeared on BitcoinEthereumNews.com. Visions of future technology are often prescient about the broad strokes while flubbing the details. The tablets in “2001: A Space Odyssey” do indeed look like iPads, but you never see the astronauts paying for subscriptions or wasting hours on Candy Crush.  Channel factories are one vision that arose early in the history of the Lightning Network to address some challenges that Lightning has faced from the beginning. Despite having grown to become Bitcoin’s most successful layer-2 scaling solution, with instant and low-fee payments, Lightning’s scale is limited by its reliance on payment channels. Although Lightning shifts most transactions off-chain, each payment channel still requires an on-chain transaction to open and (usually) another to close. As adoption grows, pressure on the blockchain grows with it. The need for a more scalable approach to managing channels is clear. Channel factories were supposed to meet this need, but where are they? In 2025, subnetworks are emerging that revive the impetus of channel factories with some new details that vastly increase their potential. They are natively interoperable with Lightning and achieve greater scale by allowing a group of participants to open a shared multisig UTXO and create multiple bilateral channels, which reduces the number of on-chain transactions and improves capital efficiency. Achieving greater scale by reducing complexity, Ark and Spark perform the same function as traditional channel factories with new designs and additional capabilities based on shared UTXOs.  Channel Factories 101 Channel factories have been around since the inception of Lightning. A factory is a multiparty contract where multiple users (not just two, as in a Dryja-Poon channel) cooperatively lock funds in a single multisig UTXO. They can open, close and update channels off-chain without updating the blockchain for each operation. Only when participants leave or the factory dissolves is an on-chain transaction…
Share
BitcoinEthereumNews2025/09/18 00:09
Franklin Templeton CEO Dismisses 50bps Rate Cut Ahead FOMC

Franklin Templeton CEO Dismisses 50bps Rate Cut Ahead FOMC

The post Franklin Templeton CEO Dismisses 50bps Rate Cut Ahead FOMC appeared on BitcoinEthereumNews.com. Franklin Templeton CEO Jenny Johnson has weighed in on whether the Federal Reserve should make a 25 basis points (bps) Fed rate cut or 50 bps cut. This comes ahead of the Fed decision today at today’s FOMC meeting, with the market pricing in a 25 bps cut. Bitcoin and the broader crypto market are currently trading flat ahead of the rate cut decision. Franklin Templeton CEO Weighs In On Potential FOMC Decision In a CNBC interview, Jenny Johnson said that she expects the Fed to make a 25 bps cut today instead of a 50 bps cut. She acknowledged the jobs data, which suggested that the labor market is weakening. However, she noted that this data is backward-looking, indicating that it doesn’t show the current state of the economy. She alluded to the wage growth, which she remarked is an indication of a robust labor market. She added that retail sales are up and that consumers are still spending, despite inflation being sticky at 3%, which makes a case for why the FOMC should opt against a 50-basis-point Fed rate cut. In line with this, the Franklin Templeton CEO said that she would go with a 25 bps rate cut if she were Jerome Powell. She remarked that the Fed still has the October and December FOMC meetings to make further cuts if the incoming data warrants it. Johnson also asserted that the data show a robust economy. However, she noted that there can’t be an argument for no Fed rate cut since Powell already signaled at Jackson Hole that they were likely to lower interest rates at this meeting due to concerns over a weakening labor market. Notably, her comment comes as experts argue for both sides on why the Fed should make a 25 bps cut or…
Share
BitcoinEthereumNews2025/09/18 00:36
Tom Lee’s Bitmine Scoops Up 3.4% of Ethereum, Triggering a Supply Squeeze

Tom Lee’s Bitmine Scoops Up 3.4% of Ethereum, Triggering a Supply Squeeze

Bitmine Immersion now controls 3.4% of Ethereum amid shrinking exchange supply and rising institutional accumulation.
Share
Crypto Breaking News2026/01/20 16:27