The post TorchForge RL Pipelines Now Operable on Together AI’s Cloud appeared on BitcoinEthereumNews.com. Jessie A Ellis Dec 04, 2025 17:54 Together AI introduces TorchForge RL pipelines on its cloud platform, enhancing distributed training and sandboxed environments with a BlackJack training demo. TorchForge reinforcement learning (RL) pipelines are now seamlessly operable on Together AI’s Instant Clusters, offering robust support for distributed training, tool execution, and sandboxed environments, as demonstrated by an open-source BlackJack training demo, according to together.ai. The AI Native Cloud: Foundation for Next-Gen RL In the rapidly evolving field of reinforcement learning, building flexible and scalable systems necessitates compatible and efficient compute frameworks and tooling. Modern RL pipelines have transcended basic training loops, now relying heavily on distributed rollouts, high-throughput inference, and a coordinated use of CPU and GPU resources. The comprehensive PyTorch stack, inclusive of TorchForge and Monarch, now operates with distributed training capabilities on Together Instant Clusters. These clusters provide: Low-latency GPU communication: Utilizing InfiniBand/NVLink topologies for efficient RDMA-based data transfers and distributed actor messaging. Consistent cluster bring-up: Preconfigured with drivers, NCCL, CUDA, and the GPU operator, enabling PyTorch distributed jobs to run without manual setup. Heterogeneous RL workload scheduling: Optimized GPU nodes for policy replicas and trainers, alongside CPU-optimized nodes for environment and tool execution. Together AI’s clusters are aptly suited for RL frameworks that require a blend of GPU-bound model computation and CPU-bound environment workloads. Advanced Tool Integration and Demonstration A significant portion of RL workloads involves executing tools, running code, or interacting with sandboxed environments. Together AI’s platform natively supports these requirements through: Together CodeSandbox: MicroVM environments tailored for tool-use, coding tasks, and simulations. Together Code Interpreter: Facilitates fast, isolated Python execution suitable for unit-test-based reward functions or code-evaluation tasks. Both CodeSandbox and Code Interpreter integrate with OpenEnv and TorchForge environment services, allowing rollout workers to utilize these tools… The post TorchForge RL Pipelines Now Operable on Together AI’s Cloud appeared on BitcoinEthereumNews.com. Jessie A Ellis Dec 04, 2025 17:54 Together AI introduces TorchForge RL pipelines on its cloud platform, enhancing distributed training and sandboxed environments with a BlackJack training demo. TorchForge reinforcement learning (RL) pipelines are now seamlessly operable on Together AI’s Instant Clusters, offering robust support for distributed training, tool execution, and sandboxed environments, as demonstrated by an open-source BlackJack training demo, according to together.ai. The AI Native Cloud: Foundation for Next-Gen RL In the rapidly evolving field of reinforcement learning, building flexible and scalable systems necessitates compatible and efficient compute frameworks and tooling. Modern RL pipelines have transcended basic training loops, now relying heavily on distributed rollouts, high-throughput inference, and a coordinated use of CPU and GPU resources. The comprehensive PyTorch stack, inclusive of TorchForge and Monarch, now operates with distributed training capabilities on Together Instant Clusters. These clusters provide: Low-latency GPU communication: Utilizing InfiniBand/NVLink topologies for efficient RDMA-based data transfers and distributed actor messaging. Consistent cluster bring-up: Preconfigured with drivers, NCCL, CUDA, and the GPU operator, enabling PyTorch distributed jobs to run without manual setup. Heterogeneous RL workload scheduling: Optimized GPU nodes for policy replicas and trainers, alongside CPU-optimized nodes for environment and tool execution. Together AI’s clusters are aptly suited for RL frameworks that require a blend of GPU-bound model computation and CPU-bound environment workloads. Advanced Tool Integration and Demonstration A significant portion of RL workloads involves executing tools, running code, or interacting with sandboxed environments. Together AI’s platform natively supports these requirements through: Together CodeSandbox: MicroVM environments tailored for tool-use, coding tasks, and simulations. Together Code Interpreter: Facilitates fast, isolated Python execution suitable for unit-test-based reward functions or code-evaluation tasks. Both CodeSandbox and Code Interpreter integrate with OpenEnv and TorchForge environment services, allowing rollout workers to utilize these tools…

TorchForge RL Pipelines Now Operable on Together AI’s Cloud

For feedback or concerns regarding this content, please contact us at [email protected]


Jessie A Ellis
Dec 04, 2025 17:54

Together AI introduces TorchForge RL pipelines on its cloud platform, enhancing distributed training and sandboxed environments with a BlackJack training demo.

TorchForge reinforcement learning (RL) pipelines are now seamlessly operable on Together AI’s Instant Clusters, offering robust support for distributed training, tool execution, and sandboxed environments, as demonstrated by an open-source BlackJack training demo, according to together.ai.

The AI Native Cloud: Foundation for Next-Gen RL

In the rapidly evolving field of reinforcement learning, building flexible and scalable systems necessitates compatible and efficient compute frameworks and tooling. Modern RL pipelines have transcended basic training loops, now relying heavily on distributed rollouts, high-throughput inference, and a coordinated use of CPU and GPU resources.

The comprehensive PyTorch stack, inclusive of TorchForge and Monarch, now operates with distributed training capabilities on Together Instant Clusters. These clusters provide:

  • Low-latency GPU communication: Utilizing InfiniBand/NVLink topologies for efficient RDMA-based data transfers and distributed actor messaging.
  • Consistent cluster bring-up: Preconfigured with drivers, NCCL, CUDA, and the GPU operator, enabling PyTorch distributed jobs to run without manual setup.
  • Heterogeneous RL workload scheduling: Optimized GPU nodes for policy replicas and trainers, alongside CPU-optimized nodes for environment and tool execution.

Together AI’s clusters are aptly suited for RL frameworks that require a blend of GPU-bound model computation and CPU-bound environment workloads.

Advanced Tool Integration and Demonstration

A significant portion of RL workloads involves executing tools, running code, or interacting with sandboxed environments. Together AI’s platform natively supports these requirements through:

  • Together CodeSandbox: MicroVM environments tailored for tool-use, coding tasks, and simulations.
  • Together Code Interpreter: Facilitates fast, isolated Python execution suitable for unit-test-based reward functions or code-evaluation tasks.

Both CodeSandbox and Code Interpreter integrate with OpenEnv and TorchForge environment services, allowing rollout workers to utilize these tools during training.

BlackJack Training Demo

Together AI has released a demonstration of a TorchForge RL pipeline running on its Instant Clusters, interacting with an OpenEnv environment hosted on Together CodeSandbox. This demo, adapted from a Meta reference implementation, trains a Qwen 1.5B model to play BlackJack using GRPO. The RL pipeline integrates a vLLM policy server, BlackJack environment, reference model, off-policy replay buffer, and a TorchTitan trainer—connected through Monarch’s actor mesh and using TorchStore for weight synchronization.

The OpenEnv GRPO BlackJack repository includes Kubernetes manifests and setup scripts. Deployment and training initiation are streamlined with simple kubectl commands, allowing experimentation with model configurations and GRPO hyperparameter adjustments.

Additionally, a standalone integration wraps Together’s Code Interpreter as an OpenEnv environment, enabling RL agents to interact with the Interpreter like any other environment. This integration allows RL pipelines to be applied to diverse tasks such as coding and mathematical reasoning.

The demonstrations highlight that sophisticated, multi-component RL training can be conducted on the Together AI Cloud with ease, setting the stage for a flexible, open RL framework in the PyTorch ecosystem, scalable on the Together AI Cloud.

Image source: Shutterstock

Source: https://blockchain.news/news/torchforge-rl-pipelines-operable-together-ai-cloud

Market Opportunity
null Logo
null Price(null)
--
----
USD
null (null) Live Price Chart
Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact [email protected] for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.

You May Also Like

UK and US Seal $42 Billion Tech Pact Driving AI and Energy Future

UK and US Seal $42 Billion Tech Pact Driving AI and Energy Future

The post UK and US Seal $42 Billion Tech Pact Driving AI and Energy Future appeared on BitcoinEthereumNews.com. Key Highlights Microsoft and Google pledge billions as part of UK US tech partnership Nvidia to deploy 120,000 GPUs with British firm Nscale in Project Stargate Deal positions UK as an innovation hub rivaling global tech powers UK and US Seal $42 Billion Tech Pact Driving AI and Energy Future The UK and the US have signed a “Technological Prosperity Agreement” that paves the way for joint projects in artificial intelligence, quantum computing, and nuclear energy, according to Reuters. Donald Trump and King Charles review the guard of honour at Windsor Castle, 17 September 2025. Image: Kirsty Wigglesworth/Reuters The agreement was unveiled ahead of U.S. President Donald Trump’s second state visit to the UK, marking a historic moment in transatlantic technology cooperation. Billions Flow Into the UK Tech Sector As part of the deal, major American corporations pledged to invest $42 billion in the UK. Microsoft leads with a $30 billion investment to expand cloud and AI infrastructure, including the construction of a new supercomputer in Loughton. Nvidia will deploy 120,000 GPUs, including up to 60,000 Grace Blackwell Ultra chips—in partnership with the British company Nscale as part of Project Stargate. Google is contributing $6.8 billion to build a data center in Waltham Cross and expand DeepMind research. Other companies are joining as well. CoreWeave announced a $3.4 billion investment in data centers, while Salesforce, Scale AI, BlackRock, Oracle, and AWS confirmed additional investments ranging from hundreds of millions to several billion dollars. UK Positions Itself as a Global Innovation Hub British Prime Minister Keir Starmer said the deal could impact millions of lives across the Atlantic. He stressed that the UK aims to position itself as an investment hub with lighter regulations than the European Union. Nvidia spokesman David Hogan noted the significance of the agreement, saying it would…
Share
BitcoinEthereumNews2025/09/18 02:22
Will Ripple Be the Death of SWIFT?

Will Ripple Be the Death of SWIFT?

The post Will Ripple Be the Death of SWIFT? appeared first on Coinpedia Fintech News Over the past few months, Ripple vs SWIFT has been debated by many analysts. Multiple crypto experts even argued that Ripple will soon replace SWIFT because of its popularity in smooth payment transactions for overseas. Analysts also say that SWIFT’s old system will be overpowered by Ripple.  How Can Ripple Destroy SWIFT?  A crypto user …
Share
CoinPedia2025/09/18 20:28
XRP Ledger Plans to Become Native DeFi Lending Powerhouse

XRP Ledger Plans to Become Native DeFi Lending Powerhouse

The post XRP Ledger Plans to Become Native DeFi Lending Powerhouse appeared on BitcoinEthereumNews.com. The XLS-66 lending protocol, explained  The 80% validator
Share
BitcoinEthereumNews2026/03/08 15:53