<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:media="http://search.yahoo.com/mrss/"><channel><title><![CDATA[Bitdeer AI Cloud]]></title><description><![CDATA[Bitdeer's GPU Cloud is powered by NVIDIA DGX™ H100, specifically designed for large-scale HPC and AI workloads. 35x faster AI training, 20% more cost efficiency and reduce 50% in latency.]]></description><link>https://www.bitdeer.ai/en/blog/</link><image><url>https://www.bitdeer.ai/en/blog/favicon.png</url><title>Bitdeer AI Cloud</title><link>https://www.bitdeer.ai/en/blog/</link></image><generator>Ghost 5.82</generator><lastBuildDate>Thu, 30 Apr 2026 09:15:47 GMT</lastBuildDate><atom:link href="https://www.bitdeer.ai/en/blog/rss/" rel="self" type="application/rss+xml"/><ttl>60</ttl><item><title><![CDATA[Bring MultiModal Reasoning to Production with NVIDIA Nemotron 3 Nano Omni on Bitdeer AI Cloud]]></title><description><![CDATA[As AI agents evolve beyond text and into real-world workflows, the ability to understand information and reason across multiple modalities is becoming essential.]]></description><link>https://www.bitdeer.ai/en/blog/bring-omni-modal-understanding-to-production-with-nvidia-nemotron-3-nano-omni-on-bitdeer-ai-cloud/</link><guid isPermaLink="false">69eefae4b019420001ad9e7b</guid><category><![CDATA[AI Applications]]></category><dc:creator><![CDATA[Taylor Ye]]></dc:creator><pubDate>Tue, 28 Apr 2026 16:02:18 GMT</pubDate><media:content url="https://www.bitdeer.ai/en/blog/content/images/2026/04/NVIDIA-Nemotron-3-Nano-Omni-en--1-.png" medium="image"/><content:encoded><![CDATA[<img src="https://www.bitdeer.ai/en/blog/content/images/2026/04/NVIDIA-Nemotron-3-Nano-Omni-en--1-.png" alt="Bring MultiModal Reasoning to Production with NVIDIA Nemotron 3 Nano Omni on Bitdeer AI Cloud"><p>As AI agents evolve beyond text and into real-world workflows, the ability to understand information and reason across multiple modalities is becoming essential. From video and audio to documents and UI screens, modern agentic systems require models that can reason across modalities with accuracy, efficiency, and enterprise readiness.&#xA0;</p><p>Today, we&#x2019;re delighted to announce that <a href="https://developer.nvidia.com/blog/nvidia-nemotron-3-nano-omni-powers-multimodal-agent-reasoning-in-a-single-efficient-open-model?ref=bitdeer.ai">NVIDIA Nemotron&#x2122; 3 Nano Omni</a> is available at launch on our Bitdeer AI Model Studio. As part of the broader <a href="https://www.nvidia.com/en-us/ai-data-science/foundation-models/nemotron/?ref=bitdeer.ai">NVIDIA Nemotron</a> family, it represents a new step forward in open, production-ready multimodal reasoning models.</p><h2 id="what-is-nvidia-nemotron-3-nano-omni"><strong>What is NVIDIA Nemotron 3 Nano Omni</strong></h2><p>NVIDIA Nemotron 3 Nano Omni is an open multimodal foundation model designed for production agentic AI. It unifies reasoning across video, audio, images, documents, charts, and text in a single model&#x2014;eliminating the need for fragmented multimodal pipelines.</p><p>By consolidating perception and reasoning into one system, Nemotron 3 Nano Omni simplifies agent development, reduces orchestration complexity, improves efficiency and scalability, and delivers leading multimodal accuracy. Built with open weights, datasets, and recipes, it enables developers and enterprises to customize, deploy, and operate multimodal agents with full control and flexibility.</p><h3 id="key-specifications"><strong>Key Specifications:</strong></h3><p><strong>&#x25CF;&#xA0; Model Size</strong>: 30B A3B</p><p>&#x25CF;&#xA0; <strong>Modalities: </strong>Input (text, image, video, audio), Output (text).</p><p>&#x25CF;&#xA0; <strong>Architecture</strong>: Hybrid Mixture of Experts (MoE) with Transformer-Mamba design</p><p>&#x25CF;&#xA0; <strong>VIsion Encoder</strong>: CRADIO v4-H</p><p>&#x25CF;&#xA0; <strong>Audio ENcoder: </strong>Parakeet</p><p>&#x25CF;&#xA0; <strong>Context Length: </strong>256k</p><p>&#x25CF;&#xA0; &#xA0;&#xA0;&#xA0; <strong>Optimizations:</strong></p><ul><ul><li>Conv3D fortemporal video reasoning</li><li>Efficient Video Sampling (EVS) for lower inference cost</li></ul></ul><p>&#x25CF;&#xA0; &#xA0;&#xA0;&#xA0;<strong> Quantization</strong>: FP8, NVFP4</p><h2 id="why-a-unified-multimodal-model-matters"><strong>Why a Unified Multimodal Model Matters</strong></h2><p>Many enterprise AI systems still rely on stitched-together pipelines across vision, speech, OCR, and reasoning models. This approach introduces higher latency from repeated inference passes, increased operational complexity, and fragmented context across modalities.</p><p>Nemotron 3 Nano Omni addresses this by acting as a multimodal perception and reasoning layer within agent systems&#x2014;enabling a unified perception &#x2192; reasoning &#x2192; action loop.</p><h2 id="enterprise-use-cases"><strong>Enterprise use cases</strong></h2><p><strong>Customer Service Agent:</strong> A customer service agent operates in a highly multimodal environment. It analyzes recorded customer interactions, including audio and speech transcriptions. It reasons over screen recordings of customer sessions and images such as screenshots of errors or invoices. At the same time, it reads documents like knowledge&#x2011;base articles, policies, and CRM history. Nemotron 3 Nano Omni unifies all of these signals so the agent understands not just what the customer said, but what they experienced and what the business rules allow&#x2014;enabling accurate, context&#x2011;aware resolution.</p><p><strong>Financial Analyst Agent</strong>: Financial analysis depends on more than text alone. This agent reasons across documents like financial filings and earnings transcripts, images such as charts and scanned reports, audio and speech from earnings calls, and video from investor presentations. Nemotron 3 Nano Omni ties together what executives say, how numbers are presented visually, and what the underlying documents show&#x2014;producing grounded insights rather than surface&#x2011;level summaries.</p><p><strong>Computer Use Agent</strong>: The computer use agent is one of the clearest demonstrations of unified multimodality. It analyzes video and images from screen recordings to understand UI state over time, interprets instructions and system audio cues, and reads documents like task instructions and validation policies. Nemotron 3 Nano Omni enables the agent to see the interface, understand intent, read constraints, and take the correct action&#x2014;all within one reasoning loop. This collapses when perception and decision&#x2011;making are split across models.</p><p><strong>Media and Entertainment Agent:</strong> Media workflows depend on more than transcripts alone. This agent reasons across video content, dialogue, on-screen text, and visual scene changes to support richer video and speech analysis. Nemotron 3 Nano Omni can generate dense captions that capture not just what is said, but what appears and happens on screen, while also improving video search and summarization across large content libraries. This helps media teams turn raw footage into searchable, contextualized, and production-ready assets more efficiently.</p><h2 id="run-nemotron-3-nano-omni-via-api-on-bitdeer-ai-model-studio"><strong>Run Nemotron 3 Nano Omni Via API on Bitdeer AI Model Studio</strong></h2><p>You can run Nemotron 3 Nano Omni on Bitdeer AI Model Studio, our serverless inference platform designed to make access to advanced foundation models simple and scalable. With a straightforward API, Model Studio allows developers and enterprises to start using models quickly without managing underlying infrastructure, reducing deployment complexity and time to value. </p><p>This makes it easier to integrate multimodal reasoning capabilities into applications and agentic workflows, while benefiting from a more flexible and efficient path from experimentation to production in just a few steps.</p><h2 id="get-started"><strong>Get Started</strong></h2><ol><li>Log in to Bitdeer AI<a href="https://www.bitdeer.ai/en/model/explore?ref=bitdeer.ai"> </a><a href="https://www.bitdeer.ai/en/model/explore?ref=bitdeer.ai">Model Studio</a></li><li>Locate <a href="https://www.bitdeer.ai/en/model/explore/mo-d7o6c8vq7u0c73ca6img?ref=bitdeer.ai" rel="noreferrer">Nemotron 3 Nano Omni</a> in the model list</li><li>Generate API key and start making API calls</li></ol><p>This streamlined workflow enables rapid integration of multimodal reasoning into applications and agent systems.</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/04/image-5.png" class="kg-image" alt="Bring MultiModal Reasoning to Production with NVIDIA Nemotron 3 Nano Omni on Bitdeer AI Cloud" loading="lazy" width="2000" height="1006" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/04/image-5.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/04/image-5.png 1000w, https://www.bitdeer.ai/en/blog/content/images/size/w1600/2026/04/image-5.png 1600w, https://www.bitdeer.ai/en/blog/content/images/2026/04/image-5.png 2000w" sizes="(min-width: 720px) 720px"></figure><h2 id="conclusion"><strong>Conclusion</strong></h2><p>NVIDIA Nemotron 3 Nano Omni makes multimodal AI more practical for real-world deployments. Instead of managing multiple models and pipelines, teams can focus on building agentic applications, automating workflows, and delivering better user experiences.</p><p>With availability on Bitdeer AI Model Studio, organizations can move from experiment to production faster&#x2014;turning multimodal AI into measurable business impact.</p>]]></content:encoded></item><item><title><![CDATA[Why Your OpenClaw Should Run in the Cloud, Not on Your Laptop]]></title><description><![CDATA[Learn why running OpenClaw in the cloud improves reliability, security, and scalability, with guidance on model selection and real-world use cases.]]></description><link>https://www.bitdeer.ai/en/blog/why-your-openclaw-should-run-in-the-cloud-not-on-your-laptop/</link><guid isPermaLink="false">69e0909eb019420001ad9e49</guid><dc:creator><![CDATA[Yimian Ma]]></dc:creator><pubDate>Fri, 17 Apr 2026 06:00:05 GMT</pubDate><media:content url="https://www.bitdeer.ai/en/blog/content/images/2026/04/run-ai-agent-with-confidence-EN.png" medium="image"/><content:encoded><![CDATA[<img src="https://www.bitdeer.ai/en/blog/content/images/2026/04/run-ai-agent-with-confidence-EN.png" alt="Why Your OpenClaw Should Run in the Cloud, Not on Your Laptop"><p>OpenClaw has surpassed 290,000 stars on GitHub, making it one of the most popular open-source AI Agent frameworks of 2026. A growing number of developers and tech enthusiasts are experimenting with deploying their own AI assistants locally.</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/04/data-src-image-34197edf-31ce-4307-b91c-32c3838aeffc.png" class="kg-image" alt="Why Your OpenClaw Should Run in the Cloud, Not on Your Laptop" loading="lazy" width="1600" height="1156" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/04/data-src-image-34197edf-31ce-4307-b91c-32c3838aeffc.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/04/data-src-image-34197edf-31ce-4307-b91c-32c3838aeffc.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2026/04/data-src-image-34197edf-31ce-4307-b91c-32c3838aeffc.png 1600w" sizes="(min-width: 720px) 720px"></figure><p>However, local deployment presents significant challenges in practice: service interruptions when the machine shuts down, complex environment configuration, API key security risks, and more &#x2014; all of which severely limit the continuous availability of an AI Agent.</p><p>Deploying OpenClaw to the cloud addresses these issues at their root. This article examines the key advantages of cloud deployment, provides guidance on selecting the right large language model, and demonstrates real-world use cases for OpenClaw.</p><h2 id="what-is-openclaw"><strong>What Is OpenClaw?</strong></h2><p><a href="https://github.com/openclaw/openclaw?ref=bitdeer.ai"><u>OpenClaw</u></a> is an open-source personal AI Agent framework built with TypeScript. Its core capabilities include:</p><ul><li><strong>24/7 Autonomous Operation</strong>: A built-in heartbeat mechanism proactively monitors tasks and takes action without requiring continuous user intervention</li><li><strong>50+ Platform Integrations</strong>: Supports WhatsApp, Telegram, Discord, Slack, Gmail, GitHub, and other major platforms for unified cross-platform management</li><li><strong>Model Flexibility</strong>: Compatible with cloud-based LLMs such as Claude, GPT, and Grok, as well as local models via Ollama / vLLM</li><li><strong>Extensible Skills System</strong>: Skills are defined using Markdown + YAML, with 100+ pre-built skills available on <a href="https://docs.openclaw.ai/skills?ref=bitdeer.ai">ClawHub</a></li><li><strong>Privacy-First Design</strong>: All data is stored locally by default, with the memory system based on local Markdown files</li></ul><p>Unlike traditional chatbots, OpenClaw features contextual memory, proactive behavior, and tool invocation capabilities &#x2014; functioning more as a persistent AI collaborator than a simple Q&amp;A interface.</p><h2 id="cloud-deployment-vs-local-deployment"><strong>Cloud Deployment vs. Local Deployment</strong></h2><p>While OpenClaw supports local installation, cloud deployment offers clear advantages for users who require stability and continuous availability.</p>
<!--kg-card-begin: html-->
<table style="border:none;border-collapse:collapse;"><colgroup><col width="135"><col width="234"><col width="240"></colgroup><tbody><tr style="height:32.25pt"><td style="border-bottom:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><br></td><td style="border-bottom:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:700;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Local Deployment</span></p></td><td style="border-bottom:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:700;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Cloud Deployment (Bitdeer AI Cloud)</span></p></td></tr><tr style="height:32.25pt"><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:700;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Uptime</span></p></td><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Stops when the machine shuts down</span></p></td><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Runs 24/7 continuously</span></p></td></tr><tr style="height:32.25pt"><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:700;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Environment Setup</span></p></td><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Manual dependency and compatibility management</span></p></td><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Standardized images, ready to use</span></p></td></tr><tr style="height:32.25pt"><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:700;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Cost Model</span></p></td><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Upfront hardware investment + ongoing electricity</span></p></td><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Pay-as-you-go, no charge when stopped</span></p></td></tr><tr style="height:32.25pt"><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:700;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Security</span></p></td><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">API keys stored on local device</span></p></td><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Isolated cloud environment, keys secured server-side</span></p></td></tr><tr style="height:32.25pt"><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:700;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Network Quality</span></p></td><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Limited by local network conditions</span></p></td><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Enterprise-grade network, stable and low-latency</span></p></td></tr><tr style="height:32.25pt"><td style="border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:700;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Multi-Device Access</span></p></td><td style="border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Restricted to the host machine</span></p></td><td style="border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Access from any device via Telegram / Web</span></p></td></tr></tbody></table>
<!--kg-card-end: html-->
<p>Local deployment is suitable for short-term testing, while cloud deployment is what enables an AI Agent to deliver continuous, reliable service. An assistant that needs to be always available requires an environment that is always online.</p><h3 id="deployment-cost"><strong>Deployment Cost</strong></h3><p>Running OpenClaw does not require GPU resources. A basic CPU virtual machine instance (2 cores / 4GB RAM / 20GB SSD) is sufficient for stable operation. At this configuration, Bitdeer AI Cloud&apos;s on-demand pricing is approximately <strong>$0.0363/hour (around $26/month)</strong> &#x2014; providing a 24/7 online AI assistant environment at a fraction of typical infrastructure costs.</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/04/data-src-image-bd62b406-c5da-4b77-a8d8-94c3b2bae50e.png" class="kg-image" alt="Why Your OpenClaw Should Run in the Cloud, Not on Your Laptop" loading="lazy" width="1600" height="916" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/04/data-src-image-bd62b406-c5da-4b77-a8d8-94c3b2bae50e.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/04/data-src-image-bd62b406-c5da-4b77-a8d8-94c3b2bae50e.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2026/04/data-src-image-bd62b406-c5da-4b77-a8d8-94c3b2bae50e.png 1600w" sizes="(min-width: 720px) 720px"></figure><p>Bitdeer AI Cloud, as an NVIDIA Preferred Cloud Service Provider, offers compute resources spanning multiple generations of NVIDIA GPUs including H100, H200, B200, and GB200. For Agent deployment scenarios like OpenClaw, a CPU-only instance is all that&apos;s needed. For advanced use cases such as model fine-tuning or local inference, GPU instances are available for seamless scaling. Virtual machine instances are billed on-demand &#x2014; for detailed pricing, refer to the <a href="https://www.bitdeer.ai/en/pricing/gpu-compute?ref=bitdeer.ai"><u>GPU Compute Pricing</u></a> page.</p><p>Model inference is handled via API calls and billed per token. Bitdeer AI Cloud provides multi-vendor inference services covering text generation and other task types. For detailed inference pricing, refer to the <a href="https://www.bitdeer.ai/en/pricing/ai-models?ref=bitdeer.ai"><u>AI Models Pricing</u></a> page.</p><h2 id="model-selection-guide"><strong>Model Selection Guide</strong></h2><p>One of OpenClaw&apos;s core strengths is model flexibility &#x2014; it is not locked into any single LLM provider. Bitdeer AI Cloud currently offers 40+ models from 10+ providers, spanning text generation, visual understanding, reasoning, and image generation. All APIs are OpenAI-compatible, making model switching nearly effortless.</p><p>Below are representative text generation models suitable for OpenClaw, with pricing reference:</p>
<!--kg-card-begin: html-->
<table style="border:none;border-collapse:collapse;"><colgroup><col width="92"><col width="119"><col width="133"><col width="266"></colgroup><tbody><tr style="height:32.25pt"><td style="border-bottom:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:700;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Provider</span></p></td><td style="border-bottom:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:700;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Model</span></p></td><td style="border-bottom:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:700;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Input / Output (per 1M tokens)</span></p></td><td style="border-bottom:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:700;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Recommended Use Cases</span></p></td></tr><tr style="height:32.25pt"><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">DeepSeek</span></p></td><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:700;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">DeepSeek-V3.2</span></p></td><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">$0.28 / $0.42</span></p></td><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Code generation, technical reasoning &#x2014; best overall value</span></p></td></tr><tr style="height:32.25pt"><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Qwen</span></p></td><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:700;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Qwen3-235B-A22B</span></p></td><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">$0.22 / $0.88</span></p></td><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">MoE architecture, lowest input cost, multilingual support</span></p></td></tr><tr style="height:44.25pt"><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Moonshot AI</span></p></td><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:700;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Kimi-K2.5</span></p></td><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">$0.60 / $3.00</span></p></td><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Conversational AI, long-context comprehension, strong overall capability</span></p></td></tr><tr style="height:32.25pt"><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Zhipu AI</span></p></td><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:700;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">GLM-5</span></p></td><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">$1.00 / $3.20</span></p></td><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Flagship general conversation and knowledge Q&amp;A</span></p></td></tr><tr style="height:32.25pt"><td style="border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">MiniMax AI</span></p></td><td style="border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:700;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">MiniMax-M2.5</span></p></td><td style="border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">$0.30 / $1.20</span></p></td><td style="border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Reasoning model for complex logic and multi-step inference</span></p></td></tr></tbody></table>
<!--kg-card-end: html-->
<h3 id="scenario-based-recommendations"><strong>Scenario-Based Recommendations</strong></h3><ul><li><strong>Best overall value</strong> &#x2014; DeepSeek-V3.2, with the lowest combined input/output cost among flagship models and industry-leading code and reasoning capabilities</li><li><strong>Long-context input scenarios</strong> (e.g., document analysis, email summarization) &#x2014; Qwen3-235B-A22B, with input pricing as low as $0.22 per million tokens</li><li><strong>Deep conversational AI</strong> &#x2014; Kimi-K2.5, with excellent language comprehension for high-quality conversational output</li><li><strong>Ultra-low cost</strong> &#x2014; The platform also offers lightweight models such as Gemma and Ministral, with input pricing starting at $0.02 per million tokens for high-frequency, low-complexity tasks</li></ul><p>For the full model catalog and interactive demos, visit <a href="https://www.bitdeer.ai/en/model/explore?ref=bitdeer.ai"><u>Bitdeer Model Explore</u></a>.</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/04/data-src-image-4ce5c568-f2ac-466e-a35a-001059f63a02.png" class="kg-image" alt="Why Your OpenClaw Should Run in the Cloud, Not on Your Laptop" loading="lazy" width="1600" height="916" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/04/data-src-image-4ce5c568-f2ac-466e-a35a-001059f63a02.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/04/data-src-image-4ce5c568-f2ac-466e-a35a-001059f63a02.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2026/04/data-src-image-4ce5c568-f2ac-466e-a35a-001059f63a02.png 1600w" sizes="(min-width: 720px) 720px"></figure><p>OpenClaw integrates with Bitdeer&apos;s model service entirely through its configuration file &#x2014; users simply enter their API key and model name in openclaw.json, and OpenClaw handles all API communication automatically with no additional code required. All Bitdeer model APIs are OpenAI REST-compatible, allowing OpenClaw to recognize and invoke them directly.</p><p>API keys can be generated in the Bitdeer AI Cloud console under <a href="https://www.bitdeer.ai/en/model/apikeys?ref=bitdeer.ai"><u>API Keys</u></a>.</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/04/data-src-image-a987a9d8-953d-4af8-acb7-30ea135efacc.png" class="kg-image" alt="Why Your OpenClaw Should Run in the Cloud, Not on Your Laptop" loading="lazy" width="1600" height="762" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/04/data-src-image-a987a9d8-953d-4af8-acb7-30ea135efacc.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/04/data-src-image-a987a9d8-953d-4af8-acb7-30ea135efacc.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2026/04/data-src-image-a987a9d8-953d-4af8-acb7-30ea135efacc.png 1600w" sizes="(min-width: 720px) 720px"></figure><h2 id="deployment-guide"><strong>Deployment Guide</strong></h2><p>Bitdeer AI has published a comprehensive deployment tutorial covering everything from creating a cloud instance and installing OpenClaw to configuring model connections and integrating with Telegram. For the full step-by-step guide, refer to:</p><p><a href="https://www.bitdeer.ai/en/blog/installing-and-configuring-openclaw-on-bitdeer-ai-cloud/"><u>Installing and Configuring OpenClaw on Bitdeer AI Cloud</u></a></p><p>The deployment process can be summarized in five steps:</p><ol><li><strong>Create a Cloud Instance</strong> &#x2014; Provision a virtual machine on Bitdeer AI Cloud</li><li><strong>Install OpenClaw</strong> &#x2014; Run the official one-line installation script</li><li><strong>Configure Models</strong> &#x2014; Enter your Bitdeer AI API key and specify the LLM</li><li><strong>Connect Telegram</strong> &#x2014; Pair with a Telegram Bot for mobile access</li><li><strong>Run in Background</strong> &#x2014; The service runs continuously 24/7 with no manual maintenance required</li></ol><h2 id="real-world-use-cases"><strong>Real-World Use Cases</strong></h2><p>Once deployed, OpenClaw delivers practical value across multiple dimensions. Below are several representative use cases:</p><h3 id="personal-productivity"><strong>Personal Productivity</strong></h3><ul><li>Connect to Gmail to automatically read emails, extract action items, and send daily digests on a set schedule</li><li>Integrate with Google Calendar to create and manage events directly through conversation</li><li>Perform real-time multilingual translation across English, Chinese, Japanese, Korean, and more</li></ul><h3 id="research-and-analysis"><strong>Research and Analysis</strong></h3><ul><li>Execute web searches, read page content, and generate structured summaries</li><li>Extract key points and summaries from any given URL</li><li>Conduct multi-step research with comparative analysis and structured report output</li></ul><h3 id="code-generation-and-execution"><strong>Code Generation and Execution</strong></h3><ul><li>Generate and execute code in a sandboxed environment, returning results directly</li><li>Support for Python, JavaScript, and other mainstream languages for data processing and scripting</li><li>Combine web search capabilities with code generation to automatically reference documentation and produce runnable solutions</li></ul><p>These capabilities are powered by OpenClaw&apos;s skills system. Over 100 pre-built skills are available on <a href="https://docs.openclaw.ai/skills?ref=bitdeer.ai"><u>ClawHub</u></a> for immediate use. For custom skill development, refer to the <a href="https://docs.openclaw.ai/tools/creating-skills?ref=bitdeer.ai"><u>OpenClaw Skills Documentation</u></a>.</p><h2 id="conclusion"><strong>Conclusion</strong></h2><p>As a fully featured open-source AI Agent framework, OpenClaw &#x2014; combined with Bitdeer AI Cloud&apos;s infrastructure and model inference services &#x2014; provides a stable, secure, and always-online AI assistant solution.</p>
<!--kg-card-begin: html-->
<table style="border:none;border-collapse:collapse;"><colgroup><col width="198"><col width="407"></colgroup><tbody><tr style="height:19.5pt"><td style="border-bottom:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:700;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Requirement</span></p></td><td style="border-bottom:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:700;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Solution</span></p></td></tr><tr style="height:20.25pt"><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">24/7 AI assistant availability</span></p></td><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Deploy on Bitdeer AI Cloud</span></p></td></tr><tr style="height:20.25pt"><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Flexible model selection</span></p></td><td style="border-bottom:solid #000000 0.416667pt;border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Bitdeer AI inference service with multi-vendor model switching</span></p></td></tr><tr style="height:19.5pt"><td style="border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">Low deployment barrier</span></p></td><td style="border-top:solid #000000 0.416667pt;vertical-align:top;padding:4pt 8pt 4pt 8pt;overflow:hidden;overflow-wrap:break-word;"><p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:8pt;"><span style="font-size:10.5pt;font-family:Arial,sans-serif;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre;white-space:pre-wrap;">One-line install script + comprehensive deployment tutorial</span></p></td></tr></tbody></table>
<!--kg-card-end: html-->
<h3 id="get-staarted"><strong>Get Staarted</strong></h3><ul><li><a href="https://www.bitdeer.ai/en/instance/server?ref=bitdeer.ai">Sign up for Bitdeer AI Cloud</a></li><li><a href="https://www.bitdeer.ai/en/blog/installing-and-configuring-openclaw-on-bitdeer-ai-cloud/">View the full deployment tutorial</a></li><li><a href="https://github.com/openclaw/openclaw?ref=bitdeer.ai">Visit the OpenClaw GitHub repository</a></li><li><a href="https://www.bitdeer.ai/en/docs/center?ref=bitdeer.ai">Bitdeer AI Documentation Center</a></li></ul><p></p><p><em>Note:Pricing for models and GPU resources is subject to change. Please refer to the platform for the most up-to-date pricing.</em></p>]]></content:encoded></item><item><title><![CDATA[Getting Started with NemoClaw on Bitdeer AI Cloud]]></title><description><![CDATA[Deploy secure, production-ready AI agents in minutes with NVIDIA NemoClaw on Bitdeer AI Cloud. NemoClaw on Bitdeer AI Cloud brings together secure agent execution with scalable infrastructure and high-performance inference. ]]></description><link>https://www.bitdeer.ai/en/blog/getting-started-with-nemoclaw-on-bitdeer-ai-cloud/</link><guid isPermaLink="false">69d744f8185fb200015b8f9c</guid><category><![CDATA[AI Applications]]></category><dc:creator><![CDATA[Johan Sim]]></dc:creator><pubDate>Fri, 10 Apr 2026 06:10:05 GMT</pubDate><media:content url="https://www.bitdeer.ai/en/blog/content/images/2026/04/en.png" medium="image"/><content:encoded><![CDATA[<img src="https://www.bitdeer.ai/en/blog/content/images/2026/04/en.png" alt="Getting Started with NemoClaw on Bitdeer AI Cloud"><p>Deploy secure, production-ready AI agents in minutes with NVIDIA NemoClaw on Bitdeer AI Cloud. In this guide, you&#x2019;ll learn how to quickly set up a fully sandboxed environment and run powerful models like Moonshot AI&#x2019;s Kimi K2.5 through Bitdeer AI&#x2019;s high-performance inference endpoints.</p><h2 id="what-youll-build"><strong>What You&apos;ll Build</strong></h2><p>By the end of this guide, you&apos;ll have:</p><ul><li>A secure, sandboxed AI agent running on Bitdeer AI Cloud</li><li>Access to powerful models like Kimi K2.5 through Bitdeer AI&apos;s inference API</li><li>Full network policy controls and monitoring capabilities</li></ul><h2 id="why-nemoclaw-on-bitdeer-ai"><strong>Why NemoClaw on Bitdeer AI?</strong></h2><p>NemoClaw on Bitdeer AI Cloud brings together secure agent execution with scalable infrastructure and high-performance inference. While sandboxed execution with Landlock, seccomp, and network isolation is handled within the NemoClaw and OpenShell framework, Bitdeer AI Cloud can provide the underlying compute environment, enabling on-demand scaling of CPU and GPU resources with usage-based pricing. Our platform also delivers access to optimized inference endpoints for running leading models, along with network-level controls to help manage how agents interact with external services, ensuring both flexibility and operational control.</p><h2 id="prerequisites"><strong>Prerequisites</strong></h2><p><strong>NemoClaw System Requirements Minimum: </strong>4+ vCPU, 16 GB RAM, 40 GB of free disk space, and Ubuntu 22.04 LTS.</p><p><strong>One Bitdeer AI Cloud Requirements</strong></p><ul><li>Active <a href="https://www.bitdeer.ai/en/instance/server?ref=bitdeer.ai"><u>Bitdeer AI Cloud </u></a>account</li><li>API key from Bitdeer AI Cloud (for inference)</li></ul><p><strong>Bitdeer AI Instance: </strong>Select<strong> g4a.xlarge (</strong>4 vCPU, 16 GB RAM<strong>)</strong></p><p><strong>Important:</strong> The sandbox image is approximately 2.4 GB compressed. For instances with 16 GB RAM, you&apos;ll have sufficient memory for smooth operation.</p><h2 id="software-dependencies"><strong>Software Dependencies</strong></h2><ul><li>Node.js 20 or later</li><li>npm 10 or later</li><li>Docker (installed and running)</li><li>OpenShell CLI (installed by NemoClaw)</li></ul><h3 id="11-create-virtual-machine"><strong>1.1 Create Virtual Machine</strong></h3><p>1. Log in to your <a href="https://www.bitdeer.ai/en?ref=bitdeer.ai" rel="noreferrer">Bitdeer AI Cloud</a> account</p><p>2. Navigate to <strong>Virtual Machine</strong> services</p><p>3. Click <strong>Create Instance</strong> and select:</p><p><strong>Instance Type:</strong> <strong>g4a.xlarge</strong> (4 vCPU, 16 GB RAM)</p><p><strong>Storage:</strong> 40 GB SSD minimum</p><p><strong>OS:</strong> Ubuntu 22.04 LTS</p><p><strong>Region:</strong> Choose nearest to your location</p><p>4. Complete the order and wait for instance provisioning</p><p>5. Once running, note the public IP address</p><h3 id="12-access-your-instance"><strong>1.2 Access Your Instance</strong></h3><p>Connect via SSH:</p><pre><code class="language-bash">ssh ubuntu@&lt;your-instance-ip&gt;
</code></pre>
<h2 id="step-2-install-docker"><strong>Step 2: Install Docker</strong></h2><p>NemoClaw requires Docker as the container runtime:</p><pre><code class="language-bash"># Update package index
sudo apt-get update

# Install Docker (service starts automatically)
sudo apt-get install -y docker.io

# Add user to docker group
sudo usermod -aG docker $USER
</code></pre>
<p><strong>Important:</strong> Log out and log back in for group changes to take effect:</p>
<pre><code class="language-bash">exit
ssh ubuntu@&lt;your-instance-ip&gt;
</code></pre>
<p>Verify Docker installation:</p>
<pre><code class="language-bash">docker --version
docker run hello-world
</code></pre>
<h2 id="step-3-get-bitdeer-ai-api-key"><strong>Step 3: Get Bitdeer AI API Key</strong></h2><p>Before installing NemoClaw, obtain your Bitdeer AI API key:</p><p>1. Log in to Bitdeer AI Cloud</p><p>2. Navigate to <a href="https://www.bitdeer.ai/en/model/explore?ref=bitdeer.ai" rel="noreferrer"><strong>Models</strong> </a>&#x2192; <strong>API Keys</strong></p><p>3. Click <strong>Generate API Key</strong></p><p>4. Copy the generated key and save it securely</p><p><strong>&#x26A0; Security Warning:</strong> Never commit API keys to version control or share them publicly.</p><h2 id="step-4-install-and-configure-nemoclaw"><strong>Step 4: Install and Configure NemoClaw</strong></h2><p>Download and run the NemoClaw installer, which includes an interactive wizard for configuring your Bitdeer AI inference endpoint:</p><pre><code class="language-bash">curl -fsSL https://www.nvidia.com/nemoclaw.sh | bash
</code></pre>
<p><strong>During the Installation Wizard</strong></p><p>When prompted, configure Bitdeer AI as your inference provider:</p><pre><code>[2/7] Configuring inference (NIM)
&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;

Inference options:
  1) NVIDIA Endpoints (recommended)
  2) OpenAI
  3) Other OpenAI-compatible endpoint  &#x2190; Select this
  4) Anthropic
  5) Other Anthropic-compatible endpoint
  6) Google Gemini

Choose [1]: 3

OpenAI-compatible base URL: https://api-inference.bitdeer.ai/v1
Other OpenAI-compatible endpoint API key: &lt;paste-your-bitdeer-api-key&gt;

Other OpenAI-compatible endpoint model []: nvidia/NVIDIA-Nemotron-3-Super-120B-A12B
# or: moonshotai/Kimi-K2.5
</code></pre>
<p>The wizard will:</p><ul><li>Install Node.js 20+ if not present</li><li>Install npm packages</li><li>Set up the NemoClaw CLI</li><li>Configure Bitdeer AI as your inference provider</li><li>Save your API key securely to ~/.nemoclaw/credentials.json</li><li>Create your first sandbox</li></ul><p><strong>Post-Installation Summary</strong></p><p>After installation completes, you should see:</p><pre><code>&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;
Sandbox      my-bitdeer-agent (Landlock + seccomp + netns)
Model        nvidia/NVIDIA-Nemotron-3-Super-120B-A12B (Bitdeer AI Endpoint)
&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;
Run:         nemoclaw my-bitdeer-agent connect
Status:      nemoclaw my-bitdeer-agent status
Logs:        nemoclaw my-bitdeer-agent logs --follow
&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;&#x2500;

[INFO]  === Installation complete ===
</code></pre>
<p>If nemoclaw is not found after installation, run:</p><pre><code class="language-bash">source ~/.bashrc
# or for zsh users:
source ~/.zshrc
</code></pre>
<p>Verify installation:</p><pre><code class="language-bash">nemoclaw --version
</code></pre>
<h2 id="step-5-connect-and-chat-with-your-agent"><strong>Step 5: Connect and Chat with Your Agent</strong></h2><p><strong>5.1 Connect to the Sandbox</strong></p><pre><code class="language-bash">nemoclaw my-bitdeer-agent connect
</code></pre>
<p>You&apos;ll enter the sandbox shell:</p><pre><code>sandbox@my-bitdeer-agent:~$
</code></pre>
<p><strong>5.2 Interactive TUI Mode</strong></p><p>Launch the OpenClaw TUI for interactive chat:</p><pre><code class="language-bash">openclaw tui
</code></pre>
<p>Send test messages and verify responses.</p><p><strong>5.3 CLI Mode</strong></p><p>For programmatic access or long outputs, use the CLI:</p><pre><code class="language-bash">openclaw agent --agent main --local -m &quot;Hello, how can you help me today?&quot; --session-id test
</code></pre>
<p>To exit the sandbox shell:</p><pre><code class="language-bash">exit
</code></pre>
<h2 id="whats-next"><strong>What&apos;s Next?</strong></h2><p>Now that you have your NemoClaw agent running on Bitdeer AI Cloud, here are some ways to extend it:</p><ul><li><strong>Connect to Telegram</strong></li></ul><p>Enable mobile access to your agent:</p><pre><code class="language-bash">nemoclaw my-bitdeer-agent telegram-setup
</code></pre>
<ul><li><strong>Try Different Models</strong></li></ul><p>Experiment with other available models through Bitdeer AI Model Studio:</p><pre><code class="language-bash">openshell inference set --provider openai --model zai-org/GLM-5
</code></pre>
<ul><li><strong>Monitor Activity</strong></li></ul><p>Launch the OpenShell TUI to watch your agent in action:</p><pre><code class="language-bash">openshell tui
</code></pre>
<ul><li><strong>Customize Security</strong></li></ul><p>Define custom network policies to control what your agent can access.</p><h2 id="conclusion">Conclusion</h2><p>With NemoClaw running on Bitdeer AI Cloud, deploying secure, production-ready AI agents becomes both practical and scalable. By combining NemoClaw&#x2019;s sandboxed execution and policy framework with Bitdeer AI&#x2019;s flexible infrastructure and optimized inference endpoints, you can move from setup to real-world usage in just a few steps. Whether you are experimenting with AI agents or building toward production, Bitdeer AI Cloud offers a reliable environment to run, scale, and manage your workloads with confidence.</p><p></p><h3 id="resources"><strong>Resources</strong></h3><p><strong>Documentation:</strong></p><ul><li><a href="https://docs.nvidia.com/nemoclaw/latest/?ref=bitdeer.ai" rel="noreferrer">NemoClaw Docs</a></li><li><a href="https://docs.nvidia.com/openshell/latest/?ref=bitdeer.ai" rel="noreferrer">OpenShell Docs</a></li><li><a href="https://www.bitdeer.ai/en/instance/server?ref=bitdeer.ai" rel="noreferrer">Bitdeer AI Cloud</a></li></ul><p><strong>Community &amp; Support:</strong></p><ul><li><a href="https://github.com/NVIDIA/NemoClaw?ref=bitdeer.ai" rel="noreferrer">NemoClaw GitHub</a></li><li><a href="NemoClaw Discord" rel="noreferrer">NemoClaw Discord</a></li><li><a href="https://www.bitdeer.ai/en/contact?ref=bitdeer.ai" rel="noreferrer">Bitdeer AI Support</a></li></ul><p></p>]]></content:encoded></item><item><title><![CDATA[​​From Safer Runtime to Scalable Agent Deployment with NVIDIA OpenShell on Bitdeer AI Cloud]]></title><description><![CDATA[Explore how NVIDIA OpenShell enables secure, scalable AI agents with stronger privacy and control, and run safer AI agents on Bitdeer AI Cloud.]]></description><link>https://www.bitdeer.ai/en/blog/from-safer-runtime-to-scalable-agent-deployment-with-nvidia-openshell-on-bitdeer-ai-cloud/</link><guid isPermaLink="false">69ce1da1185fb200015b8f76</guid><category><![CDATA[AI Applications]]></category><dc:creator><![CDATA[Taylor Ye]]></dc:creator><pubDate>Mon, 06 Apr 2026 05:58:08 GMT</pubDate><media:content url="https://www.bitdeer.ai/en/blog/content/images/2026/04/en-blog.png" medium="image"/><content:encoded><![CDATA[<img src="https://www.bitdeer.ai/en/blog/content/images/2026/04/en-blog.png" alt="&#x200B;&#x200B;From Safer Runtime to Scalable Agent Deployment with NVIDIA OpenShell on Bitdeer AI Cloud"><p>Recently, the emergence of OpenClaw drew attention to a growing demand for a new class of AI systems. Often referred to as &#x201C;claws,&#x201D; these systems represent long-running, autonomous, self-evolving agents that can plan and execute multi-step workflows with limited user intervention. Unlike traditional AI assistants, claws are not limited to single-turn interactions. They can access local files, interact with applications and external tools, and dynamically orchestrate sub-agents to break down complex tasks. More importantly, they are able to continuously refine strategies, decide where tasks should be executed, and optimize outcomes over time.</p><p>At the same time, this evolution introduces new challenges. As agents are granted broader access to data, tools, and infrastructure, concerns around security, privacy, and governance become significantly more critical. Questions around access control, data flow, model usage, and execution boundaries are no longer theoretical, but essential for deploying these systems safely in production.</p><p>To address the growing need for safer and more controllable agent execution, at NVIDIA GTC 2026, NVIDIA introduced<a href="https://www.nvidia.com/en-us/ai/nemoclaw/?ref=bitdeer.ai"> <u>NVIDIA NemoClaw</u></a> for OpenClaw, an open source stack that adds privacy and security controls to long-running, autonomous agents. Built on<a href="https://developer.nvidia.com/nemo-agent-toolkit?ref=bitdeer.ai"> <u>NVIDIA Agent Toolkit</u></a>, NemoClaw serves as an integration layer that enables OpenClaw agents to run within the newly introduced<a href="https://docs.nvidia.com/openshell/index.html?ref=bitdeer.ai"> <u>NVIDIA OpenShell</u></a> runtime, combining open models such as<a href="https://www.nvidia.com/en-us/ai-data-science/foundation-models/nemotron/?ref=bitdeer.ai"> <u>NVIDIA Nemotron</u></a> with controlled execution environments to make always-on agents safer and more practical to deploy.</p><h2 id="nvidia-openshell-powering-safer-agent-execution-for-nemoclaw"><strong>NVIDIA OpenShell: Powering Safer Agent Execution for NemoClaw</strong></h2><p><a href="https://developer.nvidia.com/blog/run-autonomous-self-evolving-agents-more-safely-with-nvidia-openshell/?ref=bitdeer.ai"><u>NVIDIA OpenShell</u></a> is an open<u>-</u>source runtime designed to make autonomous AI agents safer and more practical to deploy.&#xA0; It addresses this by acting as a control layer between the agent and the underlying infrastructure, governing how the agent executes, what it can access, and where inference is routed.</p><p>Instead of relying solely on prompts or built-in guardrails inside the agent, OpenShell enforces policies at the runtime level. It helps bridge the gap between experimentation and production by allowing developers to deploy autonomous, self-evolving agents with clearer boundaries around privacy, security, and infrastructure access.</p><h2 id="how-openshell-is-built-for-safer-agents"><strong>How OpenShell Is Built for Safer Agents</strong></h2><p>OpenShell uses a gateway to manage one or more isolated sandboxes where agents run. To understand how it works, it helps to look at it from two complementary angles: core components and protection layers. The components describe how the system is structured and managed, while the protection layers describe where policy enforcement is applied.</p><ul><li>The gateway acts as the control plane, coordinating sandbox lifecycle, authentication, providers, and policy management.</li><li>The sandbox is the isolated runtime where the agent runs. The policy engine enforces constraints across filesystem, network, and process behavior.</li><li>The privacy router governs inference routing, helping keep sensitive context under tighter control and directing model traffic according to defined privacy and cost policies.</li></ul><p>These components enforce control across four protection layers: filesystem, network, process, and inference. Filesystem and process controls are locked when the sandbox is created, while network and inference controls can be updated at runtime. Together, this gives OpenShell a defense-in-depth model that supports long-running agents while maintaining clearer privacy, security, and governance boundaries.</p><p>To better understand how OpenShell enables safer autonomous agents, it helps to look at how its architecture is structured. The diagram below illustrates the key components that work together to govern execution, enforce policy, and control inference.&#xA0;</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/04/data-src-image-f91bcc4d-2f8e-46b3-83f7-3fab56dafaca.png" class="kg-image" alt="&#x200B;&#x200B;From Safer Runtime to Scalable Agent Deployment with NVIDIA OpenShell on Bitdeer AI Cloud" loading="lazy" width="1600" height="972" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/04/data-src-image-f91bcc4d-2f8e-46b3-83f7-3fab56dafaca.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/04/data-src-image-f91bcc4d-2f8e-46b3-83f7-3fab56dafaca.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2026/04/data-src-image-f91bcc4d-2f8e-46b3-83f7-3fab56dafaca.png 1600w" sizes="(min-width: 720px) 720px"></figure><p><em>OpenShell&#x2019;s architecture for safer autonomous agents, illustrating the core components: the sandbox, the policy engine, and the privacy router</em></p><h2 id="deploy-ai-agent-workloads-with-bitdeer-ai-cloud"><strong>Deploy AI Agent Workloads with Bitdeer AI Cloud</strong></h2><p>Bitdeer AI Cloud is among the first NVIDIA Cloud Partners (NCPs) in Asia Pacific available as a model provider in the region, helping developers support the model deployment and inference behind OpenShell-based agent workloads</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/04/image-3.png" class="kg-image" alt="&#x200B;&#x200B;From Safer Runtime to Scalable Agent Deployment with NVIDIA OpenShell on Bitdeer AI Cloud" loading="lazy" width="2000" height="1071" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/04/image-3.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/04/image-3.png 1000w, https://www.bitdeer.ai/en/blog/content/images/size/w1600/2026/04/image-3.png 1600w, https://www.bitdeer.ai/en/blog/content/images/2026/04/image-3.png 2000w" sizes="(min-width: 720px) 720px"></figure><p>Bitdeer AI Cloud provides access to a wide range of high-performance NVIDIA AI infrastructure, including<a href="https://www.nvidia.com/en-us/data-center/gb200-nvl72/?ref=bitdeer.ai"> <u>NVIDIA GB200 NVL72</u></a>,<a href="https://www.nvidia.com/en-us/data-center/hgx/?ref=bitdeer.ai"> <u>NVIDIA HGX H200</u></a>. Workloads can be deployed across multiple GPU configurations, with built-in support for load balancing and dynamic scaling to optimize performance, cost, and availability.</p><p>This enables developers to power the model deployment and inference required by agent workloads with the performance and reliability needed for production.</p><p><strong>How to Get Started</strong></p><p>Running models with Bitdeer AI Cloud is straightforward.&#xA0; Start building on<a href="http://build.nvidia.com/?ref=bitdeer.ai"> build.nvidia.com</a>, deploy to Bitdeer AI Cloud:</p><p>First, select a model from the NVIDIA model catalog, such as<a href="https://build.nvidia.com/moonshotai/kimi-k2.5/deploy.?ref=bitdeer.ai"> <u>Kimi K2.5</u></a> or Nemotron 3 Super, and navigate to the Deploy section.</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/04/image-1.png" class="kg-image" alt="&#x200B;&#x200B;From Safer Runtime to Scalable Agent Deployment with NVIDIA OpenShell on Bitdeer AI Cloud" loading="lazy" width="2000" height="1030" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/04/image-1.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/04/image-1.png 1000w, https://www.bitdeer.ai/en/blog/content/images/size/w1600/2026/04/image-1.png 1600w, https://www.bitdeer.ai/en/blog/content/images/2026/04/image-1.png 2000w" sizes="(min-width: 720px) 720px"></figure><p>Then, under Partner Endpoints, choose Bitdeer AI as your model provider. This allows the underlying model deployment and inference for agent workloads to be powered by Bitdeer AI Cloud, backed by scalable, production-grade NVIDIA AI infrastructure.</p><p>Once selected, you can proceed with deployment using serverless endpoints or dedicated GPU-backed infrastructure, depending on your workload requirements.</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/04/image-2.png" class="kg-image" alt="&#x200B;&#x200B;From Safer Runtime to Scalable Agent Deployment with NVIDIA OpenShell on Bitdeer AI Cloud" loading="lazy" width="2000" height="1032" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/04/image-2.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/04/image-2.png 1000w, https://www.bitdeer.ai/en/blog/content/images/size/w1600/2026/04/image-2.png 1600w, https://www.bitdeer.ai/en/blog/content/images/2026/04/image-2.png 2000w" sizes="(min-width: 720px) 720px"></figure><p>From there, OpenShell-based agents can leverage this setup to execute tasks, call tools, and operate within controlled environments, while running on scalable, production-grade compute.</p><p>By combining NVIDIA OpenShell&#x2019;s secure runtime with Bitdeer AI&#x2019;s infrastructure, developers can:</p><ul><li>Run long-running agents on dedicated, high-performance GPU clusters</li><li>Scale workloads without managing underlying infrastructure</li><li>Maintain stronger control over privacy, security, and inference routing</li></ul><p>This integration bridges the gap between agent development and real-world deployment, making it easier to move from experimentation to scalable production systems.</p><h2 id="closing-thoughts"><strong>Closing Thoughts</strong></h2><p>What stood out at NVIDIA GTC 2026 is that AI is shifting from prompt-based interactions to long-running, autonomous systems. As token consumption continues to grow, it is increasingly driven by agents themselves, reasoning over longer contexts, coordinating tools, and executing tasks over time.</p><p>This evolution brings new requirements beyond model performance. It becomes equally important how agents are executed, how data is governed, and how systems remain controlled in production environments. NVIDIA OpenShell reflects this shift by introducing a runtime layer designed for safer and more structured agent execution.</p><p>While OpenShell introduces a safer runtime layer for agent execution, Bitdeer AI Cloud serves as a model provider, delivering the scalable compute foundation needed to support real-world deployment of these systems. By providing scalable, high-performance AI compute, we aim to support developers in building and operating long-running agents within environments that prioritize privacy, security, and reliability.</p><p>&#xA0;</p><p><strong>Resource:</strong></p><ol><li>NVIDIA<u> </u>Developer Guide:<a href="https://docs.nvidia.com/openshell/latest/about/overview.html?ref=bitdeer.ai">&#xA0;</a> <a href="https://docs.nvidia.com/openshell/latest/about/overview.html?ref=bitdeer.ai"><u>https://docs.nvidia.com/openshell/latest/about/overview.html</u></a></li><li>NVIDIA blog: https://developer.nvidia.com/blog/run-autonomous-self-evolving-agents-more-safely-with-nvidia-openshell/</li></ol>]]></content:encoded></item><item><title><![CDATA[NVIDIA GTC 2026: The Inference Inflection and the Rise of Agentic AI Factories]]></title><description><![CDATA[Explore NVIDIA GTC 2026 keynote highlights and how Bitdeer AI Cloud powers NVIDIA OpenShell as an integrated inference provider for agent deployment.]]></description><link>https://www.bitdeer.ai/en/blog/nvidia-gtc-2026-the-inference-inflection-and-the-rise-of-agentic-ai-factories/</link><guid isPermaLink="false">69c243c568d90b0001fa6fe8</guid><category><![CDATA[AI Trends & Industry News]]></category><category><![CDATA[AI Applications]]></category><dc:creator><![CDATA[Tey Rui Jie]]></dc:creator><pubDate>Tue, 24 Mar 2026 08:12:18 GMT</pubDate><media:content url="https://www.bitdeer.ai/en/blog/content/images/2026/03/first-nvidia-nemoclaw-provider-in-apac-EN.png" medium="image"/><content:encoded><![CDATA[<h3 id="key-highlights"><strong>Key Highlights</strong></h3><ul><li>The Inference Inflection: GTC keynote explained massive shift in compute demand as AI moves from pre-training to real-time agentic inference.</li><li>The<a href="https://www.nvidia.com/en-us/data-center/technologies/rubin/?ref=bitdeer.ai"> <u>NVIDIA Vera Rubin Platform</u></a>: A &quot;generational leap&quot; featuring seven new chips (including<a href="https://www.nvidia.com/en-us/data-center/vera-cpu/?ref=bitdeer.ai"> <u>NVIDIA Vera CPU</u></a> and NVIDIA Rubin GPU) and five specialized racks designed to power the world&#x2019;s largest AI factories.</li><li>Disaggregated Inference: The newly integrated<a href="https://www.nvidia.com/en-us/data-center/lpx/?ref=bitdeer.ai"> <u>NVIDIA Groq 3 LPX rack</u></a> unites with the Vera Rubin platform to deliver up to 35x higher inference throughput per megawatt.</li></ul><img src="https://www.bitdeer.ai/en/blog/content/images/2026/03/first-nvidia-nemoclaw-provider-in-apac-EN.png" alt="NVIDIA GTC 2026: The Inference Inflection and the Rise of Agentic AI Factories"><p>The &quot;Moment of Claws&quot;: NVIDIA is driving the next AI inflection point through<a href="https://docs.nvidia.com/openshell/index.html?ref=bitdeer.ai"> <u>NVIDIA OpenShell,</u></a> an open-source runtime that enables autonomous agents to reason and act safely within isolated sandboxes.</p><h3 id="the-1-trillion-inference-inflection"><strong>The $1 Trillion Inference Inflection</strong></h3><p>At GTC 2026, the keynote underscored the massive shift from pre-training models to real-time execution, Huang revealed a staggering jump in the demand for AI infrastructure</p><p>The doubling of demand is driven by the era of agentic AI&#x2014;systems that plan tasks, run tools, and validate results. Computing is no longer retrieval-based; it is growing from generative, to reasoning-oriented and finally agentic. Now, as AI Agents begin to perform different tasks through iterative processes of planning, breaking down tasks, reasoning and reflecting, the burst of inference compute demand has grown by roughly<a href="https://blogs.nvidia.com/blog/gtc-2026-news/?ref=bitdeer.ai"> <u>1,000,000x in just two years</u></a>.</p><h3 id="vera-rubin-the-foundation-of-the-ai-factory"><strong>Vera Rubin: The Foundation of the AI Factory</strong></h3><p>NVIDIA launched the Vera Rubin platform, a supercomputer where multiple racks work together as one massive, coherent system to maximize <u>t</u>okens per watt.</p><ul><li>NVIDIA Groq 3 LPX Rack: Marks a milestone in accelerated computing by enabling disaggregated inference. While the Rubin GPU handles the prefill, decode attention, and generating massive KV cache, the NVIDIA Groq LPU manages the low-latency decode phase, providing up to 10x more revenue opportunity for trillion-parameter models.</li><li>Vera CPU Rack: The world&#x2019;s first processor purpose-built for agentic AI. It delivers results twice as efficiently and 50% faster than traditional CPUs. With 88 custom Olympus cores and 1.5TB of LPDDR5X memory, it provides the single-threaded performance required for agents to run code and validate results.</li><li><a href="https://nvidianews.nvidia.com/news/nvidia-launches-bluefield-4-stx-storage-architecture-with-broad-industry-adoption?ref=bitdeer.ai">NVIDIA</a>&#xA0; <a href="https://nvidianews.nvidia.com/news/nvidia-launches-bluefield-4-stx-storage-architecture-with-broad-industry-adoption?ref=bitdeer.ai"><u>BlueField-4 STX Rack</u></a>: An AI-native storage infrastructure<u> </u>that treats data as &quot;Context Memory.&quot; Using the<a href="https://developer.nvidia.com/networking/doca?ref=bitdeer.ai"> <u>NVIDIA DOCA Memos&#x2122; framework</u></a>, it boosts inference throughput by up to 5x for long-context reasoning</li></ul><h3 id="nvidia-openshell-enabling-the-moment-of-claws"><strong>NVIDIA OpenShell: Enabling the &quot;Moment of Claws&quot;</strong></h3><p>In early 2026, the emergence of OpenClaw drew attention to the demand for autonomous agents capable of executing multi-step workflows. By enabling agents to interact with local files and external tools, frameworks like OpenClaw illustrated a shift toward persistent, execution-oriented systems.</p><p>To support the safe deployment of these SOTA agents, NVIDIA is celebrating the &#x201C;moment of claws&#x201D; with NVIDIA OpenShell, an open-source runtime designed to govern how agents execute.</p><p>We&#x2019;re also working with NVIDIA on<a href="https://www.nvidia.com/nemoclaw?ref=bitdeer.ai"> <u>NVIDIA NemoClaw</u></a> &#x2014; an open source stack that simplifies running OpenClaw always-on assistants, more safely, with a single command. As part of the NVIDIA Agent Toolkit, it installs the NVIDIA OpenShell runtime&#x2014;a secure environment for running autonomous agents, and open source models like<a href="https://developer.nvidia.com/nemotron?ncid=pa-srch-goog-599191&amp;_bt=797127771541&amp;_bk=nvidia+nemotron&amp;_bm=p&amp;_bn=g&amp;_bg=194751055082&amp;gad_source=1&amp;gad_campaignid=23551395576&amp;gbraid=0AAAAAD4XAoGmJfKE4w2qyi2jQz5MY-SLa&amp;gclid=Cj0KCQjwmunNBhDbARIsAOndKpmxfYwV4wZZwublo2vQVn5Tm5IgcadyFiRgsbaILMgf3Mzkuyy27HoaAvjDEALw_wcB&amp;ref=bitdeer.ai"> <u>NVIDIA Nemotron</u></a>.</p><p>NVIDIA OpenShell sits between the agent and the infrastructure, running any coding agent&#x2014;OpenClaw, Claude Code, Cursor, or Codex&#x2014;in an isolated sandbox with zero code changes. Every action is policy-enforced at the infrastructure layer, ensuring privacy and security for long-running agentic workflows.</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/03/image.png" class="kg-image" alt="NVIDIA GTC 2026: The Inference Inflection and the Rise of Agentic AI Factories" loading="lazy" width="1466" height="652" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/03/image.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/03/image.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2026/03/image.png 1466w" sizes="(min-width: 720px) 720px"></figure><p><em>Source: NVIDIA GTC 2026 Keynote</em></p><p>In tandem with these global advancements, Bitdeer AI is leading the charge for secure agentic infrastructure in the Asia Pacific region. Developers and enterprises can now find Bitdeer AI featured on build.nvidia.com as a provider for the infrastructure necessary to run the underlying &quot;brains&quot; of these agents.</p><p>We are also proud to highlight that <a href="https://www.bitdeer.ai/en/model/explore/mo-d6psr309pjks73dh7hcg?ref=bitdeer.ai"><u>Nemotron 3 Super</u></a> is available on our platform. As the best-performing open model for OpenClaw, Nemotron 3 Super provides the high-level reasoning and planning capabilities required to fuel complex autonomous workflows.</p><p><strong>Start building your own safe autonomous agent today directly from build.nvidia.com:</strong></p><ol><li>Access NVIDIA OpenShell on <a href="http://build.nvidia.com/?ref=bitdeer.ai"><u>build.nvidia.com</u></a>.</li><li>Seamlessly integrate frontier open model endpoints from Bitdeer AI to power your long-running agents.</li></ol><p>Powered by<a href="https://www.nvidia.com/en-us/data-center/gb200-nvl72/?ref=bitdeer.ai"> <u>NVIDIA GB200 NVL72</u></a> systems deployed in our state-of-the-art AI data centers, Bitdeer AI Cloud delivers the high-performance infrastructure required to run autonomous, self-evolving agents in NVIDIA OpenShell at scale. By combining secure runtime governance with high-performance AI cloud, we empower developers to build the next generation of AI with stronger privacy and infrastructure reliability</p><p></p><p></p><p><em>Source: </em><a href="https://blogs.nvidia.com/blog/gtc-2026-news/?ref=bitdeer.ai"><em><u>https://blogs.nvidia.com/blog/gtc-2026-news/</u></em></a></p>]]></content:encoded></item><item><title><![CDATA[Build and Run Agentic AI with NVIDIA Nemotron 3 Super on Bitdeer AI Model Studio]]></title><description><![CDATA[Bitdeer AI Model Studio now supports NVIDIA-Nemotron-3-Super-120B-A12B, bringing a new open model for advanced reasoning, long-context understanding, and agentic AI development to the platform.]]></description><link>https://www.bitdeer.ai/en/blog/build-and-run-agentic-ai-with-nvidia-nemotron-3-super-on-bitdeer-ai-model-studio/</link><guid isPermaLink="false">69b5071c68d90b0001fa6fc5</guid><category><![CDATA[AI Applications]]></category><dc:creator><![CDATA[Taylor Ye]]></dc:creator><pubDate>Sat, 14 Mar 2026 07:18:30 GMT</pubDate><media:content url="https://www.bitdeer.ai/en/blog/content/images/2026/03/NVIDIA-Nemotron-3-Super-120B-A12B-Now-Live-on-Bitdeer-AI-Model-Studio.png" medium="image"/><content:encoded><![CDATA[<img src="https://www.bitdeer.ai/en/blog/content/images/2026/03/NVIDIA-Nemotron-3-Super-120B-A12B-Now-Live-on-Bitdeer-AI-Model-Studio.png" alt="Build and Run Agentic AI with NVIDIA Nemotron 3 Super on Bitdeer AI Model Studio"><p>Bitdeer AI <a href="https://www.bitdeer.ai/en/model/explore?ref=bitdeer.ai"><u>Model Studio</u></a> now supports NVIDIA-Nemotron-3-Super-120B-A12B, bringing a new open model for advanced reasoning, long-context understanding, and agentic AI development to the platform.</p><p>Designed for complex multi-agent applications, Nemotron 3 Super combines strong reasoning accuracy with high compute efficiency. Developers can now explore and run this model on Bitdeer AI&#x2019;s high performance NVIDIA GPU infrastructure, making it easier to experiment, deploy, and scale demanding AI workloads.</p><h2 id="what-is-nvidia-nemotron-3-super"><strong>What is NVIDIA Nemotron 3 Super</strong></h2><p>NVIDIA Nemotron 3 Super, part of the Nemotron 3 family of open models, is optimized for complex multi-agent applications. NVIDIA-Nemotron-3-Super-120B-A12B is a 120B-parameter model built on a hybrid Mamba-Transformer Mixture-of-Experts (MoE) architecture, activating only 12B parameters during inference. This design delivers strong performance on complex reasoning tasks while remaining efficient enough for production-scale deployment.</p><p>What makes this model particularly notable is its focus on agentic workloads. Rather than being optimized only for general chat, Nemotron 3 Super is built for scenarios where models need to reason across long contexts, call tools accurately, and support multi-step task execution.</p><h3 id="key-specifications"><strong>Key Specifications:</strong></h3><ul><li><strong>Model Size</strong>: 120B total parameters with 12B active parameters during inference.</li><li><strong>Architecture</strong>: Hybrid Mamba-2 + SSM + Transformer with Mixture-of-Experts (MoE) routing.</li><li><strong>Context Length</strong>: Supports up to 1M tokens for long-context reasoning.</li><li><strong>Accuracy</strong>: Leading accuracy on the Artificial Analysis Intelligence Index within its model size category.</li><li><strong>Minimum GPU Requirement</strong>: 2&#xD7; H100-80GB</li><li><strong>Multilingual Support</strong>: English, French, German, Italian, Japanese, Spanish, Chinese</li></ul><h3 id="key-architecture-innovations"><strong>Key architecture innovations</strong></h3><p>Agentic AI systems require models capable of advanced reasoning, coding, and long-context analysis, while remaining efficient enough to operate continuously at production scale.</p><p>However, multi-agent workflows introduce additional complexity. Compared with traditional chat interactions, these systems generate significantly more tokens as agents repeatedly exchange conversation history, tool outputs, and intermediate reasoning steps across multiple turns. Over long tasks, this expanded context can increase inference costs and introduce challenges such as communication overhead and context drift.</p><p>To address these requirements, Nemotron 3 Super is built around several key architectural design principles.</p><ul><li>First, the hybrid Mamba-Transformer architecture improves the model&#x2019;s ability to process long sequences efficiently while retaining the reasoning strengths of transformer layers.</li><li>Second, the Mixture-of-Experts routing mechanism dynamically selects specialized expert networks for each task, enabling higher performance without proportionally increasing inference cost.</li><li>Third, the model is designed as a fully open model stack, including open weights, training datasets, and training recipes. This allows developers to customize and fine-tune the model on their own infrastructure while maintaining control over data privacy and security.</li></ul><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/03/data-src-image-bef098f7-85bc-4877-b466-fccc74d6a2bf.png" class="kg-image" alt="Build and Run Agentic AI with NVIDIA Nemotron 3 Super on Bitdeer AI Model Studio" loading="lazy" width="1696" height="313" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/03/data-src-image-bef098f7-85bc-4877-b466-fccc74d6a2bf.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/03/data-src-image-bef098f7-85bc-4877-b466-fccc74d6a2bf.png 1000w, https://www.bitdeer.ai/en/blog/content/images/size/w1600/2026/03/data-src-image-bef098f7-85bc-4877-b466-fccc74d6a2bf.png 1600w, https://www.bitdeer.ai/en/blog/content/images/2026/03/data-src-image-bef098f7-85bc-4877-b466-fccc74d6a2bf.png 1696w" sizes="(min-width: 720px) 720px"></figure><p><em>A layer pattern diagram showing repeating blocks of Mamba-2/MoE pairs interleaved with attention layers</em></p><p><em>Source: </em><a href="https://developer.nvidia.com/blog/introducing-nemotron-3-super-an-open-hybrid-mamba-transformer-moe-for-agentic-reasoning/?ref=bitdeer.ai" rel="noreferrer"><em>NVIDIA</em></a></p><p>Together, these capabilities make Nemotron 3 Super particularly well suited for building complex multi-agent systems and enterprise AI workflows. And Nemotron 3 Super is optimized for tool calling, multi-step reasoning, and agent orchestration. These capabilities are essential for emerging AI agent systems, where models must interact with external tools, APIs, and data sources to complete tasks.</p><h3 id="benchmark-performance"><strong>Benchmark performance</strong></h3><p>Nemotron 3 Super delivers leading accuracy across several agentic and reasoning benchmarks while maintaining high throughput efficiency. The model demonstrates strong capabilities in instruction following, coding, tool use, and long-context reasoning.</p><p>The benchmark results below highlight its performance across key agentic evaluation tasks.</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/03/data-src-image-8ab1ddd9-3c76-4160-b4f7-e9a667db4cc1.png" class="kg-image" alt="Build and Run Agentic AI with NVIDIA Nemotron 3 Super on Bitdeer AI Model Studio" loading="lazy" width="1152" height="432" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/03/data-src-image-8ab1ddd9-3c76-4160-b4f7-e9a667db4cc1.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/03/data-src-image-8ab1ddd9-3c76-4160-b4f7-e9a667db4cc1.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2026/03/data-src-image-8ab1ddd9-3c76-4160-b4f7-e9a667db4cc1.png 1152w" sizes="(min-width: 720px) 720px"></figure><p><em>A chart comparing Nemotron 3 Super accuracy on key benchmarks against similarly sized open models.</em></p><p>Source: <a href="https://developer.nvidia.com/blog/introducing-nemotron-3-super-an-open-hybrid-mamba-transformer-moe-for-agentic-reasoning/?ref=bitdeer.ai"><u>NVIDIA</u></a>&#xA0;</p><h2 id="enterprise-use-cases"><strong>Enterprise use cases</strong></h2><p>Nemotron 3 Super is designed for complex enterprise AI applications, particularly those that involve multi-agent reasoning and large knowledge contexts. These capabilities make the model well suited for a variety of industry use cases where advanced reasoning and scalable AI workflows are required. Below are several representative examples&#xFF1A;&#xA0;</p><p><strong>Software Development:</strong>&#xA0;</p><p>With strong coding and tool-calling capabilities, as well as the ability to work across large codebases, Nemotron 3 Super can support end-to-end software development workflows, including code generation, automated debugging, and testing.</p><p><strong>Deep Research and Search:</strong>&#xA0;</p><p>Nemotron 3 Super has demonstrated strong performance in enterprise research workflows, with capabilities well suited for comprehensive report generation and precise factual recall. This makes it a strong fit for literature review, competitive intelligence, and research automation.</p><p>Cybersecurity: High-accuracy tool calling enables autonomous agents to select the appropriate tools more reliably, reducing execution errors in high-stakes environments such as security operations and automated cybersecurity workflows.</p><p><strong>Financial Services:</strong>&#xA0;</p><p>Nemotron 3 Super can process large volumes of financial reports and long-context documents in a single workflow, helping analyst agents maintain context over extended tasks and improving both efficiency and analytical accuracy.</p><h1 id="run-nemotron-3-super-via-api-on-bitdeer-ai-model-studio"><strong>Run Nemotron 3 Super via API on Bitdeer AI Model Studio</strong></h1><p>Bitdeer AI Model Studio is a serverless inference platform where you can immediately access and run foundation models like Nemotron 3 Super through a simple API.&#xA0;</p><p><strong>Step 1: Select the Model</strong></p><p>Log in to Bitdeer AI <a href="https://www.bitdeer.ai/en/model/explore?ref=bitdeer.ai"><u>Model Studio</u></a> , locate NVIDIA-Nemotron-3-Super-120B-A12B in the model list, and start using it for your AI workloads, priced at&#xA0; $0.30/M in &#xB7;$0.80/M out.&#xA0;</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/03/--------2-.png" class="kg-image" alt="Build and Run Agentic AI with NVIDIA Nemotron 3 Super on Bitdeer AI Model Studio" loading="lazy" width="1600" height="900" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/03/--------2-.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/03/--------2-.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2026/03/--------2-.png 1600w" sizes="(min-width: 720px) 720px"></figure><h3 id="step-2-generate-an-api-key"><strong>Step 2: Generate an API Key</strong></h3><p>Before using the model, you need an API key. Go to API Keys in the left navigation panel and click Generate API Key.&#xA0;</p><h3 id="step-3-call-the-model-via-api"><strong>Step 3: Call the Model via API</strong></h3><p>Once the API key is created, you can call the model using the Bitdeer AI inference API by including your API key in the request header.</p><p>With Model Studio and API access, developers can quickly start experimenting with Nemotron 3 Super and integrate the model into their AI applications.</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/03/data-src-image-f7cf047f-20a8-4a79-83ad-d26b8b02d638.png" class="kg-image" alt="Build and Run Agentic AI with NVIDIA Nemotron 3 Super on Bitdeer AI Model Studio" loading="lazy" width="2000" height="1012" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/03/data-src-image-f7cf047f-20a8-4a79-83ad-d26b8b02d638.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/03/data-src-image-f7cf047f-20a8-4a79-83ad-d26b8b02d638.png 1000w, https://www.bitdeer.ai/en/blog/content/images/size/w1600/2026/03/data-src-image-f7cf047f-20a8-4a79-83ad-d26b8b02d638.png 1600w, https://www.bitdeer.ai/en/blog/content/images/2026/03/data-src-image-f7cf047f-20a8-4a79-83ad-d26b8b02d638.png 2048w" sizes="(min-width: 720px) 720px"></figure><h2 id="conclusion"><strong>Conclusion</strong></h2><p>With strong reasoning capabilities, long-context support, and an efficient hybrid architecture, NVIDIA-Nemotron-3-Super-120B-A12B provides a powerful foundation for building advanced AI applications. From complex reasoning and coding tasks to large-scale knowledge processing and agentic workflows, the model enables developers to tackle demanding workloads with greater efficiency and flexibility.&#xA0;</p><p>Ready to experience NVIDIA-Nemotron-3-Super-120B-A12B in action? Explore the model on Bitdeer AI Model Studio and see how it can accelerate your next AI applications.</p>]]></content:encoded></item><item><title><![CDATA[Installing and Configuring OpenClaw on Bitdeer AI Cloud]]></title><description><![CDATA[Deploy OpenClaw on Bitdeer AI Cloud with this step-by-step guide. Build AI agents with web search, code execution, and Telegram integration.]]></description><link>https://www.bitdeer.ai/en/blog/installing-and-configuring-openclaw-on-bitdeer-ai-cloud/</link><guid isPermaLink="false">698199f468d90b0001fa6f55</guid><category><![CDATA[AI Applications]]></category><category><![CDATA[AI Trends & Industry News]]></category><dc:creator><![CDATA[Cyrus Cao]]></dc:creator><pubDate>Tue, 03 Feb 2026 10:30:59 GMT</pubDate><media:content url="https://www.bitdeer.ai/en/blog/content/images/2026/02/blog-run-openclaw-on-bitdeer-ai-cloud-EN-1.png" medium="image"/><content:encoded><![CDATA[<img src="https://www.bitdeer.ai/en/blog/content/images/2026/02/blog-run-openclaw-on-bitdeer-ai-cloud-EN-1.png" alt="Installing and Configuring OpenClaw on Bitdeer AI Cloud"><p>OpenClaw is a powerful AI agent framework that enables you to build intelligent assistants with multiple capabilities, including web search, code execution, and integration with various AI models. With Bitdeer AI Cloud, OpenClaw can be deployed online with flexible resource allocation and usage-based pricing, making it easier to run AI agents without long-term hardware constraints.</p><p> In this guide, we&apos;ll walk through the complete process of deploying OpenClaw on Bitdeer AI Cloud and configuring it to work with Telegram for an interactive AI assistant experience.</p><h2 id="prerequisites"><strong>Prerequisites</strong></h2><p>Before getting started, ensure you have:</p><ul><li>A Bitdeer AI Cloud account</li><li>Basic knowledge of Linux command line</li><li>A Telegram account for bot integration</li></ul><h2 id="step-1-setting-up-bitdeer-ai-cloud-instance"><strong>Step 1: Setting Up Bitdeer AI Cloud Instance</strong></h2><h3 id="ordering-a-virtual-machine"><strong>Ordering a Virtual Machine</strong></h3><p>First, navigate to the Bitdeer AI Cloud platform and place an order for a virtual machine instance.</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-4b830819-9a5c-49e1-b708-1a1bc3b22440.png" class="kg-image" alt="Installing and Configuring OpenClaw on Bitdeer AI Cloud" loading="lazy" width="1600" height="825" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/02/data-src-image-4b830819-9a5c-49e1-b708-1a1bc3b22440.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/02/data-src-image-4b830819-9a5c-49e1-b708-1a1bc3b22440.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-4b830819-9a5c-49e1-b708-1a1bc3b22440.png 1600w" sizes="(min-width: 720px) 720px"></figure><p>Select an appropriate instance configuration based on your needs. For OpenClaw, we recommend:</p><ul><li><strong>CPU</strong>: At least 2 cores</li><li><strong>RAM</strong>: Minimum 4GB</li><li><strong>Storage</strong>: At least 20GB SSD</li><li><strong>OS</strong>: Ubuntu 20.04 or later</li></ul><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-a24b321d-5e3f-4bbf-b16f-a2bc7e0861ef.png" class="kg-image" alt="Installing and Configuring OpenClaw on Bitdeer AI Cloud" loading="lazy" width="1600" height="834" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/02/data-src-image-a24b321d-5e3f-4bbf-b16f-a2bc7e0861ef.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/02/data-src-image-a24b321d-5e3f-4bbf-b16f-a2bc7e0861ef.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-a24b321d-5e3f-4bbf-b16f-a2bc7e0861ef.png 1600w" sizes="(min-width: 720px) 720px"></figure><p>Once your order is confirmed, wait for the instance to be provisioned.</p><h3 id="accessing-your-instance"><strong>Accessing Your Instance</strong></h3><p>After the instance is ready, you&apos;ll see it in your dashboard with a &quot;Running&quot; status.</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-80f720b7-9156-45a1-b5d7-649c7803a6d1.png" class="kg-image" alt="Installing and Configuring OpenClaw on Bitdeer AI Cloud" loading="lazy" width="1600" height="778" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/02/data-src-image-80f720b7-9156-45a1-b5d7-649c7803a6d1.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/02/data-src-image-80f720b7-9156-45a1-b5d7-649c7803a6d1.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-80f720b7-9156-45a1-b5d7-649c7803a6d1.png 1600w" sizes="(min-width: 720px) 720px"></figure><p>Connect to your instance using SSH:</p><pre><code class="language-shell">ssh username@your-instance-ip
</code></pre>
<h2 id="step-2-installing-openclaw"><strong>Step 2: Installing OpenClaw</strong></h2><h3 id="quick-installation"><strong>Quick Installation</strong></h3><p>OpenClaw provides a simple one-line installation script that handles all dependencies automatically:</p><pre><code class="language-bash">curl -fsSL https://openclaw.ai/install.sh | bash -s -- --no-onboard
</code></pre>
<figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-f9776657-d4f6-4064-8791-a85a32d1c8b9.png" class="kg-image" alt="Installing and Configuring OpenClaw on Bitdeer AI Cloud" loading="lazy" width="1600" height="1291" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/02/data-src-image-f9776657-d4f6-4064-8791-a85a32d1c8b9.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/02/data-src-image-f9776657-d4f6-4064-8791-a85a32d1c8b9.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-f9776657-d4f6-4064-8791-a85a32d1c8b9.png 1600w" sizes="(min-width: 720px) 720px"></figure><p>This command will:</p><ul><li>Install all required system dependencies</li><li>Set up Node.js environment</li><li>Install OpenClaw</li></ul><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-dc6e18c3-5fcd-4c3d-b636-ec8c9353b0ee.png" class="kg-image" alt="Installing and Configuring OpenClaw on Bitdeer AI Cloud" loading="lazy" width="1600" height="1291" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/02/data-src-image-dc6e18c3-5fcd-4c3d-b636-ec8c9353b0ee.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/02/data-src-image-dc6e18c3-5fcd-4c3d-b636-ec8c9353b0ee.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-dc6e18c3-5fcd-4c3d-b636-ec8c9353b0ee.png 1600w" sizes="(min-width: 720px) 720px"></figure><p>Once the installation is completed, verify it:</p><pre><code class="language-bash">openclaw --version
</code></pre>
<figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-f1521c56-f5c6-4bf5-ab38-57b7f7b2913b.png" class="kg-image" alt="Installing and Configuring OpenClaw on Bitdeer AI Cloud" loading="lazy" width="1600" height="1291" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/02/data-src-image-f1521c56-f5c6-4bf5-ab38-57b7f7b2913b.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/02/data-src-image-f1521c56-f5c6-4bf5-ab38-57b7f7b2913b.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-f1521c56-f5c6-4bf5-ab38-57b7f7b2913b.png 1600w" sizes="(min-width: 720px) 720px"></figure><p><strong>&#x1F4A1; Tip:</strong> If you encounter openclaw: <em>command not found</em>, you may need to source your <em>.bashrc</em> file for the current session or update your PATH:</p><blockquote>
<pre><code class="language-bash">source ~/.bashrc
# Or manually add to PATH if necessary
export PATH=&quot;$HOME/.npm-global/bin:$PATH&quot;&quot;
</code></pre>
</blockquote>
<h2 id="step-3-configuring-openclaw"><strong>Step 3: Configuring OpenClaw</strong></h2><h3 id="initial-setup"><strong>Initial Setup</strong></h3><p>Run the OpenClaw setup wizard to initialize your environment:</p><pre><code class="language-bash">openclaw setup
</code></pre>
<figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-ef45ae7d-1320-4715-9063-0e67e5b55eb9.png" class="kg-image" alt="Installing and Configuring OpenClaw on Bitdeer AI Cloud" loading="lazy" width="1600" height="1213" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/02/data-src-image-ef45ae7d-1320-4715-9063-0e67e5b55eb9.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/02/data-src-image-ef45ae7d-1320-4715-9063-0e67e5b55eb9.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-ef45ae7d-1320-4715-9063-0e67e5b55eb9.png 1600w" sizes="(min-width: 720px) 720px"></figure><p>This will create the basic configuration structure for OpenClaw.</p><h3 id="configuring-ai-models"><strong>Configuring AI Models</strong></h3><p>After the setup completes, edit the OpenClaw configuration file to add your AI model settings:</p><pre><code class="language-bash">vim ~/.openclaw/openclaw.json
</code></pre>
<figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-31495117-b78f-478e-b0d2-bc1aaa599f44.png" class="kg-image" alt="Installing and Configuring OpenClaw on Bitdeer AI Cloud" loading="lazy" width="1600" height="1291" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/02/data-src-image-31495117-b78f-478e-b0d2-bc1aaa599f44.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/02/data-src-image-31495117-b78f-478e-b0d2-bc1aaa599f44.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-31495117-b78f-478e-b0d2-bc1aaa599f44.png 1600w" sizes="(min-width: 720px) 720px"></figure><p>Update the configuration with your preferred models:</p><pre><code class="language-json">{
  &quot;agents&quot;: {
    &quot;defaults&quot;: {
      &quot;model&quot;: {
        &quot;primary&quot;: &quot;bitdeerai/moonshotai/Kimi-K2.5&quot;
      },
      &quot;workspace&quot;: &quot;/home/ubuntu/.openclaw/workspace&quot;
    }
  },
  &quot;models&quot;: {
    &quot;mode&quot;: &quot;merge&quot;,
    &quot;providers&quot;: {
      &quot;bitdeerai&quot;: {
        &quot;baseUrl&quot;: &quot;https://api-inference.bitdeer.ai/v1&quot;,
        &quot;apiKey&quot;: &quot;your-api-key-here&quot;,
        &quot;api&quot;: &quot;openai-completions&quot;,
        &quot;models&quot;: [
          {
            &quot;id&quot;: &quot;moonshotai/Kimi-K2.5&quot;,
            &quot;name&quot;: &quot;Kimi-K2.5&quot;
          }
        ]
      }
    }
  }
}
</code></pre>
<figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-ec72eaae-e769-4e7e-b456-55dafc9592ca.png" class="kg-image" alt="Installing and Configuring OpenClaw on Bitdeer AI Cloud" loading="lazy" width="1600" height="777" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/02/data-src-image-ec72eaae-e769-4e7e-b456-55dafc9592ca.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/02/data-src-image-ec72eaae-e769-4e7e-b456-55dafc9592ca.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-ec72eaae-e769-4e7e-b456-55dafc9592ca.png 1600w" sizes="(min-width: 720px) 720px"></figure><p>You can find more details in model card here:</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-7a9bc1c2-8ddf-4484-b4e6-6eb32637a70d.png" class="kg-image" alt="Installing and Configuring OpenClaw on Bitdeer AI Cloud" loading="lazy" width="1600" height="758" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/02/data-src-image-7a9bc1c2-8ddf-4484-b4e6-6eb32637a70d.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/02/data-src-image-7a9bc1c2-8ddf-4484-b4e6-6eb32637a70d.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-7a9bc1c2-8ddf-4484-b4e6-6eb32637a70d.png 1600w" sizes="(min-width: 720px) 720px"></figure><h3 id="getting-bitdeer-ai-cloud-api-key"><strong>Getting Bitdeer AI Cloud API Key</strong></h3><p>To obtain your API key from Bitdeer AI Cloud:</p><ol><li>Log in to your <a href="https://www.bitdeer.ai/en/instance/server?ref=bitdeer.ai" rel="noreferrer">Bitdeer AI Cloud</a> account</li><li>Navigate to the <a href="https://www.bitdeer.ai/en/model/apikeys?ref=bitdeer.ai" rel="noreferrer">API Keys</a> section in Models</li><li>Click &quot;Generate API Key&quot;</li><li>Copy the generated API key and save it securely</li></ol><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-fe2377cc-9ceb-4935-aff2-3c8baefb056d.png" class="kg-image" alt="Installing and Configuring OpenClaw on Bitdeer AI Cloud" loading="lazy" width="1600" height="724" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/02/data-src-image-fe2377cc-9ceb-4935-aff2-3c8baefb056d.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/02/data-src-image-fe2377cc-9ceb-4935-aff2-3c8baefb056d.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-fe2377cc-9ceb-4935-aff2-3c8baefb056d.png 1600w" sizes="(min-width: 720px) 720px"></figure><p><strong>&#x26A0;&#xFE0F; Important:</strong> Keep your API keys secure and never commit them to version control.</p><h3 id="verify-model-configuration"><strong>Verify Model Configuration</strong></h3><p>After configuring your models, verify that they are properly set up by listing available models:</p><pre><code class="language-bash">openclaw models list
</code></pre>
<figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-da9865e9-a94f-4abe-bcda-358a1092004d.png" class="kg-image" alt="Installing and Configuring OpenClaw on Bitdeer AI Cloud" loading="lazy" width="1600" height="1291" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/02/data-src-image-da9865e9-a94f-4abe-bcda-358a1092004d.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/02/data-src-image-da9865e9-a94f-4abe-bcda-358a1092004d.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-da9865e9-a94f-4abe-bcda-358a1092004d.png 1600w" sizes="(min-width: 720px) 720px"></figure><p>This command will display configured models and their availability.</p><h3 id="onboarding-process"><strong>Onboarding Process</strong></h3><p>Complete the onboarding process to finalize your OpenClaw service:</p><pre><code class="language-bash">openclaw onboard --flow quickstart
</code></pre>
<figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-596d7eda-4c65-447f-97b5-ae7547826eec.png" class="kg-image" alt="Installing and Configuring OpenClaw on Bitdeer AI Cloud" loading="lazy" width="1600" height="959" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/02/data-src-image-596d7eda-4c65-447f-97b5-ae7547826eec.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/02/data-src-image-596d7eda-4c65-447f-97b5-ae7547826eec.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-596d7eda-4c65-447f-97b5-ae7547826eec.png 1600w" sizes="(min-width: 720px) 720px"></figure><h3 id="setting-up-telegram-channel"><strong>Setting Up Telegram Channel</strong></h3><p>During the onboarding process, select Telegram as your communication channel and provide your bot token.</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-95564d25-f365-45ff-a6f6-85eebcf440d5.png" class="kg-image" alt="Installing and Configuring OpenClaw on Bitdeer AI Cloud" loading="lazy" width="1600" height="959" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/02/data-src-image-95564d25-f365-45ff-a6f6-85eebcf440d5.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/02/data-src-image-95564d25-f365-45ff-a6f6-85eebcf440d5.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-95564d25-f365-45ff-a6f6-85eebcf440d5.png 1600w" sizes="(min-width: 720px) 720px"></figure><p>After completing the onboarding process, the OpenClaw service will automatically start and run in the background.</p><h3 id="accessing-openclaw-control-panel-optional"><strong>Accessing OpenClaw Control Panel (Optional)</strong></h3><p>If you want to access the OpenClaw control panel via web interface, you can set up an SSH tunnel and access it using the gateway token.</p><p>First, find your gateway token in the configuration file:</p><pre><code class="language-bash">cat ~/.openclaw/openclaw.json
</code></pre>
<p>Look for the gateway configuration section:</p><pre><code class="language-json">{
  &quot;gateway&quot;: {
    &quot;mode&quot;: &quot;local&quot;,
    &quot;auth&quot;: {
      &quot;mode&quot;: &quot;token&quot;,
      &quot;token&quot;: &quot;your-gateway-token&quot;
    },
    &quot;port&quot;: 18789
  }
}
</code></pre>
<p>Create an SSH tunnel from your local machine to access the OpenClaw control panel:</p><pre><code class="language-bash"># Run this on your local machine
ssh -L 18789:localhost:18789 username@your-instance-ip
</code></pre>
<p>Then open your browser and navigate to:</p><pre><code>http://localhost:18789?token=your-gateway-token
</code></pre>
<p>Use the gateway token from your configuration file to authenticate.</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-5b2405c0-14c3-46b7-8aca-09fd4e86b3ce.png" class="kg-image" alt="Installing and Configuring OpenClaw on Bitdeer AI Cloud" loading="lazy" width="1600" height="959" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/02/data-src-image-5b2405c0-14c3-46b7-8aca-09fd4e86b3ce.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/02/data-src-image-5b2405c0-14c3-46b7-8aca-09fd4e86b3ce.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-5b2405c0-14c3-46b7-8aca-09fd4e86b3ce.png 1600w" sizes="(min-width: 720px) 720px"></figure><h2 id="step-4-pairing-with-telegram"><strong>Step 4: Pairing with Telegram</strong></h2><h3 id="connect-your-telegram-chat"><strong>Connect Your Telegram Chat</strong></h3><p>After the OpenClaw service is running with Telegram integration enabled, you need to pair your Telegram chat:</p><ol><li>Open Telegram and find your bot</li><li>Send /start command to initiate the conversation</li><li>The bot will respond with a pairing code</li><li>Enter the pairing code in your OpenClaw terminal:</li></ol><pre><code class="language-bash">openclaw pairing approve telegram &lt;pairing-code&gt;
</code></pre>
<figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-b29f98d9-6de1-4251-8e05-e757e52370b1.png" class="kg-image" alt="Installing and Configuring OpenClaw on Bitdeer AI Cloud" loading="lazy" width="1600" height="953" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/02/data-src-image-b29f98d9-6de1-4251-8e05-e757e52370b1.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/02/data-src-image-b29f98d9-6de1-4251-8e05-e757e52370b1.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-b29f98d9-6de1-4251-8e05-e757e52370b1.png 1600w" sizes="(min-width: 720px) 720px"></figure><h3 id="interactive-chat"><strong>Interactive Chat</strong></h3><p>Once paired, you can start chatting with your AI assistant:</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-eef248eb-8b14-469c-bb5f-2dd4a9bac7b8.png" class="kg-image" alt="Installing and Configuring OpenClaw on Bitdeer AI Cloud" loading="lazy" width="1180" height="1482" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/02/data-src-image-eef248eb-8b14-469c-bb5f-2dd4a9bac7b8.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/02/data-src-image-eef248eb-8b14-469c-bb5f-2dd4a9bac7b8.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-eef248eb-8b14-469c-bb5f-2dd4a9bac7b8.png 1180w" sizes="(min-width: 720px) 720px"></figure><p>The bot can:</p><ul><li>Answer questions using the configured AI model</li><li>Perform web searches</li><li>Execute code snippets (with proper sandboxing)</li><li>Handle file uploads and processing</li><li>Maintain conversation context</li></ul><h3 id="using-telegram-commands"><strong>Using Telegram Commands</strong></h3><p>OpenClaw provides several built-in commands:</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-2b5d8f2f-1d0d-459e-a514-39342264a078.png" class="kg-image" alt="Installing and Configuring OpenClaw on Bitdeer AI Cloud" loading="lazy" width="954" height="836" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/02/data-src-image-2b5d8f2f-1d0d-459e-a514-39342264a078.png 600w, https://www.bitdeer.ai/en/blog/content/images/2026/02/data-src-image-2b5d8f2f-1d0d-459e-a514-39342264a078.png 954w" sizes="(min-width: 720px) 720px"></figure><p>Common commands include:</p><ul><li>/start - Initialize the bot</li><li>/help - Show available commands</li><li>/status - Check bot status</li></ul><h2 id="conclusion"><strong>Conclusion</strong></h2><p>You now have a fully functional OpenClaw AI assistant running on <a href="https://www.bitdeer.ai/en/instance/server?ref=bitdeer.ai" rel="noreferrer">Bitdeer AI Cloud</a> with Telegram integration. This setup provides you with a powerful, customizable AI agent that can assist with various tasks while maintaining security and performance.</p><p>The combination of Bitdeer AI Cloud&apos;s reliable infrastructure and OpenClaw&apos;s flexible framework creates an ideal environment for building and deploying intelligent assistants. As you become more familiar with the platform, you can explore advanced features like:</p><ul><li>Advanced agent skills</li><li>Integration with other channels</li></ul><p>Happy building!</p><p></p><p>*<strong>Security Note:</strong>  Apply appropriate access controls for sensitive actions.</p><h2 id="additional-resources"><strong>Additional Resources</strong></h2><ul><li><a href="https://bitdeer.ai/en/docs/center?ref=bitdeer.ai"><u>Bitdeer AI Cloud Documentation</u></a></li><li><a href="https://github.com/openclaw/openclaw?ref=bitdeer.ai"><u>OpenClaw GitHub Repository</u></a></li></ul>]]></content:encoded></item><item><title><![CDATA[Hack to Hire Episode 2: Turning Real-World Business Friction into AI-Driven Workflows]]></title><description><![CDATA[What happens when AI meets real business pressure? Hack to Hire episode 2 explores real-world systems, diving into prototype details and their implications for enterprise workflows.]]></description><link>https://www.bitdeer.ai/en/blog/hack-to-hire-episode-2-turning-real-world-business-friction-into-ai-driven-workflows/</link><guid isPermaLink="false">697c731b68d90b0001fa6f37</guid><category><![CDATA[AI Applications]]></category><dc:creator><![CDATA[Taylor Ye]]></dc:creator><pubDate>Fri, 30 Jan 2026 09:30:28 GMT</pubDate><media:content url="https://www.bitdeer.ai/en/blog/content/images/2026/01/blog-hack-to-hire-episode-2--2-.png" medium="image"/><content:encoded><![CDATA[<img src="https://www.bitdeer.ai/en/blog/content/images/2026/01/blog-hack-to-hire-episode-2--2-.png" alt="Hack to Hire Episode 2: Turning Real-World Business Friction into AI-Driven Workflows"><p>Hack to Hire 2 has concluded earlier, but the use cases developed during the event continue to stand out. Rather than focusing on experimental models or theoretical benchmarks, the hackathon produced concrete system designs that address persistent business challenges across supply chains, retail operations, and document-heavy industries.</p><p>Built within a limited timeframe, these projects demonstrated how AI when paired with the right infrastructure and system architecture can be embedded directly into operational workflows. Even after the event, the use cases remain relevant as practical references for how enterprises can translate AI capabilities into real business value. In the following article, we will take a closer look at these prototype cases and further explore their practical applications and value.</p><h2 id="use-case-1-market-supplier-intelligence-for-sustainable-supply-chains"><strong>Use Case 1: Market &amp; Supplier Intelligence for Sustainable Supply Chains</strong></h2><h3 id="background"><strong>Background</strong></h3><p>In the textile manufacturing supply chain, sustainability and supplier quality are no longer secondary concerns. Decision-makers must continuously assess compliance, performance, and risk across a large and evolving supplier base. However, relevant data is often distributed across internal records, supplier submissions, external documents, and public news sources.</p><h3 id="technical-business-challenges"><strong>Technical &amp; Business Challenges</strong></h3><p>The primary challenge lies in fragmentation. Supplier information exists in multiple formats and systems, making it difficult to establish a unified, up-to-date view. As a result, sustainability compliance checks and quality assessments are largely manual, delaying critical decisions and increasing operational risk.</p><h3 id="solution-architecture"><strong>Solution Architecture</strong></h3><p>The proposed solution consolidates supplier data, documents, and external information through a centralized data pipeline. Within the system architecture, Bitdeer AI Cloud provides the virtual machine environment and data processing layer used to run ETL workflows, host the analytical dashboard, and support AI-driven insight generation.</p><p>Structured data is stored centrally, while AI models analyze supplier attributes to produce quality scores and contextual recommendations. The dashboard presents these insights in a unified view, enabling more informed and timely supplier management decisions</p><h3 id="outcome"><strong>Outcome</strong></h3><p>The prototype illustrates how AI-supported analytics can transform supplier evaluation from manual reviews into a more systematic and insight-driven process, particularly valuable for sustainability-focused supply chains.</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/01/data-src-image-bdf01072-ef7f-400d-aae9-9d03cdfb8989.png" class="kg-image" alt="Hack to Hire Episode 2: Turning Real-World Business Friction into AI-Driven Workflows" loading="lazy" width="1594" height="710" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/01/data-src-image-bdf01072-ef7f-400d-aae9-9d03cdfb8989.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/01/data-src-image-bdf01072-ef7f-400d-aae9-9d03cdfb8989.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2026/01/data-src-image-bdf01072-ef7f-400d-aae9-9d03cdfb8989.png 1594w" sizes="(min-width: 720px) 720px"></figure><h2 id="use-case-2-retail-store-intelligence-for-planogram-execution"><strong>Use Case 2: Retail Store Intelligence for Planogram Execution</strong></h2><h3 id="background-1"><strong>Background</strong></h3><p>Retail ground staff spend significant time executing product placement according to predefined planograms. Store managers must oversee large teams, while brand and product managers often lack direct visibility into how products are ultimately displayed on the shop floor.</p><h3 id="challenges"><strong>Challenges</strong></h3><p>Manual reporting and visual checks are time-consuming and inconsistent. Without a structured feedback mechanism, it is difficult to assess execution quality or scale oversight across multiple stores.</p><h3 id="solution-architecture-1"><strong>Solution Architecture</strong></h3><p>The system digitizes the planogram workflow from instruction to verification. Product placement requirements are recorded and distributed to ground staff via a centralized interface. After execution, staff upload photos through a mobile web browser.</p><p>Within the architecture, Bitdeer AI Cloud hosts the operational dashboard and supports model inference used to analyze uploaded images. The results are automatically recorded and aggregated, allowing managers to review execution status and performance without manual follow-up.</p><h3 id="outcome-1"><strong>Outcome</strong></h3><p>By combining visual evidence with automated reporting, the solution improves transparency across retail operations and reduces the manual burden on both frontline staff and management.</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/01/data-src-image-3421defc-a99b-4b6e-8112-635ef60b316c.png" class="kg-image" alt="Hack to Hire Episode 2: Turning Real-World Business Friction into AI-Driven Workflows" loading="lazy" width="1598" height="540" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/01/data-src-image-3421defc-a99b-4b6e-8112-635ef60b316c.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/01/data-src-image-3421defc-a99b-4b6e-8112-635ef60b316c.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2026/01/data-src-image-3421defc-a99b-4b6e-8112-635ef60b316c.png 1598w" sizes="(min-width: 720px) 720px"></figure><h2 id="use-case-3-automated-tender-document-intelligence-for-smes"><strong>Use Case 3: Automated Tender Document Intelligence for SMEs</strong></h2><h3 id="background-2"><strong>Background</strong></h3><p>Singapore&#x2019;s construction industry issues thousands of government tenders each year, many consisting of hundreds of pages of documentation. For SMEs, participating in these tenders is critical but resource-intensive.</p><h3 id="challenges-1"><strong>Challenges</strong></h3><p>Tender documents are typically received via email and reviewed manually. Processing a single tender can take several days, placing heavy strain on small quality assurance teams and increasing the risk of missed deadlines or errors.</p><h3 id="solution-architecture-2"><strong>Solution Architecture</strong></h3><p>The proposed pipeline automates tender intake and evaluation. Incoming emails trigger document processing workflows that extract, analyze, and assess tender requirements. AI and Retrieval-Augmented Generation (RAG) techniques reference internal document repositories such as certificates and project histories to perform eligibility checks.</p><p>Bitdeer AI Cloud provides the execution environment for these automated workflows, hosting the virtual machines that run document processing, AI inference, and the dashboard used to visualize results and audit pipeline actions</p><h3 id="outcome-2"><strong>Outcome</strong></h3><p>The prototype demonstrates how AI-enabled automation can significantly shorten tender review cycles, helping SMEs respond more efficiently and compete more effectively in high-volume tender environments.</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/01/data-src-image-afd50bd5-88dc-4e14-b092-31de06281e45.png" class="kg-image" alt="Hack to Hire Episode 2: Turning Real-World Business Friction into AI-Driven Workflows" loading="lazy" width="1596" height="540" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/01/data-src-image-afd50bd5-88dc-4e14-b092-31de06281e45.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/01/data-src-image-afd50bd5-88dc-4e14-b092-31de06281e45.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2026/01/data-src-image-afd50bd5-88dc-4e14-b092-31de06281e45.png 1596w" sizes="(min-width: 720px) 720px"></figure><h2 id="infrastructure-as-an-enabler-not-the-focus"><strong>Infrastructure as an Enabler, Not the Focus</strong></h2><p>Across all three use cases, Bitdeer AI Cloud functions as the underlying execution layer that supports data pipelines, AI inference, and dashboard deployment. By providing a consistent and flexible runtime environment, teams were able to focus on workflow design and problem-solving rather than infrastructure setup.</p><p>This separation between infrastructure and application logic proved critical in enabling rapid iteration within the hackathon timeframe.</p><h2 id="lasting-takeaways">Lasting Takeaways</h2><p>The solutions developed through Hack to Hire demonstrate more than short term experimentation. They provide concrete examples of how AI can be operationalized when designed around real business workflows rather than isolated models or theoretical prototypes.</p><p>Across all use cases, a consistent pattern emerges. Real value is created when AI is embedded into end to end processes and supported by infrastructure that enables rapid iteration and reliable execution. In this context, Bitdeer AI is positioned as a vertically integrated AI cloud platform that unifies AI data center infrastructure with a comprehensive AI cloud layer, providing built in tools that support the full AI lifecycle from development and training to deployment and production operations. Together, these projects illustrate a pragmatic path for enterprises adopting AI by focusing on workflow integration, operational impact, and measurable outcomes.</p>]]></content:encoded></item><item><title><![CDATA[Unlocking AI Potential with NVIDIA GB200 NVL72]]></title><description><![CDATA[An in-depth look at NVIDIA GB200 NVL72, its system architecture, technical advances over H100, and where it fits in large-scale AI training and inference.]]></description><link>https://www.bitdeer.ai/en/blog/unlocking-ai-potential-with-nvidia-gb200-nvl72/</link><guid isPermaLink="false">697331ae68d90b0001fa6ef2</guid><category><![CDATA[AI Applications]]></category><dc:creator><![CDATA[Taylor Ye]]></dc:creator><pubDate>Fri, 23 Jan 2026 09:21:22 GMT</pubDate><media:content url="https://www.bitdeer.ai/en/blog/content/images/2026/01/blog-why-nvidia-gb200-nvl72-changes-the-game-02-1.png" medium="image"/><content:encoded><![CDATA[<img src="https://www.bitdeer.ai/en/blog/content/images/2026/01/blog-why-nvidia-gb200-nvl72-changes-the-game-02-1.png" alt="Unlocking AI Potential with NVIDIA GB200 NVL72"><p>The rapid advancements in artificial intelligence require cutting-edge hardware capable of handling the immense computational demands of machine learning , deep learning , and AI-driven workloads. One of the most anticipated breakthroughs in this domain is the NVIDIA GB200 NVL72, a GPU designed specifically to meet the needs of high-performance computing and AI workloads. Leveraging the power of NVIDIA&apos;s innovative architectures, the NVIDIA GB200 NVL72 is poised to revolutionize AI development.</p><h2 id="what-is-nvidia-gb200-nvl72"><strong>What is NVIDIA GB200 NVL72?&#xA0;</strong></h2><p>The NVIDIA GB200 NVL72 is a rack-scale AI system designed for large-scale AI training and inference workloads. Built on NVIDIA&#x2019;s Blackwell architecture, it features 36 Grace Blackwell Superchips, each integrating one NVIDIA Grace CPU and two Blackwell GPUs, totaling 72 GPUs and 36 CPUs in a single liquid-cooled rack.</p><p>Through a full NVLink and NVLink Switch interconnect, all GPUs operate as a tightly coupled system, allowing the platform to function as a single, massive GPU rather than a collection of discrete accelerators. This architecture significantly reduces inter-GPU communication overhead and is optimized for workloads that involve large model sizes, frequent parameter exchange, and long-running training or inference jobs that require stable, predictable performance at scale.</p><h2 id="technical-evolution-from-h100-to-gb200-nvl72"><strong>Technical evolution from H100 to GB200 NVL72</strong></h2><p>While H100 marked a major step forward for transformer-based AI, GB200 NVL72 builds on that foundation with system-level changes. Compared with H100, GB200 NVL72 introduces next-generation Tensor Cores and an updated second-generation Transformer Engine optimized for new FP4 precision, which enables doubling the compute capacity and model size compared to the H100&#x2019;s FP8, specifically benefiting massive LLM inference. This directly benefits LLM training and inference, where matrix operations dominate runtime.</p><p>Interconnect is another key differentiator. H100 clusters rely on NVLink within nodes and high-speed networking across nodes, which can introduce latency and synchronization overhead at scale. GB200 NVL72 extends the NVLink domain to all 72 GPUs within the rack via the NVLink Switch System, offering 9x the aggregate bandwidth of previous generations and treating the entire rack as a single GPU.</p><p>Memory access patterns also improve at the system level. By enabling faster parameter sharing and collective operations, GB200 NVL72 improves utilization efficiency, especially for models that exceed the memory capacity of a single GPU. Finally, GB200 NVL72 is a liquid-cooled-native system. This shift is essential to handle the 120kW+ rack power density, ensuring sustained peak performance that would be physically impossible to achieve with traditional air-cooling.</p><p>These architectural changes are not only reflected in design principles, but also translate directly into measurable gains in real-world AI workloads. The following benchmarks illustrate how GB200 NVL72 compares with H100-based systems across large language model training, inference throughput, energy efficiency, and data processing performance.</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/01/image.png" class="kg-image" alt="Unlocking AI Potential with NVIDIA GB200 NVL72" loading="lazy" width="2000" height="802" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/01/image.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/01/image.png 1000w, https://www.bitdeer.ai/en/blog/content/images/size/w1600/2026/01/image.png 1600w, https://www.bitdeer.ai/en/blog/content/images/size/w2400/2026/01/image.png 2400w" sizes="(min-width: 720px) 720px"></figure><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2026/01/image-1.png" class="kg-image" alt="Unlocking AI Potential with NVIDIA GB200 NVL72" loading="lazy" width="2000" height="814" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2026/01/image-1.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2026/01/image-1.png 1000w, https://www.bitdeer.ai/en/blog/content/images/size/w1600/2026/01/image-1.png 1600w, https://www.bitdeer.ai/en/blog/content/images/size/w2400/2026/01/image-1.png 2400w" sizes="(min-width: 720px) 720px"></figure><p><em>source: from NVIDIA website</em></p><h2 id="business-use-cases-where-gb200-nvl72-fits-best">Business Use Cases: Where GB200 NVL72 Fits Best</h2><p>The NVIDIA GB200 NVL72 is designed for AI workloads that require high computational throughput, efficient scaling, and sustained performance. Its rack-scale architecture supports enterprise scenarios where training speed, inference efficiency, and operational cost have a direct impact on business outcomes across multiple industries, including healthcare, finance, and AI-native technology companies.</p><h3 id="large-scale-ai-model-training"><strong>Large-Scale AI Model Training</strong></h3><p>NVIDIA GB200 NVL72 is well suited for training large and complex AI models used in areas such as autonomous driving, healthcare, robotics, and natural language processing. With powerful Tensor Cores and increased system-level throughput, the platform reduces the time required to train deep learning models, enabling faster experimentation cycles and quicker progression from research to deployment.</p><h3 id="high-performance-ai-inference"><strong>High-Performance AI Inference</strong></h3><p>Beyond training, NVIDIA GB200 NVL72 supports inference workloads that demand fast and reliable response times. Applications such as video analytics, recommendation systems, and real-time decision support benefit from the platform&#x2019;s ability to execute inference workloads efficiently at scale, helping enterprises deliver more responsive AI-powered services. And in&#xA0; healthcare and life sciences, workloads like medical imaging models, genomics analysis, and drug discovery rely on large datasets and long-running training jobs. It can help reduce training time and improve resource utilization, enabling faster experimentation and more efficient progression from research to clinical or production environments.</p><h3 id="energy-efficient-ai-at-scale"><strong>Energy-Efficient AI at Scale</strong></h3><p>NVIDIA GB200 NVL72 is designed to balance performance with efficiency, making it suitable for large-scale AI deployments where power consumption and operating costs are key considerations. This is particularly relevant for cloud providers, enterprise AI platforms, and research institutions running continuous workloads, where efficiency at scale directly affects total cost of ownership.</p><h3 id="enabling-next-generation-ai-applications"><strong>Enabling Next-Generation AI Applications</strong></h3><p>By supporting massive datasets and increasingly complex models, NVIDIA GB200 NVL72 enables organizations to develop and deploy next-generation AI applications across industries. This includes multimodal AI systems, large shared foundation models, and AI platforms that serve multiple business units or customers. The platform allows enterprises to scale AI capabilities incrementally while maintaining performance and architectural consistency as model complexity grows.</p><h2 id="bitdeer-ai-and-nvidia-gb200-nvl72-clusters">Bitdeer AI and NVIDIA GB200 NVL72 Clusters</h2><p>As part of our ongoing expansion of the NVIDIA GPU offerings on Bitdeer AI Cloud platform, the addition of NVIDIA GB200 NVL72 extends the platform&#x2019;s ability to support larger and more communication-intensive AI workloads. At the same time, NVIDIA H100 remains an integral part of our platform, continuing to support a wide spectrum of training and inference use cases. Together, these GPU models form a complementary compute foundation that addresses different model sizes and deployment needs.</p><p>By offering both platforms, Bitdeer AI allows customers to select GPU configurations that best match their model size, workload characteristics, and operational requirements.</p>]]></content:encoded></item><item><title><![CDATA[Traditional Data Centers vs AI Data Centers: How Infrastructure Is Evolving to Support AI at Scale]]></title><description><![CDATA[Explore how AI data centers differ from traditional data centers in design, workloads, and business value as AI drives new infrastructure demands.]]></description><link>https://www.bitdeer.ai/en/blog/traditional-data-centers-vs-ai-data-centers-how-infrastructure-is-evolving-to-support-ai-at-scale/</link><guid isPermaLink="false">6960adce68d90b0001fa6eb9</guid><dc:creator><![CDATA[Taylor Ye]]></dc:creator><pubDate>Fri, 09 Jan 2026 07:40:21 GMT</pubDate><media:content url="https://www.bitdeer.ai/en/blog/content/images/2026/01/data-centers-en@2x.png" medium="image"/><content:encoded><![CDATA[<img src="https://www.bitdeer.ai/en/blog/content/images/2026/01/data-centers-en@2x.png" alt="Traditional Data Centers vs AI Data Centers: How Infrastructure Is Evolving to Support AI at Scale"><p>Data centers have always reflected the dominant computing paradigm of their time. For many years, enterprise software, web services, and databases shaped how infrastructure was designed and operated. These workloads emphasized reliability, steady performance, and efficient resource sharing.</p><p>Artificial intelligence introduces a fundamentally different demand profile. Training and deploying modern AI systems requires large-scale parallel computation, rapid data movement, and significantly higher power density. As a result, a new class of infrastructure has emerged alongside traditional facilities: the AI data center. This article explains how traditional data centers and AI data centers differ in design, application, and business value.</p><h2 id="what-is-a-traditional-data-center"><strong>What Is a Traditional Data Center?</strong></h2><p>A traditional data center is built to support a wide variety of general-purpose computing workloads. Its architecture prioritizes flexibility and operational stability across many applications.</p><p>Core characteristics include:</p><ul><li>CPU-focused compute, optimized for sequential and transactional processing</li><li>Moderate power density, typically compatible with air-based cooling</li><li>General-purpose networking, designed primarily for client-to-server traffic</li><li>Multi-tenant workload support, often using virtualization technologies</li></ul><p>This model remains well-suited for enterprise systems, web platforms, and business-critical applications that require predictable performance and high availability.</p><h2 id="what-is-an-ai-data-center"><strong>What Is an AI Data Center?</strong></h2><p>An AI data center is engineered specifically to support machine learning and artificial intelligence workloads. These environments are designed around accelerators rather than general-purpose processors.</p><p>Key attributes include:</p><ul><li>GPU or accelerator-centric clusters designed for parallel computation</li><li>High power and thermal density, requiring advanced cooling approaches</li><li>High-bandwidth, low-latency interconnects to support intensive data exchange</li><li>Tightly integrated compute, storage, and networking to minimize bottlenecks</li></ul><p>Instead of serving many unrelated workloads, AI data centers are optimized to run fewer but extremely compute-intensive jobs efficiently. Unlike the standard Ethernet used in traditional centers, AI hubs often employ specialized InfiniBand or RoCE (RDMA over Converged Ethernet) fabrics to ensure the network doesn&apos;t become a bottleneck for the GPUs.</p><h2 id="data-center-tier-classification-tier-i-tier-ii-and-tier-iii"><strong>Data Center Tier Classification: Tier I, Tier II, and Tier III</strong></h2><p>Before comparing traditional data centers and AI data centers, it is important to understand the commonly used data center tier classification system. Tier levels are widely used to describe a data center&#x2019;s reliability, redundancy, and maintainability, rather than its computing performance.</p><h3 id="tier-i-data-center"><strong>Tier I Data Center</strong></h3><p>A Tier I data center represents the most basic level of data center infrastructure. It typically relies on a single path for power and cooling, with little or no redundancy. Planned maintenance or unexpected failures usually require a full system shutdown. Tier I facilities are best suited for non-critical workloads where availability requirements are minimal.It has an expected uptime of 99.671% (28.8 hours of downtime annually).</p><h3 id="tier-ii-data-center"><strong>Tier II Data Center</strong></h3><p>Tier II data centers introduce limited redundancy in key components such as power supplies and cooling systems. While backup components exist, the infrastructure generally still operates on a single distribution path. Tier II improves reliability compared to Tier I, but maintenance activities may still cause service interruptions. This tier is commonly used for workloads with moderate availability requirements. It has an expected uptime of 99.741% (22 hours of downtime annually).</p><h3 id="tier-iii-data-center"><strong>Tier III Data Center</strong></h3><p>Tier III data centers are designed with concurrently maintainable infrastructure. Critical systems typically follow an N+1 redundancy model and support multiple power and cooling paths, allowing maintenance to be performed without interrupting operations. Tier III is widely adopted for enterprise-grade and cloud data centers and is suitable for mission-critical applications, including many AI workloads.It has an expected uptime of 99.982% (1.6 hours of downtime annually).</p><p>It is important to note that tier classification reflects infrastructure reliability rather than workload type or computing capability. AI data centers are often built on Tier III or higher standards, with additional design considerations for high-density power delivery and advanced cooling.</p><h2 id="architectural-and-operational-analysis"><strong>Architectural and Operational Analysis</strong></h2><p>The architectural gap between traditional and AI data centers has widened rapidly. CPUs are designed for versatility, but AI workloads demand parallel execution across thousands of identical operations. GPUs and AI accelerators fill this gap, but they also introduce new constraints.</p><p>AI data centers must address:</p><ul><li>Power delivery challenges, as high-density racks draw dramatically more energy</li><li>Thermal management, where air cooling alone is insufficient</li><li>Network congestion, since model training depends on constant GPU-to-GPU communication</li><li>Data locality, ensuring storage and compute remain tightly coupled</li></ul><p>Operationally, this shifts the focus from managing virtual machines to managing clusters as a single system. Operational success is no longer measured only in utilization rates, but in training time, inference latency, and performance per watt.</p><p><strong>Below is a key comparison between Traditional Data Center and AI Data center:</strong></p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/zh/blog/content/images/2026/01/image.png" class="kg-image" alt="Traditional Data Centers vs AI Data Centers: How Infrastructure Is Evolving to Support AI at Scale" loading="lazy" width="1272" height="638"></figure><h2 id="business-use-cases-and-industry-impact"><strong>Business Use Cases and Industry Impact</strong></h2><p>AI data centers are rapidly becoming a core source of competitive advantage for modern businesses. They enable organizations to move beyond isolated AI experiments toward large-scale model training, high-throughput inference, and AI agents that automate complex workflows such as customer support, software development, and data analysis. The ability to scale compute, data movement, and power density directly influences speed of innovation and time-to-market.</p><p>Across industries, AI data centers are reshaping how value is created. Financial institutions use them for real-time risk modeling and fraud detection, healthcare organizations for advanced medical imaging and diagnostics, and manufacturers for simulation, digital twins, and predictive maintenance. These AI-driven workloads are increasingly mission-critical rather than experimental.</p><p>As AI adoption accelerates, businesses are aligning infrastructure strategy around AI data centers as long-term strategic assets. Rather than serving as supporting systems, they are becoming the primary engines for innovation, operational efficiency, and competitive positioning, defining how quickly and effectively organizations can scale intelligence across their operations.</p><h2 id="conclusion"><strong>Conclusion</strong></h2><p>As AI systems grow in scale and complexity, infrastructure decisions are no longer just technical choices but strategic ones. The ability to combine cloud flexibility with purpose-built AI data center design is becoming essential for organizations that want to deploy advanced AI reliably and efficiently.</p><p>At Bitdeer AI, we build our cloud platform around this reality. By leveraging our group&#x2019;s global infrastructure resources, including power capacity and data center sites, and combining them with flexible cloud capabilities, we provide an AI cloud foundation designed specifically for advanced workloads on advanced NVIDIA GPUs.This integrated approach enables scalable AI deployment while maintaining performance, energy efficiency, and operational resilience. We allow our customers to focus on building and deploying intelligence, while we manage the underlying AI infrastructure securely, sustainably, and at scale.</p>]]></content:encoded></item><item><title><![CDATA[Deploying Scalable PostgreSQL Database Service on Bitdeer AI Cloud]]></title><description><![CDATA[
Simplify database management with AI Cloud RDS. Optimized for PostgreSQL, get reliable, scalable, and secure high-performance database solutions.]]></description><link>https://www.bitdeer.ai/en/blog/deploying-scalable-postgresql-database-service-on-bitdeer-ai-cloud/</link><guid isPermaLink="false">693b5b3668d90b0001fa6e93</guid><category><![CDATA[AI Applications]]></category><dc:creator><![CDATA[Taylor Ye]]></dc:creator><pubDate>Fri, 12 Dec 2025 08:46:27 GMT</pubDate><media:content url="https://www.bitdeer.ai/en/blog/content/images/2025/12/Bitdeer-AI-RDS-blog-EN.png" medium="image"/><content:encoded><![CDATA[<img src="https://www.bitdeer.ai/en/blog/content/images/2025/12/Bitdeer-AI-RDS-blog-EN.png" alt="Deploying Scalable PostgreSQL Database Service on Bitdeer AI Cloud"><p>In the era of data-driven applications and AI-powered innovations, businesses require reliable, scalable, and high-performance database solutions. To meet this growing demand, we are excited to showcase RDS (Relational Database Service) on our AI cloud platform. Designed to simplify database management while ensuring security and scalability, our RDS solution is optimized for PostgreSQL, providing a seamless experience for developers and enterprises alike.</p><h3 id="key-features-of-our-rds-service"><strong>Key Features of Our RDS Service</strong></h3><ol><li><strong>Fully Managed PostgreSQL: </strong>Our RDS supports PostgreSQL, a reliable and highly extensible open-source database. With broad community support and a rich ecosystem of extensions for analytics, geospatial data, full-text search, time-series workloads, and more, it adapts easily to a wide range of use cases.&#xA0;</li><li><strong>Database Monitoring</strong>: Monitor your databases with Prometheus, Grafana, and Loki through our RDS, enabling real-time metrics, clear visualizations, and efficient log analysis to maintain performance and database health.&#xA0;</li></ol><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2025/12/data-src-image-7290fa61-03d9-45e1-b1dc-4e2d6f5616ca.png" class="kg-image" alt="Deploying Scalable PostgreSQL Database Service on Bitdeer AI Cloud" loading="lazy" width="1600" height="884" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2025/12/data-src-image-7290fa61-03d9-45e1-b1dc-4e2d6f5616ca.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2025/12/data-src-image-7290fa61-03d9-45e1-b1dc-4e2d6f5616ca.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2025/12/data-src-image-7290fa61-03d9-45e1-b1dc-4e2d6f5616ca.png 1600w" sizes="(min-width: 720px) 720px"></figure><ol start="3"><li><strong>Security &amp; Compliance</strong>:<ul><li><strong>VPC (Virtual Private Cloud) Isolation</strong>: Provides network-level security by isolating database instances from unauthorized access.</li><li><strong>Encryption</strong>: Protects sensitive data and prevents unauthorized access, ensuring compliance with security best practices.</li></ul></li><li><strong>Maintenance-Free Server Management</strong>: Offloads the burden of hardware maintenance and OS management, allowing users to focus on application development without worrying about infrastructure upkeep.</li></ol><h3 id="advantages-of-using-our-rds-service"><strong>Advantages of Using Our RDS Service</strong></h3><p>Building on these capabilities, our RDS provides several clear advantages:</p><ul><li><strong>Simplified Database Management</strong>: Automates routine database tasks such as monitoring, reducing operational overhead.</li><li><strong>Enhanced Security</strong>: Data Encryption ensures that your data remains protected from external threats.</li><li><strong>Optimized for Performance</strong>: With PostgreSQL&#x2019;s advanced indexing, querying, and transaction capabilities, RDS provides a powerful solution for data-heavy applications.</li><li><strong>Automatic Failover</strong>: With clustered deployment mode, auto failover feature is enabled to prevent application disruption over primary node failure.</li></ul><h3 id="use-cases"><strong>Use Cases</strong></h3><p>These features enable PostgreSQL on Bitdeer AI Cloud to support a wide spectrum of real-world scenarios:</p><ul><li><strong>Backend Databases for Enterprise Applications</strong> : Support mission-critical systems such as ERP, CRM, and large-scale SaaS platforms with reliable, high-performance PostgreSQL infrastructure that ensures consistency, availability, and smooth transactional operations.</li><li><strong>Vector Database for Native RAG Workflows</strong> : Store and query high-dimensional embeddings using PostgreSQL vector extensions to power Retrieval-Augmented Generation, semantic search, recommendation engines, and other embedding-driven AI applications.</li><li><strong>Graph Database for Graph RAG and Relationship Intelligence : </strong>Leverage PostgreSQL graph and JSONB capabilities to model complex relationships, enabling graph queries, entity linking, fraud detection, and Graph RAG scenarios that require deep relational reasoning.</li><li><strong>OLAP Compatibility for Seamless Data Lake Integration</strong> : Support analytical workloads through PostgreSQL&#x2019;s OLAP-friendly features, columnar extensions, and connectors to data lake systems. Enable unified storage, fast aggregations, and ETL workflows without moving data across multiple systems.</li><li><strong>Full-Text Search for Log Storage and Querying</strong> : Use PostgreSQL&apos;s native FTS capabilities to index and query log data at scale, enabling fast retrieval, filtering, and pattern detection for observability and operational analytics.</li><li><strong>Time-Series Database for Dashboards and Visualization</strong> : Store high-volume time-series data such as metrics, events, and telemetry using PostgreSQL time-series extensions. Power real-time dashboards, monitoring systems, IoT analytics, and operational insights with efficient ingestion and querying.</li><li><strong>Web Applications &amp; SaaS</strong> :Provide scalable and dependable database infrastructure for CMS platforms, e-commerce systems, mobile applications, and multi-tenant SaaS products requiring predictable performance and strong data integrity.</li><li><strong>AI &amp; Machine Learning Workflows</strong> : Serve as a unified data layer for AI pipelines by storing structured datasets, powering real-time analytics, and integrating with AI frameworks for tasks such as inference, recommendation, and predictive modeling.</li></ul><h3 id="getting-started"><strong>Getting Started</strong></h3><p>Deploying a PostgreSQL instance on our RDS platform is straightforward. You can select your desired configurations, set up monitoring with Prometheus, and benefit from security features within a few clicks.&#xA0;</p><p>Step1: Enter the RDS dashboard</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2025/12/data-src-image-2cee703d-8ce9-4b3d-9286-765dae9c8cc2.png" class="kg-image" alt="Deploying Scalable PostgreSQL Database Service on Bitdeer AI Cloud" loading="lazy" width="1600" height="866" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2025/12/data-src-image-2cee703d-8ce9-4b3d-9286-765dae9c8cc2.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2025/12/data-src-image-2cee703d-8ce9-4b3d-9286-765dae9c8cc2.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2025/12/data-src-image-2cee703d-8ce9-4b3d-9286-765dae9c8cc2.png 1600w" sizes="(min-width: 720px) 720px"></figure><p>Step2: Select Instance Type</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2025/12/data-src-image-d0d69ff7-ea74-4c93-986b-dd257e71443f.png" class="kg-image" alt="Deploying Scalable PostgreSQL Database Service on Bitdeer AI Cloud" loading="lazy" width="1600" height="857" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2025/12/data-src-image-d0d69ff7-ea74-4c93-986b-dd257e71443f.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2025/12/data-src-image-d0d69ff7-ea74-4c93-986b-dd257e71443f.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2025/12/data-src-image-d0d69ff7-ea74-4c93-986b-dd257e71443f.png 1600w" sizes="(min-width: 720px) 720px"></figure><p>Step3: Select PostgreSQL Extensions</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2025/12/data-src-image-210d9be8-606e-4477-879a-2326b96cd869.png" class="kg-image" alt="Deploying Scalable PostgreSQL Database Service on Bitdeer AI Cloud" loading="lazy" width="1600" height="854" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2025/12/data-src-image-210d9be8-606e-4477-879a-2326b96cd869.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2025/12/data-src-image-210d9be8-606e-4477-879a-2326b96cd869.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2025/12/data-src-image-210d9be8-606e-4477-879a-2326b96cd869.png 1600w" sizes="(min-width: 720px) 720px"></figure><p>For better understanding, please see the full tutorial video below.</p><figure class="kg-card kg-embed-card"><iframe width="200" height="150" src="https://www.youtube.com/embed/Xrq8kGdsuk4?feature=oembed" frameborder="0" allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share" referrerpolicy="strict-origin-when-cross-origin" allowfullscreen title="Bitdeer AI RDS Setup: The Complete Step-by-Step Guide"></iframe></figure><h3 id="conclusion"><strong>Conclusion</strong></h3><p>Our RDS service is designed to empower businesses by providing a fully managed PostgreSQL solution that combines ease of use, scalability, and enterprise-grade security. Whether you are driving AI-powered innovation, managing enterprise workloads, or running large-scale web services, Bitdeer AI Cloud ensures optimal performance and reliability. By adopting our cloud-based RDS, organizations can streamline operations, enhance data-driven decision-making, and advance their digital transformation.</p>]]></content:encoded></item><item><title><![CDATA[How AI Is Scaling Intelligence Across Finance Industry]]></title><description><![CDATA[AI is transforming finance with smarter decisioning, stronger operations, and faster customer engagement across banking, insurance, and asset management.]]></description><link>https://www.bitdeer.ai/en/blog/how-ai-is-scaling-intelligence-across-finance-industry/</link><guid isPermaLink="false">69203a9368d90b0001fa6e75</guid><category><![CDATA[AI Applications]]></category><dc:creator><![CDATA[Taylor Ye]]></dc:creator><pubDate>Fri, 21 Nov 2025 10:36:06 GMT</pubDate><media:content url="https://www.bitdeer.ai/en/blog/content/images/2025/11/how-ai-is-shaping-the-future-of-finance-en.png" medium="image"/><content:encoded><![CDATA[<img src="https://www.bitdeer.ai/en/blog/content/images/2025/11/how-ai-is-shaping-the-future-of-finance-en.png" alt="How AI Is Scaling Intelligence Across Finance Industry"><p>AI is becoming essential to the financial industry&#x2019;s broader digital transformation goals. As data volumes rise and customer expectations shift toward immediacy, precision, and always-on service, financial institutions are recognizing AI&#x2019;s potential to improve decision-making, strengthen operations, and increase the speed and quality of customer engagement. With tools, models, and deployment frameworks becoming more mature, AI is moving beyond experimentation and into production across multiple business lines.</p><p>Across banking, insurance, and asset management, AI adoption is concentrated around three major domains where it consistently demonstrates measurable value: Customer Intelligence, Risk and Decisioning, and Market and Data Intelligence. Together, these areas represent the foundation of how modern financial institutions generate insights, manage complexity, and deliver differentiated services.</p><h2 id="customer-intelligence-and-experience-transformation"><strong>Customer Intelligence and Experience Transformation</strong></h2><p>Financial institutions are shifting from product-centric engagement to data-driven, personalised customer journeys. AI enables deeper understanding of individual behaviour, intent, and financial circumstances, leading to more relevant and timely interactions.</p><ol><li><strong>Hyper-Personalization and Targeted Offerings</strong></li></ol><p>AI synthesizes behavioral data, spending patterns, life-stage indicators, and risk profiles to tailor credit limits, loan pricing, protection products, and investment guidance. This precision improves conversion and strengthens long-term loyalty.</p><ol start="2"><li><strong>Intelligent Customer Service and Autonomous Support</strong></li></ol><p>AI-powered service agents resolve inquiries, retrieve account information, and complete tasks instantly while supporting multiple languages and channels. Institutions benefit from reduced operational costs and more consistent service quality.</p><ol start="3"><li><strong>AI-Augmented Advisory and Scenario Planning</strong></li></ol><p>Advisors and clients use AI-generated analysis portfolio diagnostics, risk-return scenarios, economic simulations to make more informed decisions. This elevates advisory capabilities and increases planning accuracy across retail and wealth segments.</p><p>Increasingly, financial institutions view customer intelligence as a strategic differentiator, using AI to power personalized journeys at scale.</p><h2 id="risk-compliance-and-core-decisioning"><strong>Risk, Compliance, and Core Decisioning</strong></h2><p>Effective decisioning lies at the heart of financial operations. AI strengthens the precision, speed, and consistency of these processes, enabling institutions to manage uncertainty with greater clarity.</p><ol><li><strong>Risk Management and Fraud Prevention</strong></li></ol><p>Machine learning models detect anomalies across transaction networks, devices, accounts, and behavioural patterns. Compared with rule-based systems, AI adapts quickly to new threats and reduces false positives which is critical for fraud, AML, and risk surveillance.</p><ol start="2"><li><strong>Automated Underwriting and Credit Decisioning</strong></li></ol><p>AI-based scoring models incorporate wider data sets and produce more holistic assessments of creditworthiness while supporting clear rationales. As a result, approval cycles shorten, portfolio quality improves, and new credit products become feasible.</p><ol start="3"><li><strong>Compliance and Reporting Automation</strong></li></ol><p>AI extracts, verifies, and reconciles information from documents, forms, and internal records to accelerate KYC reviews and reporting tasks. Intelligent validation ensures greater completeness and accuracy in financial reporting and disclosure.</p><p>AI in this domain enhances operational resilience and supports high-volume, high-accuracy decision workflows across the enterprise.</p><h2 id="market-intelligence-and-data-automation"><strong>Market Intelligence and Data Automation</strong></h2><p>In a market defined by rapid information cycles and increasingly complex data streams, AI provides institutions with analytical leverage and real-time insight.</p><ol><li><strong>Predictive Analytics and Forecasting</strong></li></ol><p>Models anticipate shifts in liquidity demand, pricing risk, loan performance, customer churn, and macroeconomic indicators. These forecasts support better planning, capital allocation, and risk mitigation.</p><ol start="2"><li><strong>Algorithmic Trading and Market Analytics</strong></li></ol><p>AI identifies patterns across large, fast-moving data sources market data, news sentiment, macro signals and supports traders with improved strategy optimisation and execution.</p><ol start="3"><li><strong>Advanced Data Synthesis, Document Intelligence, and Narrative Reporting</strong></li></ol><p>LLM-driven systems interpret filings, contracts, statements, and regulatory documents, transforming unstructured information into structured, analysable formats. AI also generates narrative explanations of results, enabling faster reporting cycles and supporting emerging &#x201C;continuous close&#x201D; practices where financial insights are updated in near real time.</p><p>These capabilities give institutions a more complete and timely view of their financial and market environment.</p><p>As this overview shows, AI is reshaping nearly every function in modern banking, with impact spanning pricing, onboarding, operations, servicing, and risk. The challenge is no longer identifying use cases, it&#x2019;s building the infrastructure to scale them consistently across the organization. This is where a unified AI platform becomes essential.</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2025/11/data-src-image-11250bc0-f27f-4b2f-9d72-9f8273fc7a8d.png" class="kg-image" alt="How AI Is Scaling Intelligence Across Finance Industry" loading="lazy" width="1052" height="1208" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2025/11/data-src-image-11250bc0-f27f-4b2f-9d72-9f8273fc7a8d.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2025/11/data-src-image-11250bc0-f27f-4b2f-9d72-9f8273fc7a8d.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2025/11/data-src-image-11250bc0-f27f-4b2f-9d72-9f8273fc7a8d.png 1052w" sizes="(min-width: 720px) 720px"></figure><p><em>Figure 1. Banking Example (Source </em><a href="https://reports.weforum.org/docs/WEF_Artificial_Intelligence_in_Financial_Services_2025.pdf?ref=bitdeer.ai" rel="noreferrer"><em>The World Economic Forum</em></a><em> )</em></p><h2 id="how-bitdeer-ai-power-financial-solution"><strong>How Bitdeer AI Power Financial Solution</strong></h2><p>To operationalize AI at scale, financial institutions need a unified environment that integrates high-performance compute, secure data infrastructure, production-ready models, and automated workflows.&#xA0;</p><p>Bitdeer AI provides a unified cloud environment that combines scalable GPU compute with secure data and model-development tools. Through AI Studio covering notebooks, RAG pipelines, and training jobs that teams can build, refine, and deploy open-source LLMs efficiently while relying on integrated databases and vector search for data-driven financial workflows.</p><p>On top of this foundation, the Agent Builder allows institutions to create custom agentic workflows using open-source models, tool integrations, and database queries. These agents automate tasks such as document understanding, KYC data extraction, validation, and summarization, helping financial teams streamline high-volume processes with greater consistency and speed.</p><h2 id="conclusion"><strong>Conclusion</strong></h2><p>AI is reshaping how financial institutions operate, powering more personalised customer experiences, more accurate decisioning, and faster, insight-driven reporting. As AI capabilities mature, the institutions that succeed will be those able to deploy them consistently and securely across their core operations. With an integrated platform designed for high-performance, production-grade AI, Bitdeer AI enables financial organizations to scale these capabilities with confidence and unlock the full value of intelligent transformation.</p>]]></content:encoded></item><item><title><![CDATA[Hack to Hire: Empowering AI Innovation Through Bitdeer AI Cloud]]></title><description><![CDATA[Bitdeer AI & GECO Asia's Hack to Hire hackathon brought together engineers to build real-world AI solutions using scalable cloud computing & pre-trained models.]]></description><link>https://www.bitdeer.ai/en/blog/hack-to-hire-empowering-ai-innovation-through-bitdeer-ai-cloud/</link><guid isPermaLink="false">6909c1b168d90b0001fa6e34</guid><category><![CDATA[AI Applications]]></category><dc:creator><![CDATA[Taylor Ye]]></dc:creator><pubDate>Tue, 04 Nov 2025 09:33:53 GMT</pubDate><media:content url="https://www.bitdeer.ai/en/blog/content/images/2025/11/turning-concepts-into-ai-prototypes-hack-to-hire-2025-EN-1.png" medium="image"/><content:encoded><![CDATA[<img src="https://www.bitdeer.ai/en/blog/content/images/2025/11/turning-concepts-into-ai-prototypes-hack-to-hire-2025-EN-1.png" alt="Hack to Hire: Empowering AI Innovation Through Bitdeer AI Cloud"><p>The Hack to Hire hackathon series, jointly organized by Bitdeer AI and <a href="https://geco.asia/?ref=bitdeer.ai" rel="noreferrer">GECO Asia</a> concluded the first event earlier in September, bringing together emerging engineers, developers, and data scientists to tackle real-world business challenges through AI. Over several days, participants used Bitdeer AI Cloud&#x2019;s GPU <a href="https://www.bitdeer.ai/en/services/virtual-machine?ref=bitdeer.ai" rel="noreferrer">virtual machines</a> and <a href="https://www.bitdeer.ai/en/services/ai-inference?ref=bitdeer.ai" rel="noreferrer">AI model library</a> to build practical prototypes across multiple industries, including publishing, logistics, and consumer goods.</p><p>The event demonstrated how access to scalable computing and pre-trained AI models can dramatically shorten development cycles, transforming ideas into deployable solutions within days. It also reflected Bitdeer AI&#x2019;s commitment to nurturing the next generation of AI talents while enabling enterprises to accelerate digital transformation through intelligent infrastructure.</p><h3 id="case1-building-data-driven-market-insights"><strong>Case1: Building Data-Driven Market Insights</strong></h3><p><strong>Background: </strong>A Singapore-based publishing company with over three decades of history is one of the largest scientific publishers in the Asia-Pacific region. It produces more than 600 new titles and 130 journals annually in multiple languages, serving a diverse academic audience worldwide.</p><p><strong>Challenges: </strong>The company&#x2019;s teams spent extensive time manually researching and categorizing information from universities, journals, and books. Without real-time analytics or insight generation, the management struggled to identify high-performing content areas, forecast demand, or understand relationships between academic institutions and sales performance.</p><p><strong>Solution</strong>: Participants built an AI-powered analytics platform on Bitdeer AI Cloud using natural language processing and data-mining models. The system analyzed large datasets to uncover hidden connections between topics, readership, and market trends. With configurable prompts, users could automatically generate customized insights such as emerging research areas or optimal sales channels without technical expertise.</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2025/11/data-src-image-dc48ac93-4ed4-4f9b-a027-7e19dcda1a96.png" class="kg-image" alt="Hack to Hire: Empowering AI Innovation Through Bitdeer AI Cloud" loading="lazy" width="1272" height="438" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2025/11/data-src-image-dc48ac93-4ed4-4f9b-a027-7e19dcda1a96.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2025/11/data-src-image-dc48ac93-4ed4-4f9b-a027-7e19dcda1a96.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2025/11/data-src-image-dc48ac93-4ed4-4f9b-a027-7e19dcda1a96.png 1272w" sizes="(min-width: 720px) 720px"></figure><p><strong>Outcome: </strong>The prototype provided a foundation for intelligent decision-making, enabling the company to shift from manual data exploration to predictive market analysis. Executives could visualize real-time performance indicators, identify growth areas, and respond proactively to industry trends.</p><h3 id="case2-automating-pharmaceutical-supply-chain-management"><strong>Case2: Automating Pharmaceutical Supply Chain Management</strong></h3><p><strong>Background: </strong>A logistics company operating across Asia focuses on pharmaceutical distribution, custom packaging, and transshipment to Japan and Indonesia. Its operations relied heavily on manual processes for warehouse inspections and inventory tracking.</p><p><strong>Challenges: </strong>Recording product information and medical kits required significant time and manual effort. Warehouse data was captured in spreadsheets, with limited analytical visibility. These inefficiencies not only delayed reporting but also introduced human error, hindering the company&#x2019;s ability to monitor operations effectively.</p><p><strong>Solution: </strong>Using Bitdeer AI Cloud&#x2019;s GPU infrastructure, participants developed a mobile-friendly web application that combined photo capture with AI image recognition. Warehouse staff could take photos of cartons or medical kits, which were then uploaded for real-time analysis. AI models automatically identified SKUs, verified packaging, and updated inventory records.</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2025/11/data-src-image-7f672b14-494e-4cf9-bcec-e3585114503a.png" class="kg-image" alt="Hack to Hire: Empowering AI Innovation Through Bitdeer AI Cloud" loading="lazy" width="1268" height="444" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2025/11/data-src-image-7f672b14-494e-4cf9-bcec-e3585114503a.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2025/11/data-src-image-7f672b14-494e-4cf9-bcec-e3585114503a.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2025/11/data-src-image-7f672b14-494e-4cf9-bcec-e3585114503a.png 1268w" sizes="(min-width: 720px) 720px"></figure><p><strong>Outcome: </strong>The system significantly reduced manual data-entry time and error rates while improving inventory accuracy. Real-time analytics dashboards gave supervisors complete visibility into logistics operations and stock movement. The prototype illustrated how cloud-based AI can drive operational efficiency and compliance in the highly regulated pharmaceutical logistics sector.</p><h3 id="case-3-building-an-integrated-ai-ecosystem-in-fmcg"><strong>Case 3: Building an Integrated AI Ecosystem in FMCG</strong></h3><p><strong>Background: </strong>With over 50 years of market presence, this leading FMCG company had built a strong brand heritage. However, its digital infrastructure had become increasingly fragmented over time. The deployment of disparate systems for customer service, marketing, and internal operations led to entrenched data silos and disrupted information flow.</p><p><strong>Challenges: </strong>The organization grappled with two core operational gaps: a lack of centralized visibility for business reviews and the absence of a 24/7 customer service channel. This left decision-makers without a consolidated view of performance, while corporate chatbots remained offline after hours. Internally, significant productivity was lost as staff manually handled routine HR and IT inquiries.</p><p><strong>Solution: </strong>The hackathon team created a comprehensive AI ecosystem using Bitdeer AI Cloud. It comprised three interconnected components:</p><ol><li>A customer-facing chatbot powered by large language models to provide real-time, natural language interaction for users seeking information or support.</li><li>A business intelligence (BI) dashboard built on Microsoft BI, aggregating marketing, sales, and inventory data into interactive visualizations for leadership review.</li><li>A staff-facing assistant chatbot that automated HR and IT queries, improving efficiency and response time for internal teams.</li></ol><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2025/11/data-src-image-144317f7-0d43-4850-bec6-639b8ddf7802.png" class="kg-image" alt="Hack to Hire: Empowering AI Innovation Through Bitdeer AI Cloud" loading="lazy" width="1284" height="592" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2025/11/data-src-image-144317f7-0d43-4850-bec6-639b8ddf7802.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2025/11/data-src-image-144317f7-0d43-4850-bec6-639b8ddf7802.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2025/11/data-src-image-144317f7-0d43-4850-bec6-639b8ddf7802.png 1284w" sizes="(min-width: 720px) 720px"></figure><p>Each module was deployed and tested using Bitdeer AI Cloud&#x2019;s model library and virtual machines, enabling fast iteration and integration across the company&#x2019;s systems.</p><p><strong>Outcome: </strong>The prototype established a unified AI foundation for both customer engagement and internal operations. It improved responsiveness, enhanced data visibility, and reduced administrative workload demonstrating how enterprise-grade AI can drive organizational agility and connected intelligence.</p><h3 id="bitdeer-ai-cloud-the-foundation-for-scalable-innovation"><strong>Bitdeer AI Cloud: The Foundation for Scalable Innovation</strong></h3><p>All three projects were developed and deployed on Bitdeer AI Cloud, which provided participants with GPU computing, pre-trained AI models, and flexible development environments. The platform allowed teams to build, train, and deploy models without complex setup or infrastructure management.</p><p>By streamlining AI experimentation and deployment, Bitdeer AI Cloud enabled participants to focus on solving business problems and delivering measurable outcomes. The hackathon served as a real-world demonstration of how robust AI infrastructure accelerates both learning and innovation.</p><h3 id="accelerating-the-future-of-ai-development"><strong>Accelerating the Future of AI Development</strong></h3><p>Hack to Hire demonstrated how advanced cloud infrastructure and AI technologies can support digital transformation across a wide range of industries. In just a few days, participants developed functioning prototypes that addressed key business challenges in sectors from manufacturing to retail, improving efficiency, operational intelligence, and innovation capabilities.</p><p>Bitdeer AI delivers high-performance, cost-effective, and production-ready AI solutions, helping organizations move projects from concept to implementation and achieve measurable business impact. The event also underscored a growing reality: scalable AI infrastructure is increasingly essential for driving intelligent growth, optimizing operations, and enabling innovation across industries. This successful round paves the way for the next as Hack to Hire returns soon with more opportunities for innovation and collaboration. </p>]]></content:encoded></item><item><title><![CDATA[Power Capacity and the Future of AI]]></title><description><![CDATA[As AI expands into a full-scale industrial ecosystem, its surging electricity demand raises a critical question on whether global power capacity can sustain it.]]></description><link>https://www.bitdeer.ai/en/blog/ai-power-capacity-shortage-and-the-future-of-ai/</link><guid isPermaLink="false">6900844068d90b0001fa6e1d</guid><category><![CDATA[AI Trends & Industry News]]></category><dc:creator><![CDATA[Taylor Ye]]></dc:creator><pubDate>Tue, 28 Oct 2025 09:29:47 GMT</pubDate><media:content url="https://www.bitdeer.ai/en/blog/content/images/2025/10/the-ai-power-explosion-EN.png" medium="image"/><content:encoded><![CDATA[<img src="https://www.bitdeer.ai/en/blog/content/images/2025/10/the-ai-power-explosion-EN.png" alt="Power Capacity and the Future of AI"><p>As artificial intelligence evolves into a full-scale industrial ecosystem, it is also consuming an unprecedented amount of electricity. Whether it is generative models, large language models (LLMs), or multimodal agents, their operation within data centers relies on a common physical foundation: power capacity.</p><p>Behind the rapid progress of AI, energy consumption has become a growing concern. Some have noted that &#x201C;the endgame of AI is compute, and the endgame of compute is power.&#x201D; As models expand and real-time inference becomes the norm, the industry faces a new and pressing question: Do we have enough electricity to power it all?</p><h2 id="the-exponential-rise-in-ai-power-demand"><strong>The Exponential Rise in AI Power Demand</strong></h2><p>According to the <a href="https://www.iea.org/news/as-energy-and-ai-links-grow-new-iea-observatory-provides-latest-data-and-analysis?utm_source=chatgpt.com"><u>International Energy Agency</u></a> (IEA), global data center electricity consumption is expected to double by 2030, reaching around 945 terawatt-hours (TWh) annually. Among them, AI-optimized high-performance data centers may consume up to four times more electricity than they do today equivalent to adding the annual power consumption of a medium-sized developed country to the global grid.</p><p>Traditional data centers typically consume between 10 and 25 megawatts (MW), while hyperscale AI-focused facilities can exceed 100 MW, with annual usage comparable to that of 100,000 households. These AI-centric facilities continue to expand rapidly to accommodate ever-larger models and rising demand for AI services.</p><p>Training GPT-3 reportedly consumed roughly 1.3 gigawatt-hours (GWh) of electricity. Generating a medium-length response of around 1,000 tokens with GPT-5 consumes an average of 18.35 watt-hours (Wh) equivalent to running a microwave oven for about 76 seconds and can reach up to 40 Wh, nearly 8.6 times higher than GPT-4&#x2019;s 2.12 Wh. Every new generation of AI models, especially those with hundreds of billions of parameters, requires longer training cycles, denser GPU clusters, and sustained high power draw.</p><p>In short, electricity has become the new currency of computation.</p><h2 id="how-power-capacity-shapes-ai-development"><strong>How Power Capacity Shapes AI Development</strong></h2><p>&#x201C;Speed-to-Power&#x201D; has emerged as a key metric of competitiveness. In the global race for AI infrastructure, the speed at which a data center can secure and activate a stable power supply often determines how fast AI capacity can be deployed. Even with sufficient land, cooling, and connectivity, projects can stall if grid access is delayed.</p><p>The expansion of electricity generation and transmission infrastructure remains slow. From planning and permitting to construction and operation, the process can take years, while AI demand grows on a quarterly cycle. Utilities must balance the needs of data centers with stable residential and industrial supply. The United States, home to the world&#x2019;s largest concentration of AI data centers, is already nearing grid stability thresholds in several regions.</p><p>AI training workloads can trigger sudden and intense power surges, placing stress on local grids. When massive GPU clusters ramp up simultaneously, they can cause voltage fluctuations, particularly in systems integrated with variable renewable energy. This highlights that significant improvements in AI-related power infrastructure are still required.</p><h2 id="bridging-the-ai-power-gap"><strong>Bridging the AI Power Gap</strong></h2><p>The accelerating power consumption of AI has become a central challenge for enterprises. As electricity availability becomes a defining factor in AI development, many technology companies are enhancing their internal optimization and diversifying their energy strategies to strengthen resilience and reduce reliance on external grid expansion or policy incentives.</p><p>The first priority is energy efficiency and computational optimization. Through model compression, sparse computation, quantization, and dynamic scheduling, AI developers are striving to extract more compute per watt. Data center architectures are also evolving, with liquid cooling and high-density rack designs driving continuous improvements in PUE (Power Usage Effectiveness).</p><p>The second approach is energy diversification and storage development. Many AI infrastructure operators are deploying distributed generation and on-site storage systems such as campus-scale solar, wind, and battery energy storage systems (BESS) to balance peak loads and increase energy flexibility.</p><p>In addition, intelligent power management systems are becoming standard. AI-driven energy monitoring platforms can track real-time power fluctuations across GPU clusters, predict peaks, and automatically adjust supply strategies to achieve simultaneous optimization of computational scheduling and energy efficiency.</p><p>Finally, enterprise collaboration is emerging as a key enabler. Leading AI companies and energy providers are co-investing in next-generation power infrastructure, signing long-term power purchase agreements (PPAs), and accelerating the integration of green electricity and regional energy interconnections.</p><p>Overall, the industry is shifting from simply scaling compute to achieving smarter energy utilization pursuing higher efficiency and stability in a world of finite electricity.</p><h2 id="from-resource-to-strategic-asset"><strong>From Resource to Strategic Asset</strong></h2><p>Electricity capacity is no longer just a technical metric. It has become a strategic lever connecting digital innovation, industrial strategy, and the global energy transition.</p><figure class="kg-card kg-image-card"><img src="https://www.bitdeer.ai/en/blog/content/images/2025/10/image-1.png" class="kg-image" alt="Power Capacity and the Future of AI" loading="lazy" width="1270" height="708" srcset="https://www.bitdeer.ai/en/blog/content/images/size/w600/2025/10/image-1.png 600w, https://www.bitdeer.ai/en/blog/content/images/size/w1000/2025/10/image-1.png 1000w, https://www.bitdeer.ai/en/blog/content/images/2025/10/image-1.png 1270w" sizes="(min-width: 720px) 720px"></figure><p>In the coming decade, competition in AI will not only revolve around models or chips but also around who can secure reliable, efficient, and sustainable power the fastest.</p><h2 id="conclusion"><strong>Conclusion</strong></h2><p>As global AI demand accelerates, power capacity has shifted from being a constraint to becoming a strategic advantage. The next breakthroughs in AI will depend not only on better semiconductors and smarter algorithms but also on who can deliver scalable, clean, and resilient power infrastructure.</p><p>At Bitdeer AI, we are expanding our dedicated AI power capacity to 200 MW of IT load by 2026 as part of our long-term strategy to ensure sustainable, high-performance infrastructure across our global data center network. We believe that in the era of intelligence, every watt of energy fuels innovation.</p>]]></content:encoded></item><item><title><![CDATA[NVIDIA GPU Evolution and the Road Ahead]]></title><description><![CDATA[Review NVIDIA GPU evolution from CUDA to NVLink/NVSwitch to Blackwell & Vera Rubin, how design advances remove bottlenecks and power modern AI.]]></description><link>https://www.bitdeer.ai/en/blog/nvidia-gpu-evolution-and-the-road-ahead/</link><guid isPermaLink="false">68f1bb6868d90b0001fa6e0d</guid><category><![CDATA[AI Applications]]></category><dc:creator><![CDATA[Taylor Ye]]></dc:creator><pubDate>Fri, 17 Oct 2025 03:54:44 GMT</pubDate><media:content url="https://www.bitdeer.ai/en/blog/content/images/2025/10/The-Evolution-of--NVIDIA-GPUs-en@2x.png" medium="image"/><content:encoded><![CDATA[<img src="https://www.bitdeer.ai/en/blog/content/images/2025/10/The-Evolution-of--NVIDIA-GPUs-en@2x.png" alt="NVIDIA GPU Evolution and the Road Ahead"><p>GPUs have become the backbone of modern Artificial Intelligence (AI), High-Performance Computing (HPC), and Generative AI. NVIDIA has played a pivotal role in this transformation, evolving from a graphics accelerator vendor into the enabler of factory-scale AI platforms. Each architectural generation has been driven by one principle: resolve systemic bottlenecks that limit compute, memory, or scalability.</p><p>This article provides a systematic review of NVIDIA GPU architecture, from CUDA programmability to NVLink and NVSwitch breakthroughs, and extends to Blackwell and the upcoming Vera Rubin platform, exploring how GPU evolution is shaping the present and future of intelligent computing.</p><h2 id="what-is-a-gpu"><strong>What is a GPU</strong></h2><p>To understand the direction of GPU evolution, it is essential to first examine its structure. A GPU (Graphics Processing Unit) was originally designed for image rendering but has since evolved into the core engine driving AI and HPC. Unlike CPUs (Central Processing Units), which emphasize low-latency execution of single threads, GPUs are designed for massive parallelism and high throughput.</p><p>This architecture gives GPUs an unrivaled advantage in handling workloads that require thousands of tasks in parallel, making them indispensable for deep learning training, inference, and scientific simulations. For more details on key GPU parameters refer to our earlier article <em>&#x201C;</em><a href="https://www.bitdeer.ai/en/blog/demystifying-gpus-for-ai-beginners-what-you-need-to-know/"><u>Demystifying GPUs for AI Beginners: What You Need to Know</u></a>&#x201D;.</p><p><strong>The Evolution of NVIDIA GPU Architectures</strong></p><p>While GPUs excel at throughput through parallel design, they have also faced long-standing bottlenecks: programmability challenges, limited memory bandwidth, power efficiency constraints, and multi-GPU communication overhead. NVIDIA&#x2019;s history of architectural innovation has been a process of systematically overcoming these constraints.</p><h3 id="from-graphics-to-general-purpose-computing-1999%E2%80%932012"><strong>From Graphics to General-Purpose Computing (1999&#x2013;2012)</strong></h3><p><strong>Key Milestones:</strong></p><ul><li>Tesla (2006): CUDA programmability opened GPUs to scientific and industrial computing</li><li>Fermi &amp; Kepler (2010&#x2013;2012): Expanded memory hierarchy, improved efficiency, enabled supercomputers</li></ul><p><strong>Industry and Business Impact:</strong></p><p>This was the stage when GPUs escaped the &#x201C;graphics only&#x201D; box. CUDA made it possible to run parallel workloads in physics, weather forecasting, and financial simulations. It was the foundation for enterprises to start viewing GPUs as compute engines rather than gaming chips. Enterprises gained access to affordable high-performance computing, which reduced barriers for R&amp;D in pharmaceuticals, finance, and engineering</p><h3 id="the-ai-breakthrough-era-2014%E2%80%932022"><strong>The AI Breakthrough Era (2014&#x2013;2022)</strong></h3><p><strong>Key Milestones:</strong></p><ul><li>Pascal (2016): FP16 precision, NVLink 1.0, enabling large-scale deep learning</li><li>Volta (2017): Tensor Cores, breakthrough for neural network training</li><li>Ampere (2020): TF32 and INT8, scaling both training and inference</li></ul><p><strong>Industry and Business Impact:</strong></p><p>This was when GPUs became synonymous with AI. Tensor Cores in particular changed the economics of training neural networks, cutting costs and time dramatically. NVLink enabled distributed AI across multiple GPUs, a prerequisite for modern LLMs. Deep learning moved from research labs into production. Voice assistants, computer vision in retail and manufacturing, and predictive analytics all became commercially viable. Companies that adopted early built durable competitive advantages in automation and personalization.</p><h3 id="the-generative-ai-era-2022%E2%80%932024"><strong>The Generative AI Era (2022&#x2013;2024)</strong></h3><p><strong>Key Milestones:</strong></p><ul><li>Hopper (2022): FP8 precision, NVLink 4.0, and the introduction of confidential computing features designed for secure large-scale training.</li><li>Blackwell (2024): NVLink 5.0, Grace CPU integration, data-center scale AI factories</li></ul><p><strong>Industry and Business Impact:</strong></p><p>These platforms weren&#x2019;t just accelerators but entire AI factories. Hopper made training trillion-parameter models possible, while Blackwell introduced factory-scale scalability, integrating CPUs and GPUs seamlessly. Enterprises could deploy generative AI copilots, real-time recommendation systems, and domain-specific AI platforms. Generative AI shifted from experimental pilots to core business strategy, transforming productivity, customer engagement, and competitive differentiation.</p><h2 id="beyond-hardware-ecosystem-and-scalability"><strong>Beyond Hardware: Ecosystem and Scalability</strong></h2><p>The evolution of NVIDIA GPUs is not only about raw compute, memory, and interconnect, but also the supporting ecosystem that translates advances into usable performance.</p><ul><li><strong>CUDA (Software Foundation):</strong> Translates GPU parallelism into programmability, offering libraries such as cuBLAS, cuDNN, and TensorRT.</li><li><strong>Tensor Cores:</strong> Redefined training efficiency for neural networks.</li><li><strong>Memory Hierarchy:</strong> Transitioned from GDDR &#x2192; HBM &#x2192; HBM3e, reaching TB/s-class bandwidth. Hopper expanded cache coherence, while Blackwell increased both capacity and speed.</li><li><strong>NVLink (Interconnect):</strong> Overcame CPU-GPU and GPU-GPU communication bottlenecks. Hopper&#x2019;s NVLink 4 delivered ~900 GB/s aggregate bandwidth per GPU; Blackwell&#x2019;s NVLink 5 doubled to ~1.8 TB/s and added unified memory addressing.</li><li><strong>NVSwitch (Scalability):</strong> Extended NVLink into full-switch fabrics, enabling multi-GPU systems to function as a single logical accelerator&#x2014;critical for distributed AI training and cluster-scale AI factories.</li></ul><p>Each step in this roadmap removed a bottleneck and expanded what enterprises could achieve with AI.</p><h2 id="business-implications"><strong>Business Implications</strong></h2><p>For AI, this represents a shift from single-device limits to industrial-scale systems. Training efficiency has accelerated with specialized compute, memory capacity has scaled with HBM innovations, and multi-GPU fabrics now enable trillion-parameter model deployment.</p><ul><li>Large-scale model training is no longer confined to hyperscalers. With access to modern GPU clusters, enterprises can now build and fine-tune large language models, creating proprietary AI assets that differentiate them in the market.</li><li>Inference services have matured into enterprise-grade platforms, enabling copilots, assistants, and automated decision systems to scale reliably across entire workforces.</li><li>Industries such as finance, manufacturing, and healthcare can now operationalize digital twins, scenario simulations, and predictive analytics that were once economically prohibitive.</li><li>At the same time, leaders must balance the benefits with rising energy costs, infrastructure complexity, and sustainability challenges and decisions that increasingly shape competitive positioning and industry consolidation.</li></ul><p>GPUs have evolved from accelerators into the backbone of enterprise-scale AI. The question is no longer how fast they run but what new possibilities they create. The focus has shifted from speed to scale and from experimentation to core infrastructure. Competitiveness now depends on aligning GPU strategy with business outcomes, turning compute capability into productivity gains, innovation capacity, and long-term market advantage.</p><h2 id="future-development-vera-rubin-2026"><strong>Future Development: Vera Rubin (2026)</strong></h2><p>Following Blackwell, NVIDIA announced in September 2025 that its next-generation <a href="https://nvidianews.nvidia.com/news/nvidia-unveils-rubin-cpx-a-new-class-of-gpu-designed-for-massive-context-inference?ref=bitdeer.ai"><u>Vera Rubin </u></a>architecture will launch in 2026. The Rubin CPX GPU and Vera Rubin NVL144 platform deliver 8 exaflops of compute, 100TB of fast memory, and 1.7 PB/s memory bandwidth, designed for million-token context processing and generative video. Vera Rubin points directly to ultra-long context AI applications, where models will no longer be limited to thousand-token dialogues but can instead process entire codebases, hours-long video, and multimodal histories, driving advances in AI agents, automated code generation, and creative media production.</p><p>While Rubin&#x2019;s capabilities are groundbreaking, they also amplify existing challenges: ensuring stability and consistency across million-token contexts, balancing extreme compute and memory with sustainable deployment, and strengthening data security and privacy in multi-tenant environments. These issues will define the Rubin era and beyond.</p><h2 id="conclusion"><strong>Conclusion</strong></h2><p>GPU evolution is not just about increasing raw compute, but about the coordinated advancement of compute, memory, and interconnect. Each architectural upgrade has solved a new bottleneck, enabling larger and more complex AI models. From programmability to distributed scalability, GPUs have become the engine that moves AI from research to industrial-scale deployment.</p><p>The upcoming Vera Rubin platform highlights the next direction: not only faster, but also more specialized. Designed for long-context, multimodal, and system-level AI, Rubin represents the transformation of GPUs from accelerators into the core infrastructure of AI factories.</p><p>At Bitdeer AI, we build cloud infrastructure designed for this evolution, combining high-density GPU clusters, optimized cooling, and resilient network architecture. Our platform supports diverse AI workloads at scale, from model training to inference, enabling enterprises to deploy large, complex AI applications efficiently. With seamless scalability and integrated management tools, our cloud makes AI simple.</p>]]></content:encoded></item></channel></rss>