Quick PSA. We were made aware by our community that the additional streamflow $COM rewards weren't being distributed correctly to stakers. This is in addition to the GPU hours. We've decided to remedy this by airdropping another 5M $COM to those who staked on streamflow, but weren't able to get rewards! We will launch a second streamflow contract without rewards as these were intended to get more people to stake initially while we were getting everything else up and running.
If you're already staking, expect to get airdropped your rewards. This should happen starting next week and until 8/31/2025. You don't need to restake to receive GPU rewards. But be aware there will be no more $COM rewards distributed via the old streamflow contract. The new streamflow contract will remove $COM rewards completetly to prevent confusion among stakers.
Thank you 🌋
If you're already staking, expect to get airdropped your rewards. This should happen starting next week and until 8/31/2025. You don't need to restake to receive GPU rewards. But be aware there will be no more $COM rewards distributed via the old streamflow contract. The new streamflow contract will remove $COM rewards completetly to prevent confusion among stakers.
Thank you 🌋
👏2
$COM streamflow staking rewards are live. Thanks for bearing with us while we got the airdrop setup. https://app.streamflow.finance/airdrops/solana/mainnet/HjCjVANUof9MxdxFVSkQKaMZtDstGbTvp1mFAiisJNyR (updated URL)
Now go explain to your mom what a B200 is. $COM fam🌋
Now go explain to your mom what a B200 is. $COM fam🌋
app.streamflow.finance
Streamflow - Token vesting, airdrops, staking & more
Streamflow is the leading multi-chain asset streaming protocol specializing in token vesting, streaming payments and treasury management solutions.
🔥6😡1
A few folks reached out privately asking for a breakdown of this chart, so let's dive in and unpack it properly.
First off, the knee-jerk reaction might be: "Oh, this is just a post-COVID shift." Not quite. COVID's major impacts wrapped up around 2022, while ChatGPT and the broader AI boom really hit the headlines in 2023. So, timing-wise, that doesn't add up.
What we're looking at here isn't total servers installed or even the number of data centers built—it's total spending, in dollars. The Y-axis tells the story: raw capital outlay.
Now, you might wonder, "Servers and PCs cost about the same today as they did in 2022—where's all this extra money going?" The key insight? These aren't your standard data centers or everyday servers. We're talking AI data centers packed with specialized AI servers—like NVIDIA's B200, H200, and H100 GPUs. These beasts are exponentially more expensive than traditional hardware, which is exactly what creates that sharp "kink" in the chart.
Zoom in further, and you'll notice the monthly growth is remarkably linear. If you're familiar with manufacturing, this makes perfect sense: it mirrors TSMC's steady ramp-up in production capacity to churn out GPUs for NVIDIA. Supply chains don't scale overnight—they build predictably, and that's reflected here.
This chart essentially captures our project's core thesis in visual form. AI doesn't have to break the bank. There's already an oversupply of GPUs out there, and everything in AI—from training to inference—can be handled efficiently in a distributed way. We're all about running open-source models, sourcing hardware globally, and delivering massive bang for your buck. But it's bigger than cost savings: it's about true control, uncensored access, and ownership over your AI workflows. We build everything in the open and aim to be the premier platform for hosting, fine-tuning, and deploying models.
And what about the top line? Sure, you could frame it as "AI replacing humans." But we see the brighter side: AI supercharges productivity, making people 10x more efficient—without chaining them to an office desk. It's empowerment, not displacement.
What do you all think? Let us know in the comments.
First off, the knee-jerk reaction might be: "Oh, this is just a post-COVID shift." Not quite. COVID's major impacts wrapped up around 2022, while ChatGPT and the broader AI boom really hit the headlines in 2023. So, timing-wise, that doesn't add up.
What we're looking at here isn't total servers installed or even the number of data centers built—it's total spending, in dollars. The Y-axis tells the story: raw capital outlay.
Now, you might wonder, "Servers and PCs cost about the same today as they did in 2022—where's all this extra money going?" The key insight? These aren't your standard data centers or everyday servers. We're talking AI data centers packed with specialized AI servers—like NVIDIA's B200, H200, and H100 GPUs. These beasts are exponentially more expensive than traditional hardware, which is exactly what creates that sharp "kink" in the chart.
Zoom in further, and you'll notice the monthly growth is remarkably linear. If you're familiar with manufacturing, this makes perfect sense: it mirrors TSMC's steady ramp-up in production capacity to churn out GPUs for NVIDIA. Supply chains don't scale overnight—they build predictably, and that's reflected here.
This chart essentially captures our project's core thesis in visual form. AI doesn't have to break the bank. There's already an oversupply of GPUs out there, and everything in AI—from training to inference—can be handled efficiently in a distributed way. We're all about running open-source models, sourcing hardware globally, and delivering massive bang for your buck. But it's bigger than cost savings: it's about true control, uncensored access, and ownership over your AI workflows. We build everything in the open and aim to be the premier platform for hosting, fine-tuning, and deploying models.
And what about the top line? Sure, you could frame it as "AI replacing humans." But we see the brighter side: AI supercharges productivity, making people 10x more efficient—without chaining them to an office desk. It's empowerment, not displacement.
What do you all think? Let us know in the comments.
🔥3😡1
Today, we squashed a couple of bugs in race conditions that only occur at our scale. We also shipped a couple of improvements to our custom container orchestration layer. This will improve security, isolation and uptime and will work better on desktop PCs and not just 8xB200's.
That's right, we always make sure our stack runs on desktop PCs and even on Windows. You'll know why soon enough.
That's right, we always make sure our stack runs on desktop PCs and even on Windows. You'll know why soon enough.
🤩1🤪1
Forwarded from Aya Hackathon Channel
🏆 Aya AI Hackathon – 🇮🇳 Bangalore Winners Announced 🏆
Our Bangalore winner tweet is live on X!
🔗 Tweet: https://x.com/theayawallet/status/1956626774806479207
How to help (takes 30–60s):
1️⃣ Like ❤️ + RT 🔁 the tweet
2️⃣ Quote-tweet with a shout-out to your favorite team
3️⃣ Tag partners & IIIT-B to boost reach
4️⃣ Drop a congrats comment 🙌
Let’s get these builders the spotlight they deserve! 📣
Our Bangalore winner tweet is live on X!
🔗 Tweet: https://x.com/theayawallet/status/1956626774806479207
How to help (takes 30–60s):
1️⃣ Like ❤️ + RT 🔁 the tweet
2️⃣ Quote-tweet with a shout-out to your favorite team
3️⃣ Tag partners & IIIT-B to boost reach
4️⃣ Drop a congrats comment 🙌
Let’s get these builders the spotlight they deserve! 📣
🤔1
Another awesome event by Aya!
These teams built their projects on @comput3ai and we gave them some GPU hours so they can keep cooking.
1️⃣ Team Hackxbot — Multi-Chain AI-Powered Crypto Manager
2️⃣ Team Delta — SnackGpt
3️⃣ Suyash — AI-Powered Crypto Assistant
These teams built their projects on @comput3ai and we gave them some GPU hours so they can keep cooking.
1️⃣ Team Hackxbot — Multi-Chain AI-Powered Crypto Manager
2️⃣ Team Delta — SnackGpt
3️⃣ Suyash — AI-Powered Crypto Assistant
🔥2
Today's the big day, we just deployed
Alright @everyone it's the big day.
This is work in progress. We are doing a ton of stuff that you can't see, but it is up and we would appreciate testing and feedback. The inferencing API is now running on B200's. Right now everyone has access to all the models, we'll be closing this down as we go along.
We'll be launching paid subscriptions, we had a lot of discussions and honestly these models are so powerful that we'll probably be targeting $99/mo, but we'll start with $75/mo as a welcome offer. You won't have to use phantom, it'll be email and stripe (we'll do crypto too don't worry). We will be giving stakers access to what we consider reasonable limits (input/output tokens per month), but they'll be less than the subscriptions at least initially. Eventually we hope we can just map stakers to these plans, but for the moment we just added a bunch of 24/7 infra, which we're running in parallel to the GPU workloads most of you are familiar with.
The models on the api include the
Keep in mind these are monstrous models and you might hit the Cloudflare's request timeouts after 100 seconds. That's not us - that's a cloudflare "feature". We're working on the
Enjoy and let us know how it goes. Here's some curls to get you started.
You can access these as a normal
kimi-k2 and qwen3-coder:480b to our api. If you want to play around with this, hop on our discord: https://discord.com/invite/DmaHAdGcNyAlright @everyone it's the big day.
This is work in progress. We are doing a ton of stuff that you can't see, but it is up and we would appreciate testing and feedback. The inferencing API is now running on B200's. Right now everyone has access to all the models, we'll be closing this down as we go along.
We'll be launching paid subscriptions, we had a lot of discussions and honestly these models are so powerful that we'll probably be targeting $99/mo, but we'll start with $75/mo as a welcome offer. You won't have to use phantom, it'll be email and stripe (we'll do crypto too don't worry). We will be giving stakers access to what we consider reasonable limits (input/output tokens per month), but they'll be less than the subscriptions at least initially. Eventually we hope we can just map stakers to these plans, but for the moment we just added a bunch of 24/7 infra, which we're running in parallel to the GPU workloads most of you are familiar with.
The models on the api include the
hermes3:70b and llama3:70b, both now running at 64K context (thanks B200s), but the big news is kimi-k2 and qwen3-coder. Right now everyone with an api key has access to these models. We'll be limiting them to stakers and subscribers later this week.Keep in mind these are monstrous models and you might hit the Cloudflare's request timeouts after 100 seconds. That's not us - that's a cloudflare "feature". We're working on the
api.compute3.ai (with an 'e') deployment as we speak, and that will be naked (no cloudflare, just a couple of load balancers) so it won't have that issue. We'll be adding instructions on how to set all of this up with claude code, qwen code and probably codex as well. In the mean time feel free to test, api.comput3.ai is now running a litellm-proxy so it should handle most of the shenanigans that the coding agents need. However, we're not routing /anthropic/* (or /openai/ for that matter) to our models yet, we're looking for a reasonable way to be able to leave you control over which goes where instead of picking one for you. I do believe overloading the /model should work, but I haven't tested this yet.Enjoy and let us know how it goes. Here's some curls to get you started.
$ curl -X GET https://api.comput3.ai/v1/models | jq
{
"data": [
{
"id": "hermes3:70b",
"object": "model",
"created": 1677610602,
"owned_by": "openai"
},
{
"id": "kimi-k2",
"object": "model",
"created": 1677610602,
"owned_by": "openai"
},
{
"id": "qwen3-coder:480b",
"object": "model",
"created": 1677610602,
"owned_by": "openai"
},
{
"id": "llama3:70b",
"object": "model",
"created": 1677610602,
"owned_by": "openai"
}
],
"object": "list"
}
You can access these as a normal
/v1 API.curl -X POST https://api.comput3.ai/v1/chat/completions \
-H "Authorization: Bearer c3_api_YOUR_API_KEY" \
-H "Content-Type: application/json" \
-d '{
"model": "kimi-k2",
"messages": [
{
"role": "system",
"content": "You are a helpful assistant."
},
{
"role": "user",
"content": "What is 2+2?"
}
]
}'
{"id":"chatcmpl-007e9d34e4474e78902934247b1f72bc","created":1755501320,"model":"kimi-k2","object":"chat.completion","system_fingerprint":null,"choices":[{"finish_reason":"stop","index":0,"message":{"content":"2 + 2 = 4","role":"assistant","tool_calls":null,"function_call":null},"provider_specific_fields":{"stop_reason":163586}}],"usage":{"completion_tokens":8,"prompt_tokens":24,"total_tokens":32,"completion_tokens_details":null,"prompt_tokens_details":null},"service_tier":null,"prompt_logprobs":null,"kv_transfer_params":null}
🔥6❤4💯2👏1
vibe is live 🌋🌋🌋
export ANTHROPIC_BASE_URL=api.comput3.ai
export ANTHROPIC_AUTH_TOKEN=c3_api_YOUR_C3_API_KEY
export ANTHROPIC_MODEL=kimi-k2
👍2💯1
Line in the sand.
Today we proved that an open-source coding agent can draft, refactor and ship production-grade features entirely on a permission-less mesh of GPUs. No metered tokens. No five-figure OpenAI bills. Just the flat cost of keeping silicon warm for a month.
That is the ceiling now.
What sits below it is the floor we just raised: an agent that can bootstrap its own rent. A trading model, a data pipeline, a risk engine—pick the workload—can now reinvest a sliver of the value it creates to cover its own compute. Continuously. Transparently. On-chain.
This is not a pitch deck slide.
It’s a live circuit: open weights → open ledger → open market.
The loop is closed, the lights stay on, and no one holds the off switch.
We’re not heralding sentience; we’re delivering sovereignty. Autonomous systems that pay their way, improve without custodians, and keep their state anchored to a public chain anyone can audit. If that sounds like the original promise of crypto meeting the original promise of AI, good—you’re paying attention.
We just planted the first stake.
The grid is live.
The door is open.
Are you with us?
$COM 🌋
Today we proved that an open-source coding agent can draft, refactor and ship production-grade features entirely on a permission-less mesh of GPUs. No metered tokens. No five-figure OpenAI bills. Just the flat cost of keeping silicon warm for a month.
That is the ceiling now.
What sits below it is the floor we just raised: an agent that can bootstrap its own rent. A trading model, a data pipeline, a risk engine—pick the workload—can now reinvest a sliver of the value it creates to cover its own compute. Continuously. Transparently. On-chain.
This is not a pitch deck slide.
It’s a live circuit: open weights → open ledger → open market.
The loop is closed, the lights stay on, and no one holds the off switch.
We’re not heralding sentience; we’re delivering sovereignty. Autonomous systems that pay their way, improve without custodians, and keep their state anchored to a public chain anyone can audit. If that sounds like the original promise of crypto meeting the original promise of AI, good—you’re paying attention.
We just planted the first stake.
The grid is live.
The door is open.
Are you with us?
$COM 🌋
🔥14❤1
🚀 The word is out: we’re quietly becoming the hottest infrastructure layer in AI.
Kimi-K2 and Qwen3-Coder models are live, FP8-hosted, full context, B200-powered, and served outside China—exactly the combo the market was waiting for.
• Devs are already shipping at lightspeed—some can’t believe the integration is this easy.
• ComfyUI is dropping a fresh build that unlocks Qwen’s new image & editing models.
• GPU network is humming; next-gen LLMs are spinning up daily.
If your project needs compute or input/output tokens to power your AI apps, we’re taking those conversations now.
DMs are open, but discord is where the action is. Just get in touch with us.
https://discord.com/invite/DmaHAdGcNy
Kimi-K2 and Qwen3-Coder models are live, FP8-hosted, full context, B200-powered, and served outside China—exactly the combo the market was waiting for.
• Devs are already shipping at lightspeed—some can’t believe the integration is this easy.
• ComfyUI is dropping a fresh build that unlocks Qwen’s new image & editing models.
• GPU network is humming; next-gen LLMs are spinning up daily.
If your project needs compute or input/output tokens to power your AI apps, we’re taking those conversations now.
DMs are open, but discord is where the action is. Just get in touch with us.
https://discord.com/invite/DmaHAdGcNy
Discord
Join the Comput3.AI Discord Server!
Check out the Comput3.AI community on Discord - hang out with 406 other members and enjoy free voice and text chat.
🔥5💯1
This media is not supported in your browser
VIEW IN TELEGRAM
Quickest way to test kimi-k2 and qwen3-coder running on comput3.ai directly from your browser! 🌋
🔥2👏2❤1
5h until $COM 🌋 errupts on stage at the ElizaOS Accelerator demo day—we ignite the permissionless GPU mesh fueling tomorrow’s AI. Private workloads, open-source SOTA models, community powered AI training. Come feel the thunder. https://www.youtube.com/watch?v=neBg4yK_FwM
YouTube
elizaOS Accelerator Demo Day
Welcome to the elizaOS Accelerator Demo Day!
Over a 4-week program, we have worked closely together with amazing startups that are shaping the future of AI and Web3.
These incredible teams are now ready to pitch and take their companies to the next level.…
Over a 4-week program, we have worked closely together with amazing startups that are shaping the future of AI and Web3.
These incredible teams are now ready to pitch and take their companies to the next level.…
🔥3👏1
Are you ready? 🌋🌋🌋
https://www.youtube.com/watch?v=neBg4yK_FwM
https://www.youtube.com/watch?v=neBg4yK_FwM
YouTube
elizaOS Accelerator Demo Day
Welcome to the elizaOS Accelerator Demo Day!
Over a 4-week program, we have worked closely together with amazing startups that are shaping the future of AI and Web3.
These incredible teams are now ready to pitch and take their companies to the next level.…
Over a 4-week program, we have worked closely together with amazing startups that are shaping the future of AI and Web3.
These incredible teams are now ready to pitch and take their companies to the next level.…
🔥3💯2
Media is too big
VIEW IN TELEGRAM
$COM: The Hottest AI Infrastructure on Earth 🌋🌋🌋
Stop paying Big Tech mark-ups. Meet Comput3.AI — the infrastructure that just flipped the script on AI centralization.
* 18,000 GPU workloads launched since day one
* 7,400+ GPUs already connected
* 130+ stakers locking 25 % of the token supply
* 850 developers, 1,400 holders, 30+ projects LIVE
Demos:
1. Spin up a private GPU in 3 seconds and mint a cyberpunk AI avatar on the spot.
2. Plug any model (KimiK2, Llama-3, SDXL…) into your app via one crypto-payable API—already 40 % cheaper than AWS and the only service that accepts SOL / USDC / COM.
3. Train your own frontier model with distributed B200s—mint an NFT to lock your stake, earn 50 % higher rewards, and watch 50 % of revenue flow straight into COM buybacks.
Retail investors finally get a seat at the AI table.
No KYC. No censorship.
Just open-source, community-owned compute that prints yield while it prints the future.
👇 Smash subscribe, stake COM, and let’s decentralize the GPU monopoly—together.
Launch a GPU right now: https://launch.comput3.ai
Check out the slides: https://pitch.comput3.ai
Get an API key and try out our models: https://launch.comput3.ai
Discord: https://discord.com/invite/DmaHAdGcNy
Github: https://github.com/comput3ai
CA: J3NrhzUeKBSA3tJQjNq77zqpWJNz3FS9TrX7H7SLKcom
Stop paying Big Tech mark-ups. Meet Comput3.AI — the infrastructure that just flipped the script on AI centralization.
* 18,000 GPU workloads launched since day one
* 7,400+ GPUs already connected
* 130+ stakers locking 25 % of the token supply
* 850 developers, 1,400 holders, 30+ projects LIVE
Demos:
1. Spin up a private GPU in 3 seconds and mint a cyberpunk AI avatar on the spot.
2. Plug any model (KimiK2, Llama-3, SDXL…) into your app via one crypto-payable API—already 40 % cheaper than AWS and the only service that accepts SOL / USDC / COM.
3. Train your own frontier model with distributed B200s—mint an NFT to lock your stake, earn 50 % higher rewards, and watch 50 % of revenue flow straight into COM buybacks.
Retail investors finally get a seat at the AI table.
No KYC. No censorship.
Just open-source, community-owned compute that prints yield while it prints the future.
👇 Smash subscribe, stake COM, and let’s decentralize the GPU monopoly—together.
Launch a GPU right now: https://launch.comput3.ai
Check out the slides: https://pitch.comput3.ai
Get an API key and try out our models: https://launch.comput3.ai
Discord: https://discord.com/invite/DmaHAdGcNy
Github: https://github.com/comput3ai
CA: J3NrhzUeKBSA3tJQjNq77zqpWJNz3FS9TrX7H7SLKcom
👍7❤5🔥5💯1
🔥 Thought we were just “web3”? Nah, we’re gunning for the big dogs.
Cursor, Anthropic, OpenAI—watch your backs.
Stake $COM or swipe with $SOL / $USDC / fiat.
Either way, your bag stays safe, your mind gets blown. 🌋
https://x.com/comput3ai/status/1959233588185997365
Cursor, Anthropic, OpenAI—watch your backs.
Stake $COM or swipe with $SOL / $USDC / fiat.
Either way, your bag stays safe, your mind gets blown. 🌋
https://x.com/comput3ai/status/1959233588185997365
🔥8👍2💯2
This media is not supported in your browser
VIEW IN TELEGRAM
WE JUST DROPPED DEEPSEEK-V3.1! 🚀
It's currently the #3 open source coding model on Open Router after Qwen3-Coder:480b and Kimi-K2. We're now officially hosting all 3 and we're the only ones on B200s.
Today's the last day to vibe using any model for free, go vibe something!
It's currently the #3 open source coding model on Open Router after Qwen3-Coder:480b and Kimi-K2. We're now officially hosting all 3 and we're the only ones on B200s.
Today's the last day to vibe using any model for free, go vibe something!
🔥7👏1
This media is not supported in your browser
VIEW IN TELEGRAM
Someone in our discord did this, we swear it wasn't us (wait until it finishes). $COM 🌋
🔥7👏5❤2🥰1
🚀 $COM MEXC Airdrop is LIVE!
We’ve already showered our Raydium & Streamflow stakers with $COM, and now it’s MEXC’s turn! 💥
🔑 How to join
1️⃣ Open MEXC → Wallet → Deposit →
2️⃣ Tap On-Chain Deposit → Search COM3 (same $COM token, ticker unavailable on MEXC)
3️⃣ Screenshot the COM3 address
4️⃣ Send the screenshot to our bot: @C3MexcBot https://t.iss.one/C3MexcBot
✅ First 100 users = 10,000 $COM
💎 Hold ≥4 weeks to unlock future rewards & exclusive perks
⚡️ **Multiple MEXC addresses allowed**—submit each one via the bot to stack more!
No KYC, no purchase, just an MEXC deposit address.
Clock’s ticking—go grab your $COM screenshot now!
We’ve already showered our Raydium & Streamflow stakers with $COM, and now it’s MEXC’s turn! 💥
🔑 How to join
1️⃣ Open MEXC → Wallet → Deposit →
2️⃣ Tap On-Chain Deposit → Search COM3 (same $COM token, ticker unavailable on MEXC)
3️⃣ Screenshot the COM3 address
4️⃣ Send the screenshot to our bot: @C3MexcBot https://t.iss.one/C3MexcBot
✅ First 100 users = 10,000 $COM
💎 Hold ≥4 weeks to unlock future rewards & exclusive perks
⚡️ **Multiple MEXC addresses allowed**—submit each one via the bot to stack more!
No KYC, no purchase, just an MEXC deposit address.
Clock’s ticking—go grab your $COM screenshot now!
🔥6❤5👍4
👏 Thanks to everyone who participated 🌋
📢 Unfortunately, our current airdrop has concluded.
🚀 But don't worry! Subscribe to our channel to be the first to know about our next airdrop and exclusive opportunities:
👉 @comput3ai
📱 https://t.iss.one/comput3ai
💎 Stay tuned for more rewards and community events!
📢 Unfortunately, our current airdrop has concluded.
🚀 But don't worry! Subscribe to our channel to be the first to know about our next airdrop and exclusive opportunities:
👉 @comput3ai
📱 https://t.iss.one/comput3ai
💎 Stay tuned for more rewards and community events!
❤12🔥6🤔3👍2