Darkbloom – Private inference on idle Macs
Posted by twapi 1 day ago
Comments
Comment by kennywinker 1 day ago
Comment by eigengajesh 1 day ago
That's why we don't recommend purchasing a new machine. Existing machine is no cost for you to run this.
Electricity is one cost, but it will get paid off from every request it receives. Electricity is only deducted when you run an inference. If you have any questions, DM me @gajesh on Twitter.
Comment by mbesto 1 day ago
You misunderstood. If the ROI is there, there is enough capital in existence for you to accelerate your profit. So why even deal with the complexity of renting people's hardware when you can do it yourself?
Comment by splintercell 12 hours ago
Comment by BuildTheRobots 21 hours ago
The calculator gives numbers for nearly everything, but I can't obviously see how much space it needs for model storage or how many writes of temp files I should expect if I'm running flat out.
Comment by MithrilTuxedo 10 hours ago
Comment by sleepybrett 19 hours ago
Comment by neurostimulant 8 hours ago
Comment by stavros 1 day ago
Comment by washadjeffmad 1 day ago
Out of our >3000 currently active Apple Silicon Macs, failures due to non-physical damage are in the single digits per year. Of those, none have been from production systems with 24/7 uptime and continuous high load, which reflects your parenthetical.
Perhaps we haven't met the other end of the bathtub curve yet, but we also won't be retaining any of these very far beyond their warranty period, much less the end of their support life.
Comment by bbatsell 21 hours ago
Comment by alsetmusic 21 hours ago
It’s three years for Macs, though I believe you can pay annually for longer. Five has never been a thing to my knowledge.
Comment by stavros 1 day ago
Comment by embedding-shape 1 day ago
How much though? Say I have three Mac Minis next to each other, one that is completely idle but on, one that bursts 100% CPU every 10 minutes and one that uses 100% CPU all the time, what's the difference on how long the machines survives? Months, years or decades?
Comment by LPisGood 17 hours ago
Comment by dmitrygr 19 hours ago
That is not at all how modern chips work. Idle chips are mostly powered down, non-idle ones are working and that causes real measurable wear and tear on the silicon. CPU, RAM, NAND all wear and tear measurably with use on current manufacturing processes.
Comment by Barbing 17 hours ago
Comment by ta988 17 hours ago
Comment by dmitrygr 17 hours ago
Comment by avidphantasm 1 day ago
Comment by xhkkffbf 23 hours ago
Comment by edbaskerville 18 hours ago
Comment by dgacmu 1 day ago
And then there's a hit for overprovisioning in general. If the network is not overprovisioned somewhat, customers won't be able to get requests handled when they want, and they'll flee. But the more overprovisioned it is, the worse it is for compute seller earnings.
I suspect an optimistic view of earnings from a platform like this would be something like 1/8 utilization on a model like Gemma 4. Their calculator estimates my m4 pro mini could earn about $24/month at 3 hours/day on that model. That seems plausible.
Comment by liuliu 21 hours ago
Comment by pgporada 20 hours ago
Comment by psychoslave 1 day ago
Assuming that getting large chunk of initial investment is just a formality is out of touch with 99% of people reality out there, when it’s actually the biggest friction point in any socio-economical endeavour.
Comment by chaoz_ 1 day ago
Non-VC play (not required until you can raise on your own terms!) and clear differentiation.
If you want to go full-business-evaluation, I would be more worried about someone else implementing same thing with more commission (imo 95% and first to market is good enough).
Comment by jonplackett 1 day ago
ie. Does anyone know the payback time for a B100 used just for inference? I assume it’s more than a couple of months? Or is it just training that costs so much?
Comment by Saline9515 1 day ago
Comment by dnnddidiej 1 day ago
Prolly gonna make $50 a year tops.
Comment by CTDOCodebases 23 hours ago
When YouTubers start making videos about it you know it's too late.
Comment by dnnddidiej 15 hours ago
Comment by liuliu 22 hours ago
A H200 gives you ~4 PFLOPs, which is ~60x at only ~40x price (assuming you can get a Mac Mini at $1000). (Not to mention, BTW, RTX PRO 6000 is ~7x price for ~40x more FLOPs).
Your M4 Mac Mini only has ~20 TFLOPs.
Comment by bentobean 21 hours ago
What a time to be alive.
Comment by znnajdla 1 day ago
As a business owner, I can think of multiple reasons why a decentralized network is better for me as a business than relying on a hyperscaler inference provider. 1. No dependency on a BigTech provider who can cut me off or change prices at any time. I’m willing to pay a premium for that. 2. I get a residential IP proxy network built-in. AI scrapers pay big money for that. 3. No censorship. 4. Lower latency if inference nodes are located close to me.
Comment by kennywinker 1 day ago
Comment by znnajdla 1 day ago
Comment by lxglv 1 day ago
Comment by znnajdla 1 day ago
Comment by aacid 1 day ago
Comment by znnajdla 1 day ago
Comment by rzwitserloot 1 day ago
Running AI inference increases the power draw, and requires certain hardware.
Mining bitcoin increases the power draw, and requires certain hardware.
OP's point thus stands: Bad players will find places to get far cheaper power than the intended audience, and will buy dedicated hardware, at which point the money you can earn to do this will soon drop below the costs for power (for folks like you and me).
Maybe that won't happen, but why won't that happen?
Comment by znnajdla 1 day ago
Comment by kennywinker 23 hours ago
Comment by NiloCK 1 day ago
You - Darkbloom - Operator - Darkbloom - you, vs
You - Provider - you
---
On the censorship point - this is an interesting risk surface for operators. If people are drawn my decentralized model provisioning for its lax censorship, I'm pretty sure they're using it to generate things that I don't want to be liable for.
If anything, I could imagine dumber and stricter brand-safety style censorship on operator machines.
Comment by znnajdla 1 day ago
Comment by yard2010 1 day ago
Comment by thih9 1 day ago
Others are reporting low demand, eg.: https://news.ycombinator.com/item?id=47789171
Comment by p1necone 17 hours ago
Comment by gleenn 1 day ago
Comment by kennywinker 1 day ago
Comment by ffsm8 1 day ago
Comment by runako 1 day ago
Very smart play to build a platform, get scale, and prove out the software. Then either add a small network fee (this could be on money movement on/off platform), add a higher tier of service for money, and/or just use the proof points to go get access to capital and become an operator in your own pool.
Comment by nxpnsv 1 day ago
Comment by runako 1 day ago
This is essentially the same reason even the best money managers take outside money to start, even if they eventually kick out the investors.
Comment by agnosticmantis 1 day ago
- Elon Musk during Tesla's Autonomy Day in April 2019.
Comment by foota 1 day ago
Comment by kennywinker 1 day ago
Comment by Filligree 1 day ago
Also they’ve already launched a crypto token, which is a terrible sign.
Comment by znpy 1 day ago
Comment by tgma 1 day ago
In 15 minutes of serving Gemma, I got precisely zero actual inference requests, and a bunch of health checks and two attestations.
At the moment they don't have enough sustained demand to justify the earning estimates.
Comment by splittydev 1 day ago
Comment by tgma 1 day ago
Still, absolute zero is an unacceptable number. Had this running for more than an hour.
Comment by splittydev 1 day ago
Sure, it would be great if you'd immediately get hammered with hundreds of requests and start make money quickly. It would also be great if it was a bit more transparent, and you could see more stats (what counts as "idle"? Is my machine currently eligible to serve models?). But it's still very new, I'd say give it some time and let's see how it goes.
If you have it running and you get zero requests, it uses close to zero power above what your computer uses anyway. It doesn't cost you anything to have it running, and if you get requests, you make money. Seems like an easy decision to me.
Comment by usrusr 1 day ago
Comment by tgma 1 day ago
Comment by yard2010 1 day ago
Comment by jagged-chisel 17 hours ago
Comment by hamiltont 1 hour ago
This appears on their credit purchase page right now, but you have to email them to get credits (everyone starts with zero)
Comment by subroutine 1 day ago
Comment by lxglv 1 day ago
Comment by lostmsu 1 day ago
The numbers are absolute fraud. You shouldn't be installing their software cause fraud could be not just about numbers.
Comment by rjmunro 1 day ago
Comment by mhast 1 day ago
Given their estimates of a Mac being able to generate $1k (per month?) a 5090 with a lot more power would be able to generate $50k. For a $3k piece of hardware. Which is obviously not realistic. (As in, nobody is paying that much for the images, which seems to match well with no actual requests on the system.)
Comment by iepathos 19 hours ago
Comment by thatxliner 1 day ago
I was thinking of building this exact thing a year ago but my main stopper was economics: it would never make sense for someone to use the API, thus nobody can make money off of zero demand.
I guess we just have to look at how Uber and Airbnb bootstrapped themselves. Another issue with my original idea was that it was for compute in general, when the main, best use-case, is long(er)-running software like AI training (but I guess inference is long running enough).
But there already exist software out there that lets you rent out your GPU so...
Comment by tgma 1 day ago
Comment by starkeeper 1 day ago
Comment by lostmsu 1 day ago
Comment by LPisGood 16 hours ago
Comment by tnchr 1 day ago
Comment by elbac 17 hours ago
WARN STT backend failed health check — model will NOT be advertised
Comment by gleenn 1 day ago
Comment by mirashii 1 day ago
That said, their privacy posture at the cornerstone of their claims is snake oil and has gaping holes in it, so I still wouldn't trust it, but it's worth being accurate about how exactly they're messing up.
Comment by mike_hearn 1 day ago
You are right - the "nonce binding" the paper uses doesn't seem convincing. The missing link is that Apple's attestation doesn't bind app generated keys to a designated requirement, which would be required to create a full remote attestation.
Comment by mirashii 1 day ago
It only effectively allows this for applications that are in the set of things covered by SIP, but not for any third-party application. There's nothing that will allow you to attest that arbitrary third-party code is running some specific version without being tampered with, you can only attest that the base OS/kernel have not been tampered with. In their specific case, they attempt to patch over that by taking the hash of the binary, but you can simply patch it before it starts.
To do this properly requires a TEE to be available to third-party code for attestation. That's not a thing on macOS today.
Comment by mike_hearn 1 day ago
public key from SEP -> designated requirement of owning app binary
The macOS KeyStore infrastructure does track this which is why I thought it'd work. But the paper doesn't mention being able to get this data server side anywhere. Instead there's this nonce hack.It's odd that the paper considers so many angles including things like RDMA over Thunderbolt, but not the binding between platform key and app key.
Reading the paper again carefully I get the feeling the author knows or believes something that isn't fully elaborated in the text. He recognizes that this linkage problem exists, proposes a solution and offers a security argument for it. I just can't understand the argument. It appears APNS plays a role (apple push notification service) and maybe this is where app binding happens but the author seems to assume a fluency in Apple infrastructure that I currently lack.
Comment by mirashii 1 day ago
Certainly, it still doesn't get you there with their current implementation, as the attempts at blocking the debugger like PT_DENY_ATTACH are runtime syscalls, so you've got a race window where you can attach still. Maybe it gets you there with hardened runtime? I'd have to think a bit harder on that.
Comment by mike_hearn 1 day ago
I'm not quite sure why Apple haven't enabled DCAppAttest on macOS. From my understanding of the architecture, they have every piece needed. It's possible that they just don't trust the Mac platform enough to sign off on assertions about it, because it's a lot more open so it's harder to defend. And perhaps they feel the reputational risk isn't worth it, as people would generalize from a break of App Attest on macOS to App Attest on iOS where the money is. Hard to say.
Comment by ramoz 1 day ago
Apple Silicon has a Secure Enclave, but not a public SGX/TDX/SEV-style enclave for arbitrary code, so these claims are about OS hardening, not verifiable confidential execution.
It would be nice if it were possible. There's a lot of cool innovations possible beyond privacy.
Comment by mike_hearn 1 day ago
macOS has a strong enough security architecture that something like Darkbloom would have at least some credibility if there was a way to remotely attest a Mac's boot sequence and TCC configuration combined with key-to-DR binding. The OS sandbox can keep apps properly separated if the kernel is correct and unhacked. And Apple's systems are full of mitigations and roadblocks to simple exploitation. Would it be as good as a consumer SGX enclave? Not architecturally, but the usability is higher.
Comment by znnajdla 1 day ago
Comment by ramoz 1 day ago
You have no guarantees over any random connected laptop connected across the world.
Comment by znnajdla 1 day ago
Comment by nl 1 day ago
The key question here is how they avoid the outside computer being able to view the memory of the internal process:
> An in-process inference design that embeds the in- ference engine directly in a hardened process, elimi- nating all inter-process communication channels that could be observed, with optional hypervisor mem- ory isolation that extends protection from software- enforced to hardware-enforced via ARM Stage 2 page tables at zero performance cost.[1]
I was under the impression this wasn't possible if you are using the GPU. I could be misled on this though.
[1] https://github.com/Layr-Labs/d-inference/blob/master/papers/...
Comment by nitros 1 day ago
Comment by flockonus 1 day ago
And more so in particular, anyone using Darkbloom with commercial intents should only really send non-sensitive data (no tokens, customer data, ...) I'd say only classification tasks, imagine generation, etc.
Comment by joelthelion 1 day ago
Comment by mike_hearn 1 day ago
Comment by ramoz 1 day ago
Macs have secure enclaves.
Comment by nl 1 day ago
But they argue that:
> PT_DENY_ATTACH (ptrace constant 31): Invoked at process startup before any sensitive data is loaded. Instructs the macOS kernel to permanently deny all ptracerequests against this process, including from root. This blocks lldb, dtrace, and Instruments.
> Hardened Runtime: The binary is code-signed with hardened runtime options and explicitly without the com.apple.security.get-task-allow entitlement. The kernel denies task_for_pid() and mach_vm_read()from any external process.
> System Integrity Protection (SIP): Enforces both of the above at the kernel level. With SIP enabled, root cannot circumvent Hardened Runtime protections, load unsigned kernel extensions, or modify protected sys- tem binaries. Section 5.1 proves that SIP, once verified, is immutable for the process lifetime.
gives them memory protection.
To me that is surprising.
Comment by mirashii 1 day ago
[1] https://github.com/Layr-Labs/d-inference/blob/master/papers/...
Comment by mirashii 1 day ago
Comment by dinobones 1 day ago
If it's not running fully end to end in some secure enclave, then it's always just a best effort thing. Good marketing though.
Comment by mike_hearn 1 day ago
Apple is perfectly capable of doing remote attestation properly. iOS has DCAppAttest which does everything needed. Unfortunately, it's never been brought to macOS, as far as I know. Maybe this MDM hack is a back door to get RA capabilities, if so it'd certainly be intriguing, but if not as far as I know there's no way to get a Mac to cough up a cryptographic assertion that it's running a genuine macOS kernel/boot firmware/disk image/kernel args, etc.
It's a pity because there's a lot of unique and interesting apps that'd become possible if Apple did this. Darkbloom is just one example of what's possible. It'd be a huge boon to decentralization efforts if Apple activated this, and all the pipework is laid already so it's really a pity they don't go the extra mile here.
Comment by woadwarrior01 1 day ago
Apple's docs claim it's been available on macOS since macOS 11. Am I missing something here?
https://developer.apple.com/documentation/devicecheck/dcappa...
Comment by mike_hearn 23 hours ago
https://developer.apple.com/documentation/devicecheck/dcappa...
> If you read isSupported from an app running on a Mac device, the value is false. This includes Mac Catalyst apps, and iOS or iPadOS apps running on Apple silicon.
Comment by woadwarrior01 23 hours ago
Comment by saagarjha 1 day ago
Comment by vrockdub 1 day ago
Comment by jeroenhd 1 day ago
The biggest argument for remote attestation I can think of is to make sure nobody is returning random bullshit and cashing in prompt money on a massive scale.
Comment by 190n 19 hours ago
All you have to do is attach to the process before it does that, and then prevent this call from going through.
Comment by saagarjha 1 day ago
Comment by ramoz 1 day ago
Protection here is conditional, best-effort. There are no true guarantees, nor actual verifiability.
Comment by pants2 1 day ago
My M5 Pro can generate 130 tok/s (4 streams) on Gemma 4 26B. Darkbloom's pricing is $0.20 per Mtok output.
That's about $2.24/day or $67/mo revenue if it's fully utilized 24/7.
Now assuming 50W sustained load, that's about 36 kWh/mo, at ~$.25/kWh approx. $9/mo in costs.
Could be good for lunch money every once in a while! Around $700/yr.
Comment by mavamaarten 1 day ago
I'd say it's not worth it. But the idea is cool.
Comment by jorvi 1 day ago
Thermal stress from bursty workloads is much more of a wearing problem than electromigration. If you can consistently keep the SoC at a specific temperature, it'll last much longer.
This is also why it was very ironic that crypto miner GPUs would get sold at massive discounts. Everyone assumed that they had been ran ragged, but a proper miner would have undervolted the card and ran it at consistent utilization, meaning the card would be in better condition than a secondhand gamer GPU that would have constantly been shifting between 1% to 80% utilization, or rather, 30°C to 75°C
Comment by kennywinker 1 day ago
Comment by kennywinker 1 day ago
For Gemma 4 26B their math is:
single_tok/s = (307 GB/s / 4 GB) * 0.60 = 46.0 tok/s
batched_tok/s = 46.0 * 10 * 0.9 = 414.4 tok/s
tok/hr = 414.4 * 3600 = 1,492,020
revenue/hr = (1,492,020 / 1M) * $0.200000 = $0.2984
I have no idea if that is a good estimate of how much an M5 Pro can generate - but that’s what it says on their site.
They do a bit of a sneaky thing with power calculation: they subtract 12Ws of idle power, because they are assuming your machine is idling 24/7, so the only cost is the extra 18W they estimate you’ll use doing inference. Idk about you, but i do turn my machine off when i am not using it.
Comment by pants2 1 day ago
Comment by nnx 1 day ago
This seems high. At which quantization? Using LM Studio or something else?
Note: Darkbloom seems to run everything on Q8 MLX.
Comment by pants2 1 day ago
Comment by torginus 1 day ago
I figured since I already used it a lot, and I've never had a GPU fail on me, it would be fine.
The fans on it died in a month of constant use, replacing them was more money than what I made on mining.
Comment by todotask2 1 day ago
I don’t think this is a sustainable business model. For example, Cubbit tried to build decentralised storage, but I backed out because better alternatives now exist, and hardware continues to improve and become cheaper over time.
Your electricity and ownership are going to get lower return and does not actually requce CO2.
Comment by chaoz_ 1 day ago
I’d imagine 1 year of heavy usage would somehow affect its quality.
Comment by pants2 1 day ago
Comment by xendo 1 day ago
Comment by znnajdla 1 day ago
Comment by sethherr 1 day ago
Comment by tonyedgecombe 1 day ago
Comment by znnajdla 1 day ago
Comment by dgacmu 1 day ago
Oh, also, you seem to have some bugs:
Gemma: WARN [vllm_mlx] RuntimeError: Failed to load the default metallib. This library is using language version 4.0 which is not supported on this OS. library not found library not found library not found
cohere: 2026-04-16T14:25:10.541562Z WARN [stt] File "/Users/dga/.darkbloom/bin/stt_server.py", line 332, in load_model 2026-04-16T14:25:10.541614Z WARN [stt] from mlx_audio.stt.models.cohere_asr import audio as audio_mod 2026-04-16T14:25:10.541643Z WARN [stt] ModuleNotFoundError: No module named 'mlx_audio.stt.models.cohere_asr'
Trying to download the flux image models fails with:
curl: (56) The requested URL returned error: 404
darkbloom earnings does not work
your documentation is inconstent between saying 100% of revenue to providers vs 95%
I think .. this needs a little more care and feeding before you open it up widely. :) And maybe lay off the LLM generated text before it gets you in trouble for promising things you're not delivering.
Comment by canarias_mate 20 hours ago
Comment by haspok 1 day ago
Problem is, from a technical point of view, what kind of made sense back then (most people running desktops, fans always on, energy saving minimal) is kind of stupid today (even if your laptop has no fan, would you want it to be always generating heat?)...
I definitely want my laptops to be cool, quiet and idle most of the time.
Comment by kamranjon 1 day ago
Comment by vorticalbox 1 day ago
This is short bursts of heat 5-10 m during the render I would not be happy with that for multiple hours a day. I am sure that would have a negative effect on battery health.
Comment by dchuk 23 hours ago
I wish this was self hostable, even for a license fee. Many businesses have fleets of Macs, sometimes even in stock as returned equipment from employees. Would allow for a distributed internal inference network, which has appeal for many orgs who value or require privacy.
Comment by TuringNYC 1 day ago
Comment by damezumari 1 day ago
Comment by zozbot234 1 day ago
Comment by utopiah 1 day ago
Comment by TuringNYC 23 hours ago
Comment by utopiah 7 hours ago
Comment by frankfrank13 22 hours ago
- Convincing labs to run distributed, burst-y inference
- Convincing people to run their Mac all day, hoping to make a little profit
- Convincing users to trust a distributed network of un-trusted devices
I had a similar idea, pre-AI, just for compute in general. But solving even 1 of those 3 (swap AI lab for managed-compute-type-company, eg Supabase, Vercel) is nearly impossible.
Comment by MyUltiDev 20 hours ago
Comment by pants2 1 day ago
Comment by rvz 1 day ago
Comment by kennywinker 1 day ago
;P
Comment by BingBingBap 1 day ago
What could possibly go wrong?
Comment by MicBook56 1 day ago
Comment by jzig 21 hours ago
Comment by creamyhorror 21 hours ago
Comment by NiloCK 1 day ago
Right now the dashboards show 78 providers online, but someone in-thread here said that they spun one up and got no requests. Surely someone would be willing to beat the posted rate and swallow up the demand?
I expect this is a migration target, but a tactical omission from V1 comms both for legitimate legibility reasons (I can sell x for y is easier to parse than 'I can participate in a marketplace') and slightly illegitimate legibility reasons (obscuring likely future price collapse).
Still - neat project that I hope does well.
[1] Layer Labs, formerly EigenLayer, is company built around a protocol to abstract and recycle economic security guarantees from Ethereum proof of stake.
Comment by Inferlane 1 day ago
Comment by woadwarrior01 1 day ago
Got the latest v0.3.8 version from the list here: https://api.darkbloom.dev/v1/releases/latest
Three binaries and a Python file: darkbloom (Rust)
eigeninference-enclave (Swift)
ffmpeg (from Homebrew, lol)
stt_server.py (a simple FastAPI speech-to-text server using mlx_audio).
The good parts: All three binaries are signed with a valid Apple Developer ID and have Hardened runtime enabled.
Bad parts: Binaries aren't notarized. Enrolls the device for remote MDM using micromdm. Downloads and installs a complete Python runtime from Cloudflare R2 (Supply chain risk). PT_DENY_ATTACH to make debugging harder. Collects device serial numbers.
TL;DR: No, not touching that.
Comment by amdivia 1 day ago
Comment by jdironman 16 hours ago
Comment by poorman 20 hours ago
Comment by qurren 20 hours ago
Too expensive. It's probably producing 200 watts average for 8 hours a day. That's 1600 watt hours, which is about $1.60 at PG&E prices. That would take 187 days to recoup the cost of just the panel.
If you include installation costs and "what PG&E steals if you wire it to the same grid" it's probably more like 4x that, which is too long.
Tell me when we can have 400 watt solar panels for $50. Stupid capitalism literally forces solar panel prices to make it unprofitable.
People should never have to take out loans for solar. Solar should be subsidized and forced by the government to be so cheap that it repays for its cost within a month. Then we're talking. Most things I buy to save money, I expect them to repay within a month. Maybe 2 months max.
Comment by dgacmu 19 hours ago
Installation costs and inverters not included, however.
Comment by poorman 20 hours ago
Comment by qurren 20 hours ago
Just pointing out why capitalism + solar is a failure. Capitalism reprices the good thing to be equally expensive to the bad thing, so that nobody buys the good thing anymore.
Comment by icedrift 17 hours ago
Comment by 0xbadcafebee 1 day ago
Comment by dr_kiszonka 1 day ago
When your Mac is idle (no inference requests), it consumes minimal power — you don't lose significant money waiting for requests. The electricity costs shown only apply during active inference.
Text models typically see the highest and most consistent demand. Image generation and transcription requests are bursty — high volume during peaks, quiet otherwise."
Comment by alexpotato 1 day ago
I believe the idea was that people could submit big workloads, the server would slice them up and then have the clients download and run a small slice. You as the computer owner would then get some payout.
Intersting to see this coming back again.
Comment by willquack 16 hours ago
I added Python execution support via Pyodide (cpython compiled to wasm) and worked on a bunch of other random stuff like WebLLM inferencing during my time there.
Apart from Distributive, there's also the "Golem network", "Salad", "Koii" and various other similar projects.
---
I'm not sure if I'm convinced by the "Uber for compute" use case with compute buyers and compute workers (sellers), but if you're a university and you have 1000 Windows machines across all your computer labs, it'd be nice to leverage that compute for running research or something idk - especially with the price of ram / cloud offerings these days...
Comment by thekid314 1 day ago
Comment by Jn2G3Np8 1 day ago
But trying it out it still needs work, I couldn't download a model successfully (and their list of nodes at https://console.darkbloom.dev/providers suggests this is typical).
And as a cursory user, it took me some digging to find out that to cash out you need a Solana address (providers > earnings).
Comment by dkroy 19 hours ago
Comment by poorman 19 hours ago
Comment by heddycrow 1 day ago
We’ve been building something similar for image/video models for the past few months, and it’s made me think distribution might be the real bottleneck.
It’s proving difficult to get enough early usage to reach the point where the system becomes more interesting on its own.
Curious how others have approached that bootstrap problem. Thanks in advance.
Comment by jaffee 1 day ago
Comment by utkarsh_apoorva 1 day ago
They lost me with just one microcopy - “start earning”. Huge red signal.
Comment by Hamuko 1 day ago
Comment by miki123211 1 day ago
Is there some actual cryptography behind this, or just fundamentally-breakable DRM and vibes?
Comment by auslegung 22 hours ago
Comment by WatchDog 1 day ago
Available models (2):
CohereLabs/cohere-transcribe-03-2026 (4.6 GB)
flux_2_klein_9b_q8p.ckpt (20.2 GB)
...
Advertising 0 model(s) (only loaded models)
Also the benchmark just doesn't work.Interesting idea, but needs some work.
Comment by eigengajesh 1 day ago
Comment by daniel_iversen 23 hours ago
Comment by 0xc133 1 day ago
Comment by SlavikCA 23 hours ago
Comment by gndp 1 day ago
Comment by podviaznikov 1 day ago
on 15.1 it failed to serve models.
updated to latest 15.5 and it fails to run binary.
Comment by Schiendelman 1 day ago
Comment by jboggan 1 day ago
Comment by drob518 1 day ago
Comment by v9v 1 day ago
Comment by smooth968 23 hours ago
Comment by AustinDev 17 hours ago
Comment by puttycat 1 day ago
Afaik you will need to decrypt the data the moment it needs to be fed into the model.
How do they do this then?
Comment by mr_mitm 1 day ago
Remember, all encryption is E2EE if you're not picky about the ends.
Comment by czk 21 hours ago
Comment by ponyous 1 day ago
Comment by dangoodmanUT 1 day ago
Comment by ripped_britches 1 day ago
Comment by koliber 1 day ago
Comment by chaoz_ 1 day ago
Guess there are limitations on size of the models, but if top-tier models will getting democratized I don’t see a reason not to use this API. The only thing that comes to me is data privacy concerns.
I think batch-evals for non-sensitive data has great PMF here.
Comment by 59nadir 1 day ago
Heh, what did they win exactly? This is just a way for another company to extract value out of the single region of the world where Apple is a relevant vendor, and it happens to be the one where it's the easiest to pull people into schemes.
Comment by rvz 1 day ago
Because they were already at the finish line with Apple Silicon.
> I don’t see a reason not to use this API. The only thing that comes to me is data privacy concerns.
The whole inference is end-to-end encrypted so none of the nodes can see the prompts or the messages.
Comment by chaoz_ 1 day ago
That would finally be a crypto thing which is backed by value I believe in.
Comment by resonanormal 1 day ago
Comment by Havoc 1 day ago
Comment by throwatdem12311 1 day ago
Comment by bentt 1 day ago
Comment by DeathArrow 1 day ago
Comment by btown 1 day ago
> Apple’s attestation servers will only generate the FreshnessCode for a genuine device that checks in via APNs. A software-only adversary cannot forge the MDA certificate chain (Assumption 3). Com- bined with SIP enforcement (preventing binary replace- ment) and Secure Boot (preventing bootloader tampering), this provides strong evidence that the signing key resides in genuine Apple hardware.
Comment by saagarjha 1 day ago
Comment by btown 1 day ago
Comment by saagarjha 1 hour ago
Comment by nl 1 day ago
NVidia data center GPUs have a similar path, but not their consumer ones. Not sure about the NVidia Spark.
It's possible AMD Strix Halo can do this, but unlikely for any other PC based GPU environments.
Comment by MrDrMcCoy 1 day ago
Comment by stryakr 1 day ago
Comment by subpixel 1 day ago
Comment by ianpurton 1 day ago
Comment by chakintosh 1 day ago
Comment by grvbck 1 day ago
Macbook Air M2 8GB 12h/day -> $647/month
Mac Mini M4 32GB 12h/day -> $290/month
I mean, I'd be happy to buy a few used M2 Airs with minimal specs and start printing money but…Comment by bprasanna 1 day ago
Comment by autodidacticon 18 hours ago
Comment by egorfine 1 day ago
Comment by dcreater 1 day ago
Comment by logicallee 23 hours ago
Comment by Fokamul 1 day ago
Comment by rvz 1 day ago
Away this looks like a great idea and might have a chance at solving the economic issue with running nodes for cheap inference and getting paid for it.
Comment by sharts 1 day ago
Comment by jaylane 1 day ago
Comment by jonhohle 1 day ago
I cringe every time I see this sentence structure. I know the joke is about emdashes, but the “Its not …. It’s ….” drives me crazy.
Comment by sergiusignacius 1 day ago
Comment by rustyhancock 23 hours ago
Comment by parasubvert 1 day ago
Comment by projektfu 1 day ago
Tired: That is not a technology problem. It is a marketplace problem.
Wired: This is a marketplace problem, not a technology problem.
Comment by rustyhancock 23 hours ago
And I don't really know why both patterns don't appear equally.
Intuitively I just can't grasp how it can be so selective with the phrases it generates.
Comment by nnevatie 1 day ago
Comment by biztos 1 day ago
But the real solution is to do this other thing. If you'd like I can give you the three-step guide to fixing the thing you asked me to fix.
/s
Comment by jiusanzhou 1 day ago
Comment by 0xelpabl0 1 day ago
Comment by eddie-wang 1 day ago
Comment by bojangleslover 1 day ago
Comment by smooth968 23 hours ago
Comment by jstlykdat 1 day ago