About a year ago, Kurt found a secret stash of A100s somewhere and we thought, “hmm, wouldn’t it be great if we could get Fly machines to work with GPUs?” We’ve been working on the idea ever since. Right now we have new A100 40G PCIes and 80G SXMs (read - big bois) coming online every day. They live in ord
, iad
, ams
, sjc
, and syd
with more regions to come.
How Do I Use These?
We wanted working with GPUs to feel a lot like working with Fly machines, so that’s what we built - machines that are attached to GPUs. As an aside, Firecracker doesn’t play well with GPUs so we used Cloud Hypervisor, but the experience of working with these VMs is very similar.
To deploy an app on a GPU machine, you’d run something like this in flyctl (but don’t do it now, it won’t work until your org is GPU-blessed):
fly deploy --vm-gpu-kind a100-pcie-40gb --volume-initial-size 100
What Can I Use Them For?
Judging by our first batch of GPU users, a variety of things!
AI is, of course, popular - mostly inference with existing open-source models but also with custom models. We’re not set up for training large models, but I could imagine someone using us to fine-tune the last few layers of a model on their own data.
Some users have asked about rendering, and we’re into that. And of course some of you are exploring running your platform on top of Fly GPUs, which we love.
It’s still early, but you can check out some of the demos we’ve put together in our docs.
What’s the Cost?
Right now, the on-demand cost is $2.50/hr per GPU for the A100 40G PCIes and $3.50/hr per GPU for the A100 80G SXMs. There’s no minimums for usage. You decide the CPU, RAM, and storage you need. We have discounted pricing for reserved GPU machines and dedicated hosts.
How Do I Get One?
Everyone can use them today!
What are YOU interested in building with Fly GPUs? What other GPU-enabled shenanigans would you like to see Fly support? Leave us a note.