HACKER Q&A
📣 soheil

Best way to spend $25k in GCP


I have $25k remaining credit from Google Cloud Platforms, which expires in about a month. I got the credit over 2 years ago and had hundreds of severs for various websites, database storage, elasticsearch instances, dedicated servers for random cron tasks, etc. But I still haven't been able to exhaust the credit. I even used it to train several models using multiple Tesla GPUs. But beyond a point I felt like I didn't know what I was doing with the models and stopped wasting GPU cycles on them.

I feel like if I don't use it for something useful it'll be wasted. Any ideas?


  👤 sillysaurusx Accepted Answer ✓
Dear soheil,

I happen to need around few thousand in GCP credits in order to egress around 30TB of data from our cloud buckets. It's training data, training logs, model snapshots, and pretty much all of our ML history (http://wiki.tensorfork.com/).

The real tally is actually 83TB, but around 30TB of that is incremental model snapshots which we don't need to download (turns out, you can generate a lot of data when you're using TPUs for training), and around ~30TB is whatever-data that no one cares about / duplicated data across regions. (To use TPUs in EU and US, you need to place the training data in both EU and US.)

Normally I wouldn't ask, but we've been playing "hot potato" with this data for several months now, since our last credits ran out. We've basically been making new GCP accounts (where you get $300 credit) and then letting it sit there for a month (because they give you a month before deleting it, even if the credits run out).

If you'd be willing to help, it would be quite simple: I would transfer the buckets to your account, and then download the data to a Hetzner server. I'd then make the data publicly available, such as it is. You can DM me on twitter (https://twitter.com/theshawwn) or email me (see profile). Thanks for considering!

EDIT: For an example of how we've amassed terabytes of training logs, you can see some of our BigGAN outputs here: https://twitter.com/search?q=from%3Atheshawwn%20biggan&src=t...



👤 option_greek
It really shouldn't, but stories like this irks me a lot because I have been trying to get credits for my bootstrapped startup for sometime and there is just no way to get them unless you are part of a accelerator or took funding from well known investor.

Of course, Gcp, Aws, DO don't owe me anything and these programs must be successful based on their data but still shows the irony of the world where resources aren't matched well against needs.


👤 aborsy
Don’t use it if not needed. On the contrary, it contributes to the environmental waste.

👤 Jugurtha
You mention training models... Do you want to donate them to ML researchers or students?

Maybe people can use these on our platform (https://iko.ai) which is on GCP and GKE. The platform provides real-time collaborative notebooks so people can train, track, package, deploy, and monitor machine learning models on the users' Kubernetes clusters (AWS, GCP, Azure, Digital Ocean, etc. So if you have credits elsewhere, this may be doable).

You could transfer your credits to us, we'll generate a special link, then you donate that link to people or entities who, when signing up with the link, could use your resources.


👤 Cilvic
How about seeing how big of an index of the web one can build for 25k?

I am inspired by this: https://news.ycombinator.com/item?id=28550764

I was thinking about building a domain name index, this would give the project a nice deadline and budget :-)

If you want to collaborate, my email is in my profile.


👤 AtlasBarfed
Can't you just buy some really long term upfront storage contract?

You'll always need storage.


👤 aynyc
How did you manage to get that much? I’ve been trying with no avail to get some decent credit to run my side project. I’m playing the whack-a-mole game to signing up new account every time my $300 runs out.

👤 tester34
How about running heavy computations for somebody else for some $$?

"ROI"


👤 unixhero
Maybe you can mine some Ethereum. Not sure if that breaks their TOS, you'd have to check to see that this is allowed.

👤 cprecioso
Setup a cloud gaming rig and max it out :D