HACKER Q&A
📣 herodoturtle

Real-world anecdotes of MySQL at scale?


What is the biggest MySQL database you've every worked with?

Any interesting stories to tell?

I'm trying to get a sense of what MySQL is like at massive scale.

Thanks ^_^


  👤 unoti Accepted Answer ✓
I used to work at Second Life. Their player inventory and monetary transaction system was in MySQL. Every single item a player had or made was a record in this database. If you’ve never spent time in SL then the scope of this is probably a lot bigger than you imagine. As an active creator in SL I had 10s of thousands of records in there myself for inventory, and lots more activity for financial transactions and history. The database was distributed into somewhere between 30 and 50 servers, each of which had a read only replica. User inventory was divided amongst these shards. Any transactions that could execute on the read only replica did, and transactions that needed to be placed on the main server for a user executed there.

There were always concerns about the latency involved in the replication. Whenever I did large scale batch operations that needed to write lots of records in master servers, I would monitor the replication queue length and suspend operations when it got too far behind to keep things sane and healthy.

The biggest failure in this architecture was sharing sql queries as opposed to sharing service calls. This is something that was being fixed as I left. Switching it to work with service calls gives more flexibility in the underlying implementation and opens more options for improving the system.

For most kinds of large systems I’d recommend thinking about the architecture in terms of distributed Actors hosted in something like DAPR, Orleans, Akka, or Erlang/OTP and then the technical details of the underlying database become less of a constraint. It grants you flexibility for your underlying storage that you would not otherwise have.

Fun side fact: Bos, the CTO of meta, was a key person working on making all the magic happen there back then.


👤 jmalicki
Facebook is backed by MySQL https://engineering.fb.com/2021/07/22/data-infrastructure/my...

(though it's only used as the backing data store for their graph database https://www.usenix.org/system/files/conference/atc13/atc13-b...)


👤 hbrn
Assuming "at scale" means you have no other option than horizontal scaling.

At this point differences between databases start to shrink. Most of the complexity and features are being moved to the application layer, and your database is becoming dumber.

More subtle (and frankly boring) features are becoming important. Like how cost-effective is it to run, observe, backup/recover, etc.

Honestly, I feel like most interesting stories are before "at scale". Stories about how you delayed the need to "scale" for several years (and probably saved the company by allowing to focus on product instead of infra). But unfortunately those stories aren't something people tend to brag about. Nobody writes stories about how they chose to be pragmatic and not scale because they could just buy a bigger server or add a caching layer.


👤 ransom1538
I operate around $300k monthly mysql infra. We service most schools in the US with a netflix type service. We have burst traffic (at noon EST we x50 traffic). Here is what we learned:

1) RDS/Cloudsql is awesome but too expensive, we can't afford it. The things it provides are not that special with VMs anyway (images, backups, replication).

2) The industry moved to vitess db, but converting to this is too daunting, we are focused on automation and cost reduction instead.

3) Use index based sharding (no math shards % total, no consistent hashing, no drama). When you get a new user, setup where they should be on each shard and keep that on a single table [user_id,video_cluster_id,session_cluster_id,..]. This way moves are easy. Then video_cluster_id is a load balanced host infront of a fleet of replicas.

4) The hardest thing you will run into is write saturation. You can't write fast enough to a disk. Advanced topologies and 3) helps here. Example: [MasterA] <- [Slaves,0..n] <- [MasterB] <- slaveb,0..k]. This way you can push more writes to MasterB, and add replication filters so that slaveb doesn't need to perform all the writes MasterA needs.

5) SSDs. Don't bother with anything else.

6) We would pay a million dollars for this tool: https://docs.percona.com/percona-toolkit/pt-online-schema-ch...

7) GTID took away the hard parts of mysql replication.

8) Hardcore mysql experience is disappearing. With RDS/Cloudsql/Postgres less and less developers are going able to patch a production problem.


👤 itake
I was in a system design interview where the interviewer didn't like that I chose MySQL for storage, because it doesn't scale. I pointed that I literally used to work at a b2c app in this exact problem space that exclusively used MySQL storage. I haven't heard back yet from the company and its been 2 weeks :-/.

👤 natefox
Pinterest uses (used?) mysql. Blog post is a bit old but a fun read https://medium.com/pinterest-engineering/sharding-pinterest-...

👤 evanelias
One thing to keep in mind when looking at older MySQL scale-out presentations and war-stories: they're mostly MySQL 5.0-5.6, and a chunk of them are bare metal rather than cloud. The problems you'll encounter today -- scaling out in a modern cloud environment on MySQL 8 -- may not line up too well with what folks hit in the past. Ditto for best practices on architecture and ideal max db instance size.

I gave a bunch of MySQL conference talks from 2011 through 2017, including several focused on scaling and sharding at social networks. Looking back through them today, a lot of content from the scaling/sharding ones honestly isn't too useful if you're trying to scale up a new MySQL deployment right now.


👤 manimino
Are you referring to distributed MySQL such as Vitess? It is the backend for Slack and GitHub; also was the backend for YouTube in the past.

https://vitess.io/


👤 wbeckler
Google "mysql booking.com" for a beast of an example. There's lots of content out there about that one.

👤 alberth
There always seems to be countless examples of companies using MySQL at extremely large scale.

What are some Postgres examples?

As a Postgres fan, I'm genuinely curious. Please don't take my comments as troll-y.


👤 ksec
It is very rare we get the subject of MySQL on HN Front-page considering HN have been pretty much Anti-MySQL, so I will take this opportunity to ask;

Does anyone know if MySQL 9 is on the roadmap? Or are we going to get mostly small iteration and bug fix release only?


👤 sambostock
Shopify Engineering has blog posts you may find interesting:

https://www.google.com/search?q=mysql+site%3Ashopify.enginee...


👤 akgerber
Etsy, via sharding & a custom ORM — the 'Flickr architecture'.

In 2021: "We saw 1.8M queries per second on our massively sharded MySQL databases." https://twitter.com/codeascraft/status/1466174452459196420

Here's a post from a decade ago: https://www.etsy.com/codeascraft/two-sides-for-salvation


👤 kodah
Vitess is by far the biggest I've seen in practice.

Sharding is at the root of storage management, which let's users design a strategy that fits their storage and request benchmarks. The biggest difference that I've seen is that you no longer get to use many of the MySQL operations and functions that you may be used to like relationships. It's analogous to the best and slightly painful parts of MySQL and NoSQL. Things like relationships get relegated to the application.


👤 jmyeet
There are any number of presentations on Facebook operating MySQL at scale (eg [1][2]). I don't know how many servers they have but half a million wouldn't surprise me.

[1]: https://www.youtube.com/watch?v=NfS5ZLNPxS4

[2]: https://www.youtube.com/watch?v=kP6undC_HDE


👤 palashkulsh
We have around 35tb MySQL deployment in master slave config. The most interesting nightmare happened to us when doing the master fail over using ghost and panic manual intervention created 2 active master scenario leading to chaos which we had to clean for days along with monetary loss : ) Lesson learnt don't do manual intervention when ghost is doing fail over

👤 exabrial
Really the only thing we run into trouble with is `count(id)` where `id` is the pk of the table. This is surprisingly slow when you get above a couple hundred million rows. We're 3rd normal form, with proper fk's and indexes.

I _highly_ recommend having someone or some team in charge of db architecture. Use the "remarks" column to make a data dictionary that's stored right in the table. Have an _exact_ specication for table names (plural nouns), column names (agent_id (always a fk), agent_name), index names (udx-agents-agent_name), fk names (fk-leads->agents).

But, here in lies the rub: ALWAYS implement today's requirements, not next-years. If your app can be powered by a single mysql instance, _just do it_. You'll be in a much better position to scale later that if you pick some excessive technology you _think_ you might need in the future.


👤 mscccc

👤 20after4
Wikipedia uses MySQL and although I worked there for 7 years I never did much work directly with the DBs.

It's been 9 years since I worked at deviantart.com but last I knew the site still operated on a large sharded MySQL cluster divided into something like 32 individual shards where each user's data was on a shard determined by their userid.

Essentially just `shard = userID % shardCount`;

The application would connect to all of the servers and for many queries you have to send the query to all 32 servers, or some subset based on the records you seek, and then merge the results of all of those queries.


👤 karmakaze
The largest databases I've worked with were sharded, each with replicas and failover of writers. With 100s of shards each database is a smaller database and maybe doesn't even count as the largest I've worked with.

Large but smaller ones were federated (divided by tables rather than rows) also with group replication and replicas. The most interesting case was switching over a MySQL 5.6 InnoDB to a 5.7 TokuDB storage engine. Bringing up a new replica used to take hours to catch up depending on how old a snapshot you were starting with. Using TokuDB it took 10s of minutes, well under an hour. TokuDB isn't quite as effective as LSMTrees but MyRocks/RocksDB wasn't (isn't?) ready for 5.7 and last I checked had some stricter requirements.

The best think about MySQL is how (relatively) easy it is to set up replication with multiple-writers, and even load-balance/failover behind a TCP proxy like keepalived. Backups are pretty easy to script, stop replication on an instance, stop it, copy files, resume (and let it catch up replication, then re-add to query pool).

When you scale vertically to 512GB memory servers using NUMA memory, things get tricky. Had to tweak the BIOSes to get the same spec'd machines to behave mostly the same w.r.t. thread utilization and settle on a memory buffer partitioning scheme that minimized lock waits. Also split off slow queries to 'reporting' read-replica instance(s).

Nowadays folks would probably prefer managed services like AWS Aurora.


👤 OliverJones
This answer to your question is more about scaling out (many machines) than scaling up (big machines).

Many budget hosting providers supporting CMSs like WordPress and Joomla run very large multitenant MySQL or MariaDB instances. These instances often get into performance trouble when one or more of the customer sites on them grow large or get traffic spikes. The CMSs offer elaborate caching strategies (based on redis, memcache/d) to help mitigate this. But of course the multitenant nature of these instances means that individual customers must configure the caching. And the lowest-end hosting providers don't offer redis or memcache/d.

It's possible to monitor various operational parameters (via `SHOW GLOBAL STATUS` and similar SQL statements) to detect problems. It's very hard for customers to mitigate configuration problems (like insufficient buffer pool sizes or too few maximum connections). But customers can change table indexes.

Other hosting providers operate large numbers of small separate MariaDB and MySQL instances, one per customer. Considering that WordPress powers something like 30 - 40% of the sites on the internet, the total number of instances is vast.

WordPress.org, a vast site, uses multiple read replicas.

Version upgrade inertia is a big problem for hosting providers. One provider (whom I will not name) runs a MySQL version that reached end-of-life well over a year ago. New-start MySQL / MariaDB project designs should make provisions for DBMS version updates. Updates are worth applying because the development teams do lots of performance-enhancement work.


👤 biermic
I have one running with a ~5TB big unpartitioned table. Adding ~9M rows per day, with significant read load via 2 replicas. Works just fine on a 32 GB memory RDS instance. There is another table half the amount of rows and I join the two 1000s of times a day.

Took a bit of fiddling around to have things run stable. Currently migrating to Redshift, which has stellar performance but comes with its own problems.


👤 taldo
A more down-to-earth example: Evernote is (almost entirely) on MySQL. Since most of the stuff is "yours", it's fairly easy to shard partitioning by user. And the global stuff which is not that much (users/auth/etc) is stored in beefy replicated instances.

Newer features are being built on other data stores, and some things are being migrated away, but the core stuff is still in MySQL.


👤 efficax
https://github.blog/2018-06-20-mysql-high-availability-at-gi...

depends what you mean by massive scale, but mysql can certainly handle a website of say, stackoverflows size, on beefy hardware with read replicas. facebook would be another story entirely


👤 aq9
You need to qualify massive scale, it can be on multiple dimensions: * "Large" data (these days, 10s of TB of data, maybe with multi-TB individual tables, billions of rows in the large tables) * "High" QPS - reads * "High" QPS - writes

High QPS is in the eye of the beholder, but let's say something like a few hundred thousand QPS and up.

If you have scale in just one of these 3 dimensions, the solutions are relatively easy (except maybe for high write QPS, depends on the nature of the writes). If you are at scale on all three dimensions, solutions get harder (sharding, non-innodb storage engines).

Of course, there is also the wildcard: the nature of your queries/transactions. Large transactions along with scale will make your life a living hell.



👤 zactato
When I left Fitbit we were still using MySQL as the primary datastore for all tracker data. Admittedly only in the 10s of millions of users, but tracker data was stored at sub second precision and indefinitely. It was definitely up in the 10s or 100s of TB. Unlike most DBs it was more write heavy than read heavy.

Our scale has handled by sharding based on userId and geo. We could dynamically shift the shards when scaling the cluster up and down. I wish we had open sourced that sharding layer, it was pretty nifty.


👤 brutuscat
I remember coming back to "Optimised Pagination using MySQL"[1] to avoid common pagination "slowness".

[1]: https://www.xarg.org/2011/10/optimized-pagination-using-mysq...

[more]: https://stackoverflow.com/a/32360867


👤 kingnothing
Uber uses an in-house db called Schemaless which is built on top of MySQL for many of their largest datasets (many petabytes each).

https://www.uber.com/blog/schemaless-part-one-mysql-datastor...


👤 Thaxll
EA uses MySQL for a lot of things, largest game like FIFA / FUT, they have thousands of servers.

👤 vitalya
You can read about one of major Slack outages here https://slack.engineering/slacks-incident-on-2-22-22/. MySQL included.

👤 xnx
Google AdWords was run on MySQL for a long time: https://news.ycombinator.com/item?id=7645703

👤 ksec
Look at Youtube, or PlanetScale.

👤 aprdm
Shopify, during the interview , pretty crazy stuff

👤 t3rabytes
A ton of pain without a DBA team dedicated to it.