1

Over the next 2-3 years we will be scaling up data collection for a project. As a result the amount of data will grow 10-fold. Our current MySQL installation can keep up with the 2Tb of data but for larger queries there is a fair amount of IOWait.

Im investigating a migration to a clustered solution to spread out the IO but am wondering about NDB and what happens to data that doesn't get accessed very often. The impression I get from reading about MySQL cluster is that it relies on memory tables for most of the data. What happens with tables that don't get accessed very often (or at all)? And how does backup work? Can I use MYSQLDUMP or is there a better solution?

Stefan Lasiewski
  • 23,667
  • 41
  • 132
  • 186
ethrbunny
  • 2,369
  • 4
  • 41
  • 75

1 Answers1

2

NDB has some limitations for large tables due to it's in-memory model. For you situation, is does not make sense.

Recently, we tested Percona Cluster with some very promising results. It supports master-master replication, and has full InnoDB ACID compliance. One thing to note is the speed of queries is limited to the slowest node in the cluster - which many MySQL installations that are Master/Slave usually have a much faster Master.

If you need really fast access, you could also convert the really large data sets to MongoDB or Cassandra. Both of these are considerably faster than RDBMS like MySQL. Clustering in these NoSQL databases is native.

gmck
  • 132
  • 1