Skip to content

Latest commit

 

History

History
289 lines (238 loc) · 11 KB

goodbye-mysql-sharding-use-scale-out-mysql-alternative-to-store-160-tb-of-data.md

File metadata and controls

289 lines (238 loc) · 11 KB
title author date summary tags aliases image customer customerCategory logo
Goodbye, MySQL Sharding: Use a Scale-Out MySQL Alternative to Store 160+ TB of Data
Chunlei Liu
Kai Xuan
2021-01-14
This post talks about how TiDB, a highly available MySQL alternative, helps 58.com avoid MySQL sharding, scale out their databases, and simplify their database maintenance.
MySQL
No sharding
Scalability
High availability
HTAP
Real-time analytics
/case-studies/no-sharding-no-etl-use-scale-out-mysql-alternative-to-store-160-tb-of-data
/images/blog/no-mysql-sharding-no-etl-scale-out-mysql-alternative.jpg
58.com
Internet
/images/blog/customers/58-logo.png

Industry: Advertising

Authors:

  • Chunlei Liu (Senior DBA at 58.com)
  • Kai Xuan (Former Senior DBA at 58.com)

Transcreator: Caitin Chen; Editor: Tom Dewan

No MySQL sharding, No ETL, scale-out MySQL alternative

58.com is China's leading online marketplace for classifieds covering various categories, such as jobs, real estate, automotive, financing, used goods, and local services. Merchants and consumers can publish their advertisements on our online platform so that they can connect, share information, and conduct business. By the end of 2018, we had more than 500 million users, and our total revenue in 2019 was nearly US $2.23 billion.

As our businesses grew, large amounts of data flooded into our databases. But standalone MySQL databases couldn't store so much data, and MySQL sharding was an undesirable solution. To achieve MySQL high availability, we needed an external tool. To perform Online Analytical Processing (OLAP), we had to use complicated and tedious extract, transform, load (ETL) jobs. We looked for a scalable, easy-to-maintain, highly available database to solve these issues.

After an investigation, we adopted TiDB, an open-source, distributed, Hybrid Transactional/Analytical Processing (HTAP) database. Now, our production environment has 52 TiDB clusters that store 160+ TB of data, with 320+ servers running in 15 applications. To maintain such large-scale clusters, we only need two DBAs.

In this post, we'll deep dive into why we migrated from MySQL to TiDB and how we use TiDB to scale out our databases, perform real-time analytics, and achieve high availability.

Our pain points

We stored our private cloud monitoring data in MySQL. Because our data size was huge, we needed to regularly clean our tables. This was a lot of work for our DBAs.

When we used MySQL, we encountered these problems:

  • A standalone MySQL database has limited capacity. It can't store enough data to meet our needs.
  • Sharding was troublesome. At 58.com, we didn't have a middleware team, and our developers needed to operate and maintain sharding scenarios by themselves. After we sharded our database, it took a lot of work to aggregate the data.
  • To perform OLAP analytics, we must do complex and boring ETL tasks. An ETL process was time-consuming, and this hindered our ability to do real-time data analytics.
  • To achieve high availability, MySQL relies on an external tool. We used Master High Availability (MHA) to implement MySQL high availability, but it increased our maintenance cost.
  • In the primary-secondary database framework, MySQL has high latency on the secondary database. When we performed data description language (DDL) operations, high latency occurred in the secondary database. This greatly affected real-time reads.
  • A standalone MySQL database couldn't support large amounts of writes. When our writes in a standalone MySQL database reached about 15,000 rows of data, we encountered a database performance bottleneck.

Therefore, we looked for a new database solution with the following capabilities:

  • It has an active community. If its community is not active, when we find issues or bugs, we can't get solutions.
  • It's easy to operate and maintain.
  • It can solve our current problems, such as issues brought by sharding and lots of writes and deletes.
  • It is suitable for many application scenarios and provides multiple solutions.

Why we chose TiDB, a NewSQL database

TiDB is an open-source, cloud-native, distributed SQL database built by PingCAP and its open-source community. It is MySQL compatible and features horizontal scalability, strong consistency, and high availability. It's a one-stop solution for both Online Transactional Processing (OLTP) and OLAP workloads. You can learn more about TiDB's architecture here.

We adopted TiDB, because it met all our requirements for the database:

  • TiDB uses distributed storage, and it can easily scale out. We no longer need to worry that a single machine's capacity is limited.
  • TiDB is highly available. We don't need to use an additional high-availability service. So TiDB helps us eliminate extra operation and maintenance costs.
  • TiDB supports writes from multiple nodes. This prevents a database performance bottleneck when we write about 15,000 rows of data to a single node.
  • TiDB provides data aggregation solutions for sharding. With TiDB, we no longer need to shard databases.
  • TiDB has a complete monitoring system. We don't need to build our own monitoring software.

How we're using TiDB as an alternative to MySQL

So far, we've deployed 52 TiDB clusters in the production environment to store 160+ TB of data, with 320+ servers running in 15 applications. Our databases have 5.5 billion daily visits. To maintain such large-scale clusters, we only need two DBAs, and, meanwhile, they're also managing MySQL databases.

TiDB's architecture at 58.com

In a TiDB cluster, our application separates read and write domain names, and the backend uses load balancers to balance read and write loads. By default, a single cluster has four TiDB nodes, one for writes and three for reads. If write performance becomes a bottleneck, we can add TiDB nodes.

TiDB's scale-out database architecture at 58.com

TiDB's architecture at 58.com

From TiDB 2.0 to TiDB 4.0.2

The following table summarizes how we've used TiDB over the years and how our reliance on TiDB has grown.

Date TiDB version Environment
April 2018 TiDB 2.0 was introduced in production
  • Migrated our monitoring log to TiDB
  • About 2,000 visits per second
  • 7 TB–8 TB of data
December 2018 Upgraded to TiDB 2.1 Four TiDB clusters in production storing all our private cloud's log system data
September 2019 Upgraded to TiDB 3.02 in production and test environments
  • 80 TiDB physical machines storing 40 TB of data
  • 1 billion daily visits
  • 2,300 MySQL clusters and 11 TiDB clusters
  • (We gradually migrated from MySQL to TiDB.)
September 2020 Upgraded all TiDB clusters to 4.0.2
  • The Technology and Engineering Group (TEG) Payment Management Center (PMC) order statement application is our core business, and we guarantee that not a single row of its data would be lost. It stored 20 TB of data, with 4,000 read queries per second (QPS) and 1,000 write QPS.
  • It took us only 9 minutes to upgrade our 20 TB cluster!

Our applications running on TiDB

We're running 242 TiDB databases in 15 applications. To name a few:

TiDB clusters Applications Application details Detailed information Comment
TEG WList management backend and audio configuration WList is a general-purpose distributed key-list storage platform developed by the 58 Architecture Platform Department Data size: 6.3 billion rows of data
  • 7.2 TB
  • In the past 6 months, 7 flash memory cards have been damaged in these applications. But thanks to TiDB's high availability, they didn't affect our businesses.
WTable management backend The WTable cluster application accesses monitoring data such as machine CPU, I/O, and network Data size: 26.2 billion rows of data
Search index It stores data about user search terms used in the last year
  • Current data size: 300 million rows of data
  • Daily increase: 500 thousand to 800 thousand rows of data
Anjuke (a real estate information service platform) Operation log Operation log Daily increase: 40 million rows of data
Statement log Statement log
  • QPS: 10 thousand at peak
  • Disk: 1 TB
User growth 58 information News
  • Data size: 50 million rows of data
  • Disk: 1 TB
Information security Verification center Verification center Daily increase: 1 million rows of data
Financial company The financial real-time data warehouse's underlying data storage The financial real-time data warehouse's underlying data storage Thousands of tables

Conclusion

Thanks to TiDB's horizontal scalability, high availability, and HTAP capabilities, we could say goodbye to troublesome MySQL sharding and time-consuming ETL. It's so easy to scale out our databases and maintain such large-scale clusters.

Now, at 58.com, we use databases including MySQL, Redis, MongoDB, Elasticsearch, and TiDB. We receive about 800 billion visits per day. We have 4,000+ clusters, with 1,500+ physical servers.

If you have any questions or you'd like to learn more about our experience with TiDB, you can join the TiDB community on Slack.