June 17, 2024


Databases are designed for particular schemas, queries, and throughput, however when you’ve got knowledge that will get queried extra continuously for a time period, it’s possible you’ll wish to scale back the load in your database by introducing a cache layer. 

On this put up, we’ll take a look at the horizontally scalable Google Cloud Bigtable, which is nice for high-throughput reads and writes. Efficiency will be optimized by guaranteeing rows are queried considerably uniformly throughout the database. If we introduce a cache for extra continuously queried rows, we pace up our utility in two methods: we’re decreasing the load on hotspotted rows and rushing up responses by regionally colocating the cache and computing. 

Memcached is an in-memory key-value retailer for small chunks of arbitrary knowledge, and I will use the scalable, totally managed Memorystore for Memcached, since it’s effectively built-in with the Google Cloud ecosystem.


  1. Create a brand new Google Cloud undertaking or use an current undertaking and database of your alternative. The examples right here will present Cloud Bigtable, however Spanner or Firestore could be good choices too.

  2. I will present gcloud instructions for many of the steps, however you are able to do most of this within the Google Cloud Console in the event you favor.

  3. Create a Cloud Bigtable occasion and a desk with one row utilizing these instructions:

cbt createinstance bt-cache “Bigtable with cache” bt-cache-c1 us-central1-b 1 SSD &&  

cbt -instance=bt-cache createtable mobile-time-series “households=stats_summary” &&  

cbt -instance=bt-cache set mobile-time-series cellphone#4c410523#20190501 stats_summary:os_build=PQ2A.190405.003 stats_summary:os_name=android &&  

cbt -instance=bt-cache learn mobile-time-series

The code

The generic logic for a cache will be outlined within the following steps: 

  1. Choose a row key to question.

  2. If row key’s in cache

     three. In any other case

  • Search for the row in Cloud Bigtable.
  • Add the worth to the cache with an expiration.
  • Return the worth.

For Cloud Bigtable, your code may seem like this (full code on GitHub):


Source link

Leave a Reply

Your email address will not be published. Required fields are marked *