Skip to content
This repository has been archived by the owner on Aug 25, 2024. It is now read-only.
/ MongoShake Public archive
forked from alibaba/MongoShake

MongoShake is a universal data replication platform based on MongoDB's oplog. Redundant replication and active-active replication are two most important functions. 基于mongodb oplog的集群复制工具,可以满足迁移和同步的需求,进一步实现灾备和多活功能。

License

Notifications You must be signed in to change notification settings

optimus-ohya/MongoShake

 
 

Repository files navigation

This is a brief introduction of Mongo-Shake, please visit english wiki or chinese wiki if you want to see more details including architecture, data flow, performance test, business showcase and so on.

Mongo-Shake


Mongo-Shake is developed and maintained by Nosql Team in Alibaba-Cloud.
Mongo-Shake is a universal platform for services based on MongoDB's oplog. It fetches oplog from source mongo database, and replays in the target mongo database or sends to other ends in different tunnels. If the target side is mongo database which means replay oplog directly, it's like a syncing tool that used to copy data from source MongoDB to another MongoDB to build redundant replication or active-active replication. Except for this direct way, there are others tunnel types such like rpc, file, tcp, kafka. Receivers wrote by users must define their own interfaces to connecting to these tunnels respectively. Users can also define there own tunnel type which is pluggable. If connecting to a third-party message middleware like kafka, the consumer can get the subscriber data in an asynchronous way in pub/sub module flexibly. Here comes general data flow,
pic1
The source can be either single mongod, replica set or sharding while target can be mongod or mongos. If the source is replica set, we suggest fetching data from secondary/hidden to ease the primary pressure. If the source is sharding, every shard should connect to Mongo-Shake. There can be several mongos on the target side to keep high availability, and different data will be hashed and written to different mongos.

Parallel Replication


There are three options for parallel replication which we call 'shad_key': id, collection and auto. id means the concurrency granularity is document while collection means the granularity is collection/table. auto option is decided by if there has unique index of any collection, it will change to collection with unique index exist otherwise id.

High Availability


Mongo-Shake periodically persistent its context into register center which by default is the source database. Currently, the context is checkpoint which marks the position of successfully replay oplog.
Hypervisor mechanism is also supported so that it will restart immediately when dies(master_quorum in configuration).

Filter


Support filtering database and collection namespace with whitelist and blacklist.

DDL Syncing


Starting with version 1.5, MongoShake supports syncing DDL by using global barrier. Once fetching DDL oplog, MongoShake adds a barrier so that all the belowing oplogs waitting in the queue until this oplog is written into the target MongoDB or tunnel and the checkpoint is updated. Currently, DDL is only support for ReplicaSet on the source side(target side can be RelicaSet or Sharding), we will support Sharding in the later version.
ddl

Global ID


In Aliyun internal version, global id(also called gid) is supported which marks the id of the database. It can be used to avoid loop when two databases become backup of each other. Mongo-Shake only fetches the oplogs equal to source database id, all the oplogs are be fetched when no gid gave. For current opensource version, it's not supported limited by the modification of MongoDB kernel.
If you want to build active-active replication without gid supported, please visit FAQ document to see more details.

Tunnel


As mentioned above, we support several tunnel types such like: rcp, tcp, file, kafka, mock and direct. rpc and tcp means connecting to receiver synchronously by net/rcp and TCP respectively; file means writing output into file; kafka is an asynchronous way of sending the message; mock is used by testing that throws away all the data; direct means writing into target MongoDB directly. Users can also add or modify current tunnel type.
We offer receiver to connect to different tunnels like: rpc, tcp, file, mock and kafka. Please visit FAQ document to see more details.

Compressor


Gzip, zlib, deflate compressor are supported in batched oplogs before sending.

Monitor & Debug


User can monitor or debug Mongo-Shake through RESTful API, please visit FAQ document to see more details.

Other Details


Mongo-Shake uses go-driver to fetch oplogs from source MongoDB which is later than the given timestamp in configuration. Then, it filters oplogs based on whitelist, blacklist, and gid. All the oplogs will be transferred at least once which is acceptable because of idempotent of oplog DML. We use seq and ack to make sure the package is received which is similar to the sequence and acknowledgment numbers in TCP.
The oplogs are batched together in the handling pipeline.
Users can adjust the worker concurrency and executor concurrency according to the different environment.
Please see the detail documents listed at the beginning if you want to see more details.

Code branch rules

version rules: a.b.c.

  • a: major version
  • b: minor version. even number means stable version. e.g. 1.2.x, 1.4.x, 2.0.x are stable while 1.5.x, 2.1.x aren't.
  • c: bugfix version
branch name rules
master master branch, do not allowed push code. store the latest stable version.
develop(main branch) develop branch. all the bellowing branches fork from this.
feature-* new feature branch. forked from develop branch and then merge back after finish developing, testing, and code review.
bugfix-* bugfix branch. forked from develop branch and then merge back after finish developing, testing, and code review.
improve-* improvement branch. forked from develop branch and then merge back after finish developing, testing, and code review.

tag rules: add tag when releasing: "release-v{version}-{date}". for example: "release-v1.0.2-20180628"

Usage


Run ./bin/collector.darwin or collector.linux which is built in OSX and Linux respectively.

Or you can build mongo-shake yourself according to the following steps(go version needs >= 15.10):

please note: user must modify collector.conf first to match needs. You can also use "start.sh" script which supports hypervisor mechanism in Linux OS only.

Shake series tool


We also provide some tools for synchronization in Shake series.

Thanks


Username Mail
lydarkforest linyunads1379@163.com
diggzhang diggzhang@gmail.com
ManleyLiu daywbdb@qq.com
hustchensi chensi_04@126.com
HelloCodeMing huanmingwong@163.com
cocoakekeyu cocoakekeyu@gmail.com
lixj1103 244769542@qq.com
xzshinan shinan@gongchang.com
tzjavadmg codyzeng@163.com
dx8439 171390022@qq.com
monkeyWie
raydy.yan yajuyan@hotmail.com
loda507 741536172@qq.com
骑着蜗牛的兔子 348978774@qq.com
lijwww 2530877879@qq.com
nanmu42 i@nanmu.me
zemul zemiaozhou@gmail.com

About

MongoShake is a universal data replication platform based on MongoDB's oplog. Redundant replication and active-active replication are two most important functions. 基于mongodb oplog的集群复制工具,可以满足迁移和同步的需求,进一步实现灾备和多活功能。

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Go 94.5%
  • Python 4.5%
  • Other 1.0%