码迷,mamicode.com
首页 > 数据库 > 详细

DataX安装使用实现MySQL到MySQL数据同步

时间:2020-07-26 11:17:36      阅读:129      评论:0      收藏:0      [点我收藏+]

标签:使用   shu   mic   tab   oracle   hdf   刷新   数据   default   

DataX安装使用实现MySQL到MySQL数据同步

1.前置条件:

1.1jdk安装

  • jdk安装前往官网,这里我安装jdk-8u261

  • 解压

    sudo mkdir -p /opt/moudle
    sudo tar -zxvf jdk-8u261-linux-x64.tar.gz -C /opt/moudle/
    
    
  • 设置环境变量

    export JAVA_HOME=/opt/moudle/jdk1.8.0_261
    export JRE_HOME=${JAVA_HOME}/jre
    export CLASSPATH=.:${JAVA_HOME}/lib:${JRE_HOME}/lib
    export PATH=${JAVA_HOME}/bin:$PATH
    
  • 刷新配置

    source /etc/profile
    
  • 检查java

    java -version
    
    # 出现下面安装成功
    java version "1.8.0_261"
    Java(TM) SE Runtime Environment (build 1.8.0_261-b12)
    Java HotSpot(TM) 64-Bit Server VM (build 25.261-b12, mixed mode)
    

1.2python安装

  • 略(官方推荐>=2.6.X)

1.3 Hadoop单机伪分布式安装

2.安装DataX

  • DataX是阿里巴巴的一个异构数据源离线同步工具,致力于实现包括关系型数据库(MySQL、Oracle等)、HDFS、Hive、ODPS、HBase、FTP等各种异构数据源之间稳定高效的数据同步功能。

    !技术图片

  • 下载地址:http://datax-opensource.oss-cn-hangzhou.aliyuncs.com/datax.tar.gz

  • 解压

    tar -zxvf datax.tar.gz -C /opt/software/
    
  • 运行自检脚本

    cd /opt/software/datax/
    bin/datax.py job/job.json
    
  • 出现下面界面表示成功:

技术图片

  • /opt/software/datax/job/job.json格式。
{
	"content":[
		{
			"reader":{
				"name":"streamreader",# 流式读,根据DataX定义好的设置
				"parameter":{
					"column":[#把column里所有value读到流当中
						{
							"type":"string",
							"value":"DataX"
						},
						{
							"type":"long",
							"value":19890604
						},
						{
							"type":"date",
							"value":"1989-06-04 00:00:00"
						},
						{
							"type":"bool",
							"value":true
						},
						{
							"type":"bytes",
							"value":"test"
						}
					],
					"sliceRecordCount":100000
				}
			},
			"writer":{
				"name":"streamwriter",# 流式写,根据DataX定义好的设置
				"parameter":{
					"encoding":"UTF-8",
					"print":false#打印
				}
			}
		}
	],
	"setting":{
		"errorLimit":{# errorLimit错误限制
			"percentage":0.02,# 最大容忍错误限制百分比2%
			"record":0# 容忍错误记录调试 0
		},
		"speed":{# 控制并发数:通过byte或channel控制,这里默认通过byte控制
			"byte":10485760#以 sliceRecordCount乘以byte,打印数据条数占用空间
		}
	}
}

3.基本使用

3.1从stream读取数据并打印到控制台。

  • 首先查看官方json配置模版

    # 查看 streamreader --> streamwriter 模版
    python /opt/software/datax/bin/datax.py -r streamreader -w streamwriter
    # 模版如下:
    DataX (DATAX-OPENSOURCE-3.0), From Alibaba !
    Copyright (C) 2010-2017, Alibaba Group. All Rights Reserved.
    
    
    Please refer to the streamreader document:
         https://github.com/alibaba/DataX/blob/master/streamreader/doc/streamreader.md 
    
    Please refer to the streamwriter document:
         https://github.com/alibaba/DataX/blob/master/streamwriter/doc/streamwriter.md 
     
    Please save the following configuration as a json file and  use
         python {DATAX_HOME}/bin/datax.py {JSON_FILE_NAME}.json 
    to run the job.
    {
        "job": {
            "content": [
                {
                    "reader": {
                        "name": "streamreader", 
                        "parameter": {
                            "column": [], 
                            "sliceRecordCount": ""
                        }
                    }, 
                    "writer": {
                        "name": "streamwriter", 
                        "parameter": {
                            "encoding": "", 
                            "print": true
                        }
                    }
                }
            ], 
            "setting": {
                "speed": {
                    "channel": ""
                }
            }
        }
    }
    
  • 根据模版编写json文件

    {
        "job": {
            "content": [
                {
                    "reader": {
                        "name": "streamreader", 
                        "parameter": {
                            "column": [
                                {
                                    "type":"string",
                                    "value":"xujunkai, hello world!"
                                },
                                {
                                    "type":"string",
                                    "value":"徐俊凯, 你好!"
                                },
                            ], 
                            "sliceRecordCount": "10"#打印次数
                        }
                    }, 
                    "writer": {
                        "name": "streamwriter", 
                        "parameter": {
                            "encoding": "utf-8", #编码方式utf-8
                            "print": true
                        }
                    }
                }
            ], 
            "setting": {
                "speed": {#控制并发数
                    "channel": "2"#控制并发2次-->这里因为是打印所以会sliceRecordCount乘以channel 打印20遍。如果设置为mysql真的会进行并发
                }
            }
        }
    }
    
  • 创建一个json文件,在根目录

    mkdir json
    cd json/
    vim stream2stream.json
    # 将上述内容粘贴进去
    
  • 运行job

    /opt/software/datax/bin/datax.py ./stream2stream.json
    
  • 如下图:

    技术图片

3.2从MySQL到MySQL批量插入

3.2.1预先准备工作:
  • 写入和读取方准备创建库和表

    # 创建库
    create database `testdatax` character set utf8
    # 创建表
    create table user1w(
    id int not null auto_increment,
    name varchar(10) not null,
    score int not null,
    primary key(`id`))engine=InnoDB default charset=utf8;
    
  • 编写一个简单存储过程,读取数据端插入数据:

    DELIMITER //
    create PROCEDURE add_user(in num INT)
    BEGIN
    DECLARE rowid INT DEFAULT 0;
    DECLARE name CHAR(1);
    DECLARE score INT;
    WHILE rowid < num DO
    SET rowid = rowid + 1;
    set name = SUBSTRING(‘abcdefghijklmnopqrstuvwxyz‘,ROUND(1+25*RAND()),1);
    set score= FLOOR(40 + (RAND()*60));
    insert INTO user1w (name,score) VALUES (name,score);
    END WHILE;
    END //
    DELIMITER ;
    
  • 执行插入数据

    call add_user(10000);
    
3.2.2 查看一下mysql到mysql的json配置
  • python /opt/software/datax/bin/datax.py -r mysqlreader -w mysqlwriter,json文件配置:

    {
        "job": {
            "content": [
                {
                    "reader": {
                        "name": "mysqlreader", # 读取端,根据DataX定义好的设置
                        "parameter": {
                            "column": [], # 读取端需要同步的列
                		   "splitPk": "",# 数据抽取时指定字段进行数据分片
                            "connection": [
                                {
                                    "jdbcUrl": [], #读取端连接信息
                                    "table": []# 读取端指定的表
                                }
                            ], 
                            "password": "", #读取端账户
                            "username": "", #读取端密码
                            "where": ""# 描述筛选条件
                        }
                    }, 
                    "writer": {
                        "name": "mysqlwriter", #写入端,根据DataX定义好的设置
                        "parameter": {
                            "column": [], #写入端需要同步的列
                            "connection": [
                                {
                                    "jdbcUrl": "", # 写入端连接信息
                                    "table": []# 写入端指定的表
                                }
                            ], 
                            "password": "", #写入端密码
                            "preSql": [], # 执行写入之前做的事情
                            "session": [], 
                            "username": "", #写入端账户
                            "writeMode": ""# 操作乐星
                        }
                    }
                }
            ], 
            "setting": {
                "speed": {
                    "channel": ""#指定channel数
                }
            }
        }
    }
    
  • 我的配置json:

    {
        "job": {
            "content": [
                {
                    "reader": {
                        "name": "mysqlreader", 
                        "parameter": {
                            "username": "root",
                            "password": "123"
                            "column": ["*"],
                            "splitPk": "id",
                            "connection": [
                                {
                                    "jdbcUrl": [
                                        "jdbc:mysql://读取端IP:3306/testdatax?useUnicode=true&characterEncoding=utf8"
                                    ], 
                                    "table": ["user1w"]
                                }
                            ]
                        }
                    }, 
                    "writer": {
                        "name": "mysqlwriter", 
                        "parameter": {
                            "column": ["*"], 
                            "connection": [
                                {
                                    "jdbcUrl": "jdbc:mysql://写入端IP:3306/testdatax?useUnicode=true&characterEncoding=utf8", 
                                    "table": ["user1w"]
                                }
                            ], 
                            "password": "123", 
                            "preSql": [
                                "truncate user1w"
                            ], 
                            "session": [
                                "set session sql_mode=‘ANSI‘"
                            ], 
                            "username": "root", 
                            "writeMode": "insert"
                        }
                    }
                }
            ], 
            "setting": {
                "speed": {
                    "channel": "5"
                }
            }
        }
    }
    
  • cd到datax下bin目录执行:

    python2 datax.py /root/json/mysql2mysql.json
    
  • 会打印同步数据信息完毕。更多配置见github-dataX

3.3从数据库MySQL数据导入到HDFS中

python /opt/software/datax/bin/datax.py -r mysqlreader -w mysqlwriter
  • 未完待续...

DataX安装使用实现MySQL到MySQL数据同步

标签:使用   shu   mic   tab   oracle   hdf   刷新   数据   default   

原文地址:https://www.cnblogs.com/xujunkai/p/13378942.html

(0)
(0)
   
举报
评论 一句话评论(0
登录后才能评论!
© 2014 mamicode.com 版权所有  联系我们:gaon5@hotmail.com
迷上了代码!