image-20230223162950484

image-20230223163008047

初识Zookeeper

image-20230223163119145

image-20230223163202580

ZooKeeper的安装与配置

1.1下载安装

1、环境准备

ZooKeeper服务器是用Java创建的,它运行在JVM之上。需要安装JDK 7或更高版本。

2、上传

将下载的ZooKeeper放到/opt/ZooKeeper目录下

image-20230223163923744

3、解压

将tar包解压到/opt/zookeeper目录下

1
tar -zxvf apache-zookeeper-3.5.6-bin.tar.gz 

1.2 配置启动

1、配置zoo.cfg

进入到conf目录拷贝一个zoo_sample.cfg并完成配置

1
2
3
4
#进入到conf目录
cd /opt/zooKeeper/apache-zookeeper-3.5.6-bin/conf/
#拷贝
cp zoo_sample.cfg zoo.cfg

修改zoo.cfg

1
2
3
4
5
6
#打开目录
cd /opt/zooKeeper/
#创建zooKeeper存储目录
mkdir zkdata
#修改zoo.cfg
vim /opt/zooKeeper/apache-zookeeper-3.5.6-bin/conf/zoo.cfg

image-20230223170618800

修改存储目录:dataDir=/opt/zookeeper/zkdata

2、启动ZooKeeper

1
2
3
cd /opt/zooKeeper/apache-zookeeper-3.5.6-bin/bin/
#启动
./zkServer.sh start

image-20230223170523243

3、查看ZooKeeper状态

zookeeper启动成功。standalone代表zk没有搭建集群,现在是单节点

image-20230223170852473

ZooKeeper 命令操作

image-20230223171028658

image-20230223171126447

Zookeeper 数据模型

image-20230223171220112

Zookeeper 服务端常用命令

image-20230223171352166

image-20230223171400816

Zookeeper 客户端常用命令

image-20230223171746381

image-20230223171801531

连接zookeeper服务端

1
./zkCli.sh -server localhost:2181

image-20230223172335534

查看指定目录下节点

1
ls /

image-20230223172443145

创建节点包含数据

1
create /app1 itheima

image-20230223172722062

创建节点不包含数据

1
create /app2

image-20230223173012829

获取节点数据

image-20230223173128251

设置节点数据

image-20230223173231351

image-20230223173456624

ZooKeeper JavaAPI 操作

image-20230223174207969

Curator 介绍

image-20230223174249531

lCurator API 常用操作

image-20230223174536406

image-20230223174546904

建立连接

搭建maven工程并导入坐标

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
<?xml version="1.0" encoding="UTF-8"?>
<project xmlns="http://maven.apache.org/POM/4.0.0"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd">
<modelVersion>4.0.0</modelVersion>

<groupId>com.itheima</groupId>
<artifactId>zk-client</artifactId>
<version>1.0-SNAPSHOT</version>


<dependencies>

<dependency>
<groupId>junit</groupId>
<artifactId>junit</artifactId>
<version>4.10</version>
<scope>test</scope>
</dependency>

<!--curator-->
<dependency>
<groupId>org.apache.curator</groupId>
<artifactId>curator-framework</artifactId>
<version>4.0.0</version>
</dependency>

<dependency>
<groupId>org.apache.curator</groupId>
<artifactId>curator-recipes</artifactId>
<version>4.0.0</version>
</dependency>
<!--日志-->
<dependency>
<groupId>org.slf4j</groupId>
<artifactId>slf4j-api</artifactId>
<version>1.7.21</version>
</dependency>

<dependency>
<groupId>org.slf4j</groupId>
<artifactId>slf4j-log4j12</artifactId>
<version>1.7.21</version>
</dependency>

</dependencies>


<build>
<plugins>
<plugin>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-compiler-plugin</artifactId>
<version>3.1</version>
<configuration>
<source>1.8</source>
<target>1.8</target>
</configuration>
</plugin>
</plugins>
</build>

</project>

导入日志配置文件

image-20230223175326344

1
2
3
4
5
6
log4j.rootLogger=off,stdout

log4j.appender.stdout = org.apache.log4j.ConsoleAppender
log4j.appender.stdout.Target = System.out
log4j.appender.stdout.layout = org.apache.log4j.PatternLayout
log4j.appender.stdout.layout.ConversionPattern = [%d{yyyy-MM-dd HH/:mm/:ss}]%-5p %c(line/:%L) %x-%m%n

编写测试方法

image-20230223180603472

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
public void testConnect() {

/*
*
* @param connectString 连接字符串。zk server 地址和端口 "192.168.149.135:2181,192.168.149.136:2181"
* @param sessionTimeoutMs 会话超时时间 单位ms
* @param connectionTimeoutMs 连接超时时间 单位ms
* @param retryPolicy 重试策略
*/
/* //重试策略
RetryPolicy retryPolicy = new ExponentialBackoffRetry(3000,10);
//1.第一种方式
CuratorFramework client = CuratorFrameworkFactory.newClient("192.168.149.135:2181",
60 * 1000, 15 * 1000, retryPolicy);*/
//重试策略
RetryPolicy retryPolicy = new ExponentialBackoffRetry(3000, 10);
//2.第二种方式
//CuratorFrameworkFactory.builder();
client = CuratorFrameworkFactory.builder()
.connectString("10.211.55.5:2181")
.sessionTimeoutMs(60 * 1000)
.connectionTimeoutMs(15 * 1000)
.retryPolicy(retryPolicy)
.namespace("itheima")
.build();

//开启连接
client.start();

}

修改结构将客户端端放入全局变量

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
package com.itheima.curator;

import org.apache.curator.RetryPolicy;
import org.apache.curator.framework.CuratorFramework;
import org.apache.curator.framework.CuratorFrameworkFactory;
import org.apache.curator.framework.api.BackgroundCallback;
import org.apache.curator.framework.api.CuratorEvent;
import org.apache.curator.retry.ExponentialBackoffRetry;
import org.apache.zookeeper.CreateMode;
import org.apache.zookeeper.data.Stat;
import org.junit.After;
import org.junit.Before;
import org.junit.Test;

import java.util.List;

public class CuratorTest {


private CuratorFramework client;

/**
* 建立连接
*/
// @Before
@Before
public void testConnect() {

/*
*
* @param connectString 连接字符串。zk server 地址和端口 "192.168.149.135:2181,192.168.149.136:2181"
* @param sessionTimeoutMs 会话超时时间 单位ms
* @param connectionTimeoutMs 连接超时时间 单位ms
* @param retryPolicy 重试策略
*/
/* //重试策略
RetryPolicy retryPolicy = new ExponentialBackoffRetry(3000,10);
//1.第一种方式
CuratorFramework client = CuratorFrameworkFactory.newClient("192.168.149.135:2181",
60 * 1000, 15 * 1000, retryPolicy);*/
//重试策略
RetryPolicy retryPolicy = new ExponentialBackoffRetry(3000, 10);
//2.第二种方式
//CuratorFrameworkFactory.builder();
client = CuratorFrameworkFactory.builder()
.connectString("10.211.55.5:2181")
.sessionTimeoutMs(60 * 1000)
.connectionTimeoutMs(15 * 1000)
.retryPolicy(retryPolicy)
.namespace("itheima")
.build();

//开启连接
client.start();

}


@After
public void close() {
if (client != null) {
client.close();
}
}


}

添加节点

1
2
3
4
5
6
7
/**
* 创建节点:create 持久 临时 顺序 数据
* 1. 基本创建 :create().forPath("")
* 2. 创建节点 带有数据:create().forPath("",data)
* 3. 设置节点的类型:create().withMode().forPath("",data)
* 4. 创建多级节点 /app1/p1 :create().creatingParentsIfNeeded().forPath("",data)
*/

基本创建

1
2
3
4
5
6
7
@Test
public void testCreate2() throws Exception {
//1. 基本创建
//如果创建节点,没有指定数据,则默认将当前客户端的ip作为数据存储
String path = client.create().forPath("/app1");
System.out.println(path);
}

image-20230223183156011

1
2
3
4
5
[zk: localhost:2181(CONNECTED) 21] ls /
[itheima, zookeeper]
[zk: localhost:2181(CONNECTED) 22] ls /itheima
[app1]
[zk: localhost:2181(CONNECTED) 23]

创建节点 带有数据

1
2
3
4
5
6
7
8
@Test
public void testCreate() throws Exception {
//2. 创建节点 带有数据
//如果创建节点,没有指定数据,则默认将当前客户端的ip作为数据存储
String path = client.create().forPath("/app2", "hehe".getBytes());
System.out.println(path);

}
1
2
3
4
[zk: localhost:2181(CONNECTED) 23] ls /itheima
[app1, app2]
[zk: localhost:2181(CONNECTED) 24] get /itheima/app2
hehe

设置节点的类型

1
2
3
4
5
6
7
@Test
public void testCreate3() throws Exception {
//3. 设置节点的类型
//默认类型:持久化
String path = client.create().withMode(CreateMode.EPHEMERAL).forPath("/app3");
System.out.println(path);
}

创建多级节点

1
2
3
4
5
6
7
@Test
public void testCreate4() throws Exception {
//4. 创建多级节点 /app1/p1
//creatingParentsIfNeeded():如果父节点不存在,则创建父节点
String path = client.create().creatingParentsIfNeeded().forPath("/app4/p1");
System.out.println(path);
}
1
2
3
4
[zk: localhost:2181(CONNECTED) 25] ls /itheima
[app1, app2, app4]
[zk: localhost:2181(CONNECTED) 26] ls /itheima/app4
[p1]

查询节点

1
2
3
4
5
6
/**
* 查询节点:
* 1. 查询数据:get: getData().forPath()
* 2. 查询子节点: ls: getChildren().forPath()
* 3. 查询节点状态信息:ls -s:getData().storingStatIn(状态对象).forPath()
*/

查询数据

1
2
3
4
5
6
@Test
public void testGet1() throws Exception {
//1. 查询数据:get
byte[] data = client.getData().forPath("/app1");
System.out.println(new String(data));
}

image-20230223184921631

查询子节点

1
2
3
4
5
6
7
@Test
public void testGet2() throws Exception {
// 2. 查询子节点: ls
List<String> path = client.getChildren().forPath("/");

System.out.println(path);
}

image-20230223185032097

查询节点状态信息

1
2
3
4
5
6
7
8
@Test
public void testGet3() throws Exception {
Stat status = new Stat();
System.out.println(status);
//3. 查询节点状态信息:ls -s
client.getData().storingStatIn(status).forPath("/app1");
System.out.println(status);
}

image-20230223185335704

修改节点

1
2
3
4
5
6
7
8
/**
* 修改数据
* 1. 基本修改数据:setData().forPath()
* 2. 根据版本修改: setData().withVersion().forPath()
* * version 是通过查询出来的。目的就是为了让其他客户端或者线程不干扰我。
*
* @throws Exception
*/

基本修改数据

1
2
3
4
@Test
public void testSet() throws Exception {
client.setData().forPath("/app1", "itcast".getBytes());
}
1
2
[zk: localhost:2181(CONNECTED) 31] get /itheima/app1
itcast

根据版本修改

1
2
3
4
5
6
7
8
9
10
11
12
@Test
public void testSetForVersion() throws Exception {

Stat status = new Stat();
//3. 查询节点状态信息:ls -s
client.getData().storingStatIn(status).forPath("/app1");


int version = status.getVersion();//查询出来的 3
System.out.println(version);
client.setData().withVersion(version).forPath("/app1", "hehe".getBytes());
}
1
2
[zk: localhost:2181(CONNECTED) 32] get /itheima/app1
hehe

当版本不一致时直接报错

image-20230223191004563

删除节点

1
2
3
4
5
6
7
8
/**
* 删除节点: delete deleteall
* 1. 删除单个节点:delete().forPath("/app1");
* 2. 删除带有子节点的节点:delete().deletingChildrenIfNeeded().forPath("/app1");
* 3. 必须成功的删除:为了防止网络抖动。本质就是重试。 client.delete().guaranteed().forPath("/app2");
* 4. 回调:inBackground
* @throws Exception
*/
1
2
3
4
5
@Test
public void testDelete() throws Exception {
// 1. 删除单个节点
client.delete().forPath("/app1");
}
1
2
3
4
5
@Test
public void testDelete2() throws Exception {
//2. 删除带有子节点的节点
client.delete().deletingChildrenIfNeeded().forPath("/app4");
}
1
2
3
4
5
@Test
public void testDelete3() throws Exception {
//3. 必须成功的删除
client.delete().guaranteed().forPath("/app2");
}
1
2
3
4
5
6
7
8
9
10
11
12
@Test
public void testDelete4() throws Exception {
//4. 回调
client.delete().guaranteed().inBackground(new BackgroundCallback(){

@Override
public void processResult(CuratorFramework client, CuratorEvent event) throws Exception {
System.out.println("我被删除了~");
System.out.println(event);
}
}).forPath("/app1");
}

image-20230223191539845

Watch事件监听

image-20230223210040614

image-20230223210047460

编写测试方法

image-20230223213105760

NodeCache

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29

/**
* 演示 NodeCache:给指定一个节点注册监听器
*/

@Test
public void testNodeCache() throws Exception {
//1. 创建NodeCache对象
final NodeCache nodeCache = new NodeCache(client,"/app1");
//2. 注册监听
nodeCache.getListenable().addListener(new NodeCacheListener() {
@Override
public void nodeChanged() throws Exception {
System.out.println("节点变化了~");

//获取修改节点后的数据
byte[] data = nodeCache.getCurrentData().getData();
System.out.println(new String(data));
}
});

//3. 开启监听.如果设置为true,则开启监听是,加载缓冲数据
nodeCache.start(true);


while (true){

}
}

修改app1的值

1
2
3
4
5
6
7
8
9
[zk: localhost:2181(CONNECTED) 44] get /itheima/app1 
www
[zk: localhost:2181(CONNECTED) 45] set /itheima/app1 eeee
[zk: localhost:2181(CONNECTED) 46]
WATCHER::

WatchedEvent state:SyncConnected type:NodeDataChanged path:/itheima/app1

[zk: localhost:2181(CONNECTED) 46]

image-20230223213630042

PathChildrenCache

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
/**
* 演示 PathChildrenCache:监听某个节点的所有子节点们
*/

@Test
public void testPathChildrenCache() throws Exception {
//1.创建监听对象
PathChildrenCache pathChildrenCache = new PathChildrenCache(client,"/app2",true);

//2. 绑定监听器
pathChildrenCache.getListenable().addListener(new PathChildrenCacheListener() {
@Override
public void childEvent(CuratorFramework client, PathChildrenCacheEvent event) throws Exception {
System.out.println("子节点变化了~");
System.out.println(event);
//监听子节点的数据变更,并且拿到变更后的数据
//1.获取类型
PathChildrenCacheEvent.Type type = event.getType();
//2.判断类型是否是update
if(type.equals(PathChildrenCacheEvent.Type.CHILD_UPDATED)){
System.out.println("数据变了!!!");
byte[] data = event.getData().getData();
System.out.println(new String(data));

}
}
});
//3. 开启
pathChildrenCache.start();

while (true){

}
}

app2下新增一个节点p1

1
2
3
4
5
6
[zk: localhost:2181(CONNECTED) 47] ls /itheima/app2
[]
[zk: localhost:2181(CONNECTED) 48] create /itheima/app2/p1
Created /itheima/app2/p1
[zk: localhost:2181(CONNECTED) 49] ls /itheima/app2
[p1]

image-20230223214629556

设置p1的值

1
[zk: localhost:2181(CONNECTED) 50] set /itheima/app2/p1 qqq

image-20230223214748350

修改p1的值

1
[zk: localhost:2181(CONNECTED) 51] set /itheima/app2/p1 rrr

image-20230223214847509

TreeCache

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
/**
* 演示 TreeCache:监听某个节点自己和所有子节点们
*/

@Test
public void testTreeCache() throws Exception {
//1. 创建监听器
TreeCache treeCache = new TreeCache(client,"/app2");

//2. 注册监听
treeCache.getListenable().addListener(new TreeCacheListener() {
@Override
public void childEvent(CuratorFramework client, TreeCacheEvent event) throws Exception {
System.out.println("节点变化了");
System.out.println(event);
}
});

//3. 开启
treeCache.start();

while (true){

}
}
1
[zk: localhost:2181(CONNECTED) 53] set /itheima/app2 aaa

image-20230223215612315

1
[zk: localhost:2181(CONNECTED) 54] set /itheima/app2/p1 ttt

image-20230223215649013

image-20230223215744681

分布式锁

image-20230224214848253

image-20230224214856576

ZooKeeper分布式锁原理

image-20230224215501816

模拟12306售票案例

image-20230226202647264

image-20230226202656500

image-20230226202715656

Ticket12306.java

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
package com.itheima.curator;

import org.apache.curator.RetryPolicy;
import org.apache.curator.framework.CuratorFramework;
import org.apache.curator.framework.CuratorFrameworkFactory;
import org.apache.curator.framework.recipes.locks.InterProcessMutex;
import org.apache.curator.retry.ExponentialBackoffRetry;

import java.util.concurrent.TimeUnit;

public class Ticket12306 implements Runnable{

private int tickets = 10;//数据库的票数

private InterProcessMutex lock ;


public Ticket12306(){
//重试策略
RetryPolicy retryPolicy = new ExponentialBackoffRetry(3000, 10);
//2.第二种方式
//CuratorFrameworkFactory.builder();
CuratorFramework client = CuratorFrameworkFactory.builder()
.connectString("10.211.55.5:2181")
.sessionTimeoutMs(60 * 1000)
.connectionTimeoutMs(15 * 1000)
.retryPolicy(retryPolicy)
.build();

//开启连接
client.start();

lock = new InterProcessMutex(client,"/lock");
}

@Override
public void run() {

while(true){
//获取锁
try {
lock.acquire(3, TimeUnit.SECONDS);
if(tickets > 0){

System.out.println(Thread.currentThread()+":"+tickets);
Thread.sleep(100);
tickets--;
}
} catch (Exception e) {
e.printStackTrace();
}finally {
//释放锁
try {
lock.release();
} catch (Exception e) {
e.printStackTrace();
}
}



}

}
}

LockTest.java

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
package com.itheima.curator;

import org.apache.curator.RetryPolicy;
import org.apache.curator.framework.CuratorFramework;
import org.apache.curator.framework.CuratorFrameworkFactory;
import org.apache.curator.framework.recipes.cache.*;
import org.apache.curator.retry.ExponentialBackoffRetry;
import org.junit.After;
import org.junit.Before;
import org.junit.Test;

public class LockTest {


public static void main(String[] args) {
Ticket12306 ticket12306 = new Ticket12306();

//创建客户端
Thread t1 = new Thread(ticket12306,"携程");
Thread t2 = new Thread(ticket12306,"飞猪");

t1.start();
t2.start();
}

}

image-20230226204010581

image-20230226204241517

ZooKeeper 集群搭建

image-20230226204417018

Zookeeper 集群介绍

image-20230226204436223

ZooKeeper 集群搭建

搭建要求

准备三台虚拟机

image-20230226205329568

准备工作

以下在三台服务器上进行相同操作

安装JDK

1
2
sudo apt-get update
sudo apt-get install openjdk-8-jdk

Zookeeper压缩包上传到服务器

将Zookeeper解压指定文件夹并修改文件夹名称

1
2
3
4
5
sudo mkdir /usr/local/zookeeper-cluster
sudo chmod 777 /usr/local/zookeeper-cluster
cd /usr/local/zookeeper-cluster
tar -zxvf apache-zookeeper-3.5.6-bin.tar.gz
sudo mv apache-zookeeper-3.5.6-bin/ zookeeper/

创建data目录 ,并且将 conf下zoo_sample.cfg 文件改名为 zoo.cfg

1
2
sudo mkdir /usr/local/zookeeper-cluster/zookeeper/data
mv /usr/local/zookeeper-cluster/zookeeper/conf/zoo_sample.cfg /usr/local/zookeeper-cluster/zookeeper/conf/zoo.cfg

配置每一个Zookeeper 的dataDir 和 clientPort

修改/usr/local/zookeeper-cluster/zookeeper/conf/zoo.cfg

1
2
3
4
vim /usr/local/zookeeper-cluster/zookeeper/conf/zoo.cfg

clientPort=2181
dataDir=/usr/local/zookeeper-cluster/zookeeper/data

配置集群

  1. 在每个zookeeper的 data 目录下创建一个 myid 文件,内容分别是1、2、3 。这个文件就是记录每个服务器的ID

    Linux1

1
2
3
sudo chmod 777 /usr/local/zookeeper-cluster/zookeeper/data
echo 1 >/usr/local/zookeeper-cluster/zookeeper/data/myid

​ linux2

1
2
sudo chmod 777 /usr/local/zookeeper-cluster/zookeeper/data
echo 2 >/usr/local/zookeeper-cluster/zookeeper/data/myid

​ linux3

1
2
sudo chmod 777 /usr/local/zookeeper-cluster/zookeeper/data
echo 3 >/usr/local/zookeeper-cluster/zookeeper/data/myid
  1. 在每一个zookeeper 的 zoo.cfg配置客户端访问端口(clientPort)和集群服务器IP列表。

集群服务器IP列表如下

1
2
3
4
5
6
vim /usr/local/zookeeper-cluster/zookeeper/conf/zoo.cfg
quorumListenOnAllIPs=true
#虚拟机需要上面这条命令
server.1=10.211.55.5:2881:3881
server.2=10.211.55.6:2881:3881
server.3=10.211.55.7:2881:3881

解释:server.服务器ID=服务器IP地址:服务器之间通信端口:服务器之间投票选举端口

启动集群

启动集群就是分别启动每个实例。

1
/usr/local/zookeeper-cluster/zookeeper/bin/zkServer.sh start

image-20230227103134636

启动后我们查询一下每个实例的运行状态

1
/usr/local/zookeeper-cluster/zookeeper/bin/zkServer.sh status

image-20230227103504219

image-20230227103554850

image-20230227103640371

模拟集群异常

  1. 首先我们先测试如果是从服务器挂掉,会怎么样

把3号服务器停掉,观察1号和2号,发现状态并没有变化

1
/usr/local/zookeeper-cluster/zookeeper/bin/zkServer.sh stop

image-20230227104928349

image-20230227105015942

image-20230227105030431

由此得出结论,3个节点的集群,从服务器挂掉,集群正常

  1. 我们再把1号服务器(从服务器)也停掉,查看2号(主服务器)的状态,发现已经停止运行了。

image-20230227105202200

image-20230227105301216

由此得出结论,3个节点的集群,2个从服务器都挂掉,主服务器也无法运行。因为可运行的机器没有超过集群总数量的半数。

  1. 我们再次把1号服务器启动起来,发现2号服务器又开始正常工作了。而且依然是领导者。

image-20230227105419244

image-20230227105433752

  1. 我们把3号服务器也启动起来,把2号服务器停掉,停掉后观察1号和3号的状态。

image-20230227105547603

image-20230227105607585

image-20230227105638537

发现新的leader产生了~

由此我们得出结论,当集群中的主服务器挂了,集群中的其他服务器会自动进行选举状态,然后产生新得leader

  1. 我们再次测试,当我们把2号服务器重新启动起来启动后,会发生什么?2号服务器会再次成为新的领导吗?我们看结果

image-20230227105731893

image-20230227105756409

我们会发现,2号服务器启动后依然是跟随者(从服务器),3号服务器依然是领导者(主服务器),没有撼动3号服务器的领导地位。

由此我们得出结论,当领导者产生后,再次有新服务器加入集群,不会影响到现任领导者。

Zookeeper 核心理论

image-20230227115954918

image-20230227120338104

完结撒花🌹🌹