|
|
|
@ -133,12 +133,12 @@ def test_register_repl_instances_then_coordinators():
|
|
|
|
|
return sorted(list(execute_and_fetch_all(coordinator3_cursor, "SHOW INSTANCES")))
|
|
|
|
|
|
|
|
|
|
expected_cluster_coord3 = [
|
|
|
|
|
("coordinator_1", "127.0.0.1:10111", "", True, "coordinator"),
|
|
|
|
|
("coordinator_2", "127.0.0.1:10112", "", True, "coordinator"),
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", True, "coordinator"),
|
|
|
|
|
("instance_1", "", "127.0.0.1:10011", True, "replica"),
|
|
|
|
|
("instance_2", "", "127.0.0.1:10012", True, "replica"),
|
|
|
|
|
("instance_3", "", "127.0.0.1:10013", True, "main"),
|
|
|
|
|
("coordinator_1", "127.0.0.1:10111", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_2", "127.0.0.1:10112", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", "unknown", "coordinator"),
|
|
|
|
|
("instance_1", "", "127.0.0.1:10011", "up", "replica"),
|
|
|
|
|
("instance_2", "", "127.0.0.1:10012", "up", "replica"),
|
|
|
|
|
("instance_3", "", "127.0.0.1:10013", "up", "main"),
|
|
|
|
|
]
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_coord3, check_coordinator3)
|
|
|
|
|
|
|
|
|
@ -147,24 +147,23 @@ def test_register_repl_instances_then_coordinators():
|
|
|
|
|
def check_coordinator1():
|
|
|
|
|
return sorted(list(execute_and_fetch_all(coordinator1_cursor, "SHOW INSTANCES")))
|
|
|
|
|
|
|
|
|
|
# TODO: (andi) Waiting for product decision about this
|
|
|
|
|
expected_cluster_not_shared = [
|
|
|
|
|
("coordinator_1", "127.0.0.1:10111", "", True, "coordinator"),
|
|
|
|
|
("coordinator_2", "127.0.0.1:10112", "", True, "coordinator"),
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", True, "coordinator"),
|
|
|
|
|
("instance_1", "", "", False, "replica"),
|
|
|
|
|
("instance_2", "", "", False, "replica"),
|
|
|
|
|
("instance_3", "", "", False, "main"),
|
|
|
|
|
expected_cluster_shared = [
|
|
|
|
|
("coordinator_1", "127.0.0.1:10111", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_2", "127.0.0.1:10112", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", "unknown", "coordinator"),
|
|
|
|
|
("instance_1", "", "", "unknown", "replica"),
|
|
|
|
|
("instance_2", "", "", "unknown", "replica"),
|
|
|
|
|
("instance_3", "", "", "unknown", "main"),
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_not_shared, check_coordinator1)
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_shared, check_coordinator1)
|
|
|
|
|
|
|
|
|
|
coordinator2_cursor = connect(host="localhost", port=7691).cursor()
|
|
|
|
|
|
|
|
|
|
def check_coordinator2():
|
|
|
|
|
return sorted(list(execute_and_fetch_all(coordinator2_cursor, "SHOW INSTANCES")))
|
|
|
|
|
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_not_shared, check_coordinator2)
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_shared, check_coordinator2)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
def test_register_coordinator_then_repl_instances():
|
|
|
|
@ -190,12 +189,12 @@ def test_register_coordinator_then_repl_instances():
|
|
|
|
|
return sorted(list(execute_and_fetch_all(coordinator3_cursor, "SHOW INSTANCES")))
|
|
|
|
|
|
|
|
|
|
expected_cluster_coord3 = [
|
|
|
|
|
("coordinator_1", "127.0.0.1:10111", "", True, "coordinator"),
|
|
|
|
|
("coordinator_2", "127.0.0.1:10112", "", True, "coordinator"),
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", True, "coordinator"),
|
|
|
|
|
("instance_1", "", "127.0.0.1:10011", True, "replica"),
|
|
|
|
|
("instance_2", "", "127.0.0.1:10012", True, "replica"),
|
|
|
|
|
("instance_3", "", "127.0.0.1:10013", True, "main"),
|
|
|
|
|
("coordinator_1", "127.0.0.1:10111", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_2", "127.0.0.1:10112", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", "unknown", "coordinator"),
|
|
|
|
|
("instance_1", "", "127.0.0.1:10011", "up", "replica"),
|
|
|
|
|
("instance_2", "", "127.0.0.1:10012", "up", "replica"),
|
|
|
|
|
("instance_3", "", "127.0.0.1:10013", "up", "main"),
|
|
|
|
|
]
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_coord3, check_coordinator3)
|
|
|
|
|
|
|
|
|
@ -204,21 +203,23 @@ def test_register_coordinator_then_repl_instances():
|
|
|
|
|
def check_coordinator1():
|
|
|
|
|
return sorted(list(execute_and_fetch_all(coordinator1_cursor, "SHOW INSTANCES")))
|
|
|
|
|
|
|
|
|
|
# TODO: (andi) This should be solved eventually
|
|
|
|
|
expected_cluster_not_shared = [
|
|
|
|
|
("coordinator_1", "127.0.0.1:10111", "", True, "coordinator"),
|
|
|
|
|
("coordinator_2", "127.0.0.1:10112", "", True, "coordinator"),
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", True, "coordinator"),
|
|
|
|
|
expected_cluster_shared = [
|
|
|
|
|
("coordinator_1", "127.0.0.1:10111", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_2", "127.0.0.1:10112", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", "unknown", "coordinator"),
|
|
|
|
|
("instance_1", "", "", "unknown", "replica"),
|
|
|
|
|
("instance_2", "", "", "unknown", "replica"),
|
|
|
|
|
("instance_3", "", "", "unknown", "main"),
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_not_shared, check_coordinator1)
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_shared, check_coordinator1)
|
|
|
|
|
|
|
|
|
|
coordinator2_cursor = connect(host="localhost", port=7691).cursor()
|
|
|
|
|
|
|
|
|
|
def check_coordinator2():
|
|
|
|
|
return sorted(list(execute_and_fetch_all(coordinator2_cursor, "SHOW INSTANCES")))
|
|
|
|
|
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_not_shared, check_coordinator2)
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_shared, check_coordinator2)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
def test_coordinators_communication_with_restarts():
|
|
|
|
@ -240,10 +241,13 @@ def test_coordinators_communication_with_restarts():
|
|
|
|
|
)
|
|
|
|
|
execute_and_fetch_all(coordinator3_cursor, "SET INSTANCE instance_3 TO MAIN")
|
|
|
|
|
|
|
|
|
|
expected_cluster_not_shared = [
|
|
|
|
|
("coordinator_1", "127.0.0.1:10111", "", True, "coordinator"),
|
|
|
|
|
("coordinator_2", "127.0.0.1:10112", "", True, "coordinator"),
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", True, "coordinator"),
|
|
|
|
|
expected_cluster_shared = [
|
|
|
|
|
("coordinator_1", "127.0.0.1:10111", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_2", "127.0.0.1:10112", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", "unknown", "coordinator"),
|
|
|
|
|
("instance_1", "", "", "unknown", "replica"),
|
|
|
|
|
("instance_2", "", "", "unknown", "replica"),
|
|
|
|
|
("instance_3", "", "", "unknown", "main"),
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
coordinator1_cursor = connect(host="localhost", port=7690).cursor()
|
|
|
|
@ -251,20 +255,20 @@ def test_coordinators_communication_with_restarts():
|
|
|
|
|
def check_coordinator1():
|
|
|
|
|
return sorted(list(execute_and_fetch_all(coordinator1_cursor, "SHOW INSTANCES")))
|
|
|
|
|
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_not_shared, check_coordinator1)
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_shared, check_coordinator1)
|
|
|
|
|
|
|
|
|
|
coordinator2_cursor = connect(host="localhost", port=7691).cursor()
|
|
|
|
|
|
|
|
|
|
def check_coordinator2():
|
|
|
|
|
return sorted(list(execute_and_fetch_all(coordinator2_cursor, "SHOW INSTANCES")))
|
|
|
|
|
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_not_shared, check_coordinator2)
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_shared, check_coordinator2)
|
|
|
|
|
|
|
|
|
|
interactive_mg_runner.kill(MEMGRAPH_INSTANCES_DESCRIPTION, "coordinator_1")
|
|
|
|
|
interactive_mg_runner.start(MEMGRAPH_INSTANCES_DESCRIPTION, "coordinator_1")
|
|
|
|
|
coordinator1_cursor = connect(host="localhost", port=7690).cursor()
|
|
|
|
|
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_not_shared, check_coordinator1)
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_shared, check_coordinator1)
|
|
|
|
|
|
|
|
|
|
interactive_mg_runner.kill(MEMGRAPH_INSTANCES_DESCRIPTION, "coordinator_1")
|
|
|
|
|
interactive_mg_runner.kill(MEMGRAPH_INSTANCES_DESCRIPTION, "coordinator_2")
|
|
|
|
@ -274,11 +278,11 @@ def test_coordinators_communication_with_restarts():
|
|
|
|
|
coordinator1_cursor = connect(host="localhost", port=7690).cursor()
|
|
|
|
|
coordinator2_cursor = connect(host="localhost", port=7691).cursor()
|
|
|
|
|
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_not_shared, check_coordinator1)
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_not_shared, check_coordinator2)
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_shared, check_coordinator1)
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_shared, check_coordinator2)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
# TODO: (andi) Test when dealing with distributed coordinators that you can register on one coordinator and unregister from any other coordinator
|
|
|
|
|
# # TODO: (andi) Test when dealing with distributed coordinators that you can register on one coordinator and unregister from any other coordinator
|
|
|
|
|
@pytest.mark.parametrize(
|
|
|
|
|
"kill_instance",
|
|
|
|
|
[True, False],
|
|
|
|
@ -287,7 +291,12 @@ def test_unregister_replicas(kill_instance):
|
|
|
|
|
safe_execute(shutil.rmtree, TEMP_DIR)
|
|
|
|
|
interactive_mg_runner.start_all(MEMGRAPH_INSTANCES_DESCRIPTION)
|
|
|
|
|
|
|
|
|
|
coordinator1_cursor = connect(host="localhost", port=7690).cursor()
|
|
|
|
|
coordinator2_cursor = connect(host="localhost", port=7691).cursor()
|
|
|
|
|
coordinator3_cursor = connect(host="localhost", port=7692).cursor()
|
|
|
|
|
|
|
|
|
|
assert add_coordinator(coordinator3_cursor, "ADD COORDINATOR 1 ON '127.0.0.1:10111'")
|
|
|
|
|
assert add_coordinator(coordinator3_cursor, "ADD COORDINATOR 2 ON '127.0.0.1:10112'")
|
|
|
|
|
execute_and_fetch_all(
|
|
|
|
|
coordinator3_cursor, "REGISTER INSTANCE instance_1 ON '127.0.0.1:10011' WITH '127.0.0.1:10001'"
|
|
|
|
|
)
|
|
|
|
@ -299,6 +308,12 @@ def test_unregister_replicas(kill_instance):
|
|
|
|
|
)
|
|
|
|
|
execute_and_fetch_all(coordinator3_cursor, "SET INSTANCE instance_3 TO MAIN")
|
|
|
|
|
|
|
|
|
|
def check_coordinator1():
|
|
|
|
|
return sorted(list(execute_and_fetch_all(coordinator1_cursor, "SHOW INSTANCES")))
|
|
|
|
|
|
|
|
|
|
def check_coordinator2():
|
|
|
|
|
return sorted(list(execute_and_fetch_all(coordinator2_cursor, "SHOW INSTANCES")))
|
|
|
|
|
|
|
|
|
|
def check_coordinator3():
|
|
|
|
|
return sorted(list(execute_and_fetch_all(coordinator3_cursor, "SHOW INSTANCES")))
|
|
|
|
|
|
|
|
|
@ -308,10 +323,21 @@ def test_unregister_replicas(kill_instance):
|
|
|
|
|
return sorted(list(execute_and_fetch_all(main_cursor, "SHOW REPLICAS")))
|
|
|
|
|
|
|
|
|
|
expected_cluster = [
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", True, "coordinator"),
|
|
|
|
|
("instance_1", "", "127.0.0.1:10011", True, "replica"),
|
|
|
|
|
("instance_2", "", "127.0.0.1:10012", True, "replica"),
|
|
|
|
|
("instance_3", "", "127.0.0.1:10013", True, "main"),
|
|
|
|
|
("coordinator_1", "127.0.0.1:10111", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_2", "127.0.0.1:10112", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", "unknown", "coordinator"),
|
|
|
|
|
("instance_1", "", "127.0.0.1:10011", "up", "replica"),
|
|
|
|
|
("instance_2", "", "127.0.0.1:10012", "up", "replica"),
|
|
|
|
|
("instance_3", "", "127.0.0.1:10013", "up", "main"),
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
expected_cluster_shared = [
|
|
|
|
|
("coordinator_1", "127.0.0.1:10111", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_2", "127.0.0.1:10112", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", "unknown", "coordinator"),
|
|
|
|
|
("instance_1", "", "", "unknown", "replica"),
|
|
|
|
|
("instance_2", "", "", "unknown", "replica"),
|
|
|
|
|
("instance_3", "", "", "unknown", "main"),
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
expected_replicas = [
|
|
|
|
@ -331,6 +357,8 @@ def test_unregister_replicas(kill_instance):
|
|
|
|
|
),
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_shared, check_coordinator1)
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_shared, check_coordinator2)
|
|
|
|
|
mg_sleep_and_assert(expected_cluster, check_coordinator3)
|
|
|
|
|
mg_sleep_and_assert(expected_replicas, check_main)
|
|
|
|
|
|
|
|
|
@ -339,9 +367,19 @@ def test_unregister_replicas(kill_instance):
|
|
|
|
|
execute_and_fetch_all(coordinator3_cursor, "UNREGISTER INSTANCE instance_1")
|
|
|
|
|
|
|
|
|
|
expected_cluster = [
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", True, "coordinator"),
|
|
|
|
|
("instance_2", "", "127.0.0.1:10012", True, "replica"),
|
|
|
|
|
("instance_3", "", "127.0.0.1:10013", True, "main"),
|
|
|
|
|
("coordinator_1", "127.0.0.1:10111", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_2", "127.0.0.1:10112", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", "unknown", "coordinator"),
|
|
|
|
|
("instance_2", "", "127.0.0.1:10012", "up", "replica"),
|
|
|
|
|
("instance_3", "", "127.0.0.1:10013", "up", "main"),
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
expected_cluster_shared = [
|
|
|
|
|
("coordinator_1", "127.0.0.1:10111", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_2", "127.0.0.1:10112", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", "unknown", "coordinator"),
|
|
|
|
|
("instance_2", "", "", "unknown", "replica"),
|
|
|
|
|
("instance_3", "", "", "unknown", "main"),
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
expected_replicas = [
|
|
|
|
@ -354,6 +392,8 @@ def test_unregister_replicas(kill_instance):
|
|
|
|
|
),
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_shared, check_coordinator1)
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_shared, check_coordinator2)
|
|
|
|
|
mg_sleep_and_assert(expected_cluster, check_coordinator3)
|
|
|
|
|
mg_sleep_and_assert(expected_replicas, check_main)
|
|
|
|
|
|
|
|
|
@ -362,11 +402,22 @@ def test_unregister_replicas(kill_instance):
|
|
|
|
|
execute_and_fetch_all(coordinator3_cursor, "UNREGISTER INSTANCE instance_2")
|
|
|
|
|
|
|
|
|
|
expected_cluster = [
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", True, "coordinator"),
|
|
|
|
|
("instance_3", "", "127.0.0.1:10013", True, "main"),
|
|
|
|
|
("coordinator_1", "127.0.0.1:10111", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_2", "127.0.0.1:10112", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", "unknown", "coordinator"),
|
|
|
|
|
("instance_3", "", "127.0.0.1:10013", "up", "main"),
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
expected_cluster_shared = [
|
|
|
|
|
("coordinator_1", "127.0.0.1:10111", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_2", "127.0.0.1:10112", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", "unknown", "coordinator"),
|
|
|
|
|
("instance_3", "", "", "unknown", "main"),
|
|
|
|
|
]
|
|
|
|
|
expected_replicas = []
|
|
|
|
|
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_shared, check_coordinator1)
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_shared, check_coordinator2)
|
|
|
|
|
mg_sleep_and_assert(expected_cluster, check_coordinator3)
|
|
|
|
|
mg_sleep_and_assert(expected_replicas, check_main)
|
|
|
|
|
|
|
|
|
@ -375,7 +426,11 @@ def test_unregister_main():
|
|
|
|
|
safe_execute(shutil.rmtree, TEMP_DIR)
|
|
|
|
|
interactive_mg_runner.start_all(MEMGRAPH_INSTANCES_DESCRIPTION)
|
|
|
|
|
|
|
|
|
|
coordinator1_cursor = connect(host="localhost", port=7690).cursor()
|
|
|
|
|
coordinator2_cursor = connect(host="localhost", port=7691).cursor()
|
|
|
|
|
coordinator3_cursor = connect(host="localhost", port=7692).cursor()
|
|
|
|
|
assert add_coordinator(coordinator3_cursor, "ADD COORDINATOR 1 ON '127.0.0.1:10111'")
|
|
|
|
|
assert add_coordinator(coordinator3_cursor, "ADD COORDINATOR 2 ON '127.0.0.1:10112'")
|
|
|
|
|
execute_and_fetch_all(
|
|
|
|
|
coordinator3_cursor, "REGISTER INSTANCE instance_1 ON '127.0.0.1:10011' WITH '127.0.0.1:10001'"
|
|
|
|
|
)
|
|
|
|
@ -387,16 +442,35 @@ def test_unregister_main():
|
|
|
|
|
)
|
|
|
|
|
execute_and_fetch_all(coordinator3_cursor, "SET INSTANCE instance_3 TO MAIN")
|
|
|
|
|
|
|
|
|
|
def check_coordinator1():
|
|
|
|
|
return sorted(list(execute_and_fetch_all(coordinator1_cursor, "SHOW INSTANCES")))
|
|
|
|
|
|
|
|
|
|
def check_coordinator2():
|
|
|
|
|
return sorted(list(execute_and_fetch_all(coordinator2_cursor, "SHOW INSTANCES")))
|
|
|
|
|
|
|
|
|
|
def check_coordinator3():
|
|
|
|
|
return sorted(list(execute_and_fetch_all(coordinator3_cursor, "SHOW INSTANCES")))
|
|
|
|
|
|
|
|
|
|
expected_cluster = [
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", True, "coordinator"),
|
|
|
|
|
("instance_1", "", "127.0.0.1:10011", True, "replica"),
|
|
|
|
|
("instance_2", "", "127.0.0.1:10012", True, "replica"),
|
|
|
|
|
("instance_3", "", "127.0.0.1:10013", True, "main"),
|
|
|
|
|
("coordinator_1", "127.0.0.1:10111", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_2", "127.0.0.1:10112", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", "unknown", "coordinator"),
|
|
|
|
|
("instance_1", "", "127.0.0.1:10011", "up", "replica"),
|
|
|
|
|
("instance_2", "", "127.0.0.1:10012", "up", "replica"),
|
|
|
|
|
("instance_3", "", "127.0.0.1:10013", "up", "main"),
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
expected_cluster_shared = [
|
|
|
|
|
("coordinator_1", "127.0.0.1:10111", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_2", "127.0.0.1:10112", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", "unknown", "coordinator"),
|
|
|
|
|
("instance_1", "", "", "unknown", "replica"),
|
|
|
|
|
("instance_2", "", "", "unknown", "replica"),
|
|
|
|
|
("instance_3", "", "", "unknown", "main"),
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_shared, check_coordinator1)
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_shared, check_coordinator2)
|
|
|
|
|
mg_sleep_and_assert(expected_cluster, check_coordinator3)
|
|
|
|
|
|
|
|
|
|
try:
|
|
|
|
@ -410,20 +484,43 @@ def test_unregister_main():
|
|
|
|
|
interactive_mg_runner.kill(MEMGRAPH_INSTANCES_DESCRIPTION, "instance_3")
|
|
|
|
|
|
|
|
|
|
expected_cluster = [
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", True, "coordinator"),
|
|
|
|
|
("instance_1", "", "127.0.0.1:10011", True, "main"),
|
|
|
|
|
("instance_2", "", "127.0.0.1:10012", True, "replica"),
|
|
|
|
|
("instance_3", "", "127.0.0.1:10013", False, "unknown"),
|
|
|
|
|
("coordinator_1", "127.0.0.1:10111", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_2", "127.0.0.1:10112", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", "unknown", "coordinator"),
|
|
|
|
|
("instance_1", "", "127.0.0.1:10011", "up", "main"),
|
|
|
|
|
("instance_2", "", "127.0.0.1:10012", "up", "replica"),
|
|
|
|
|
("instance_3", "", "127.0.0.1:10013", "down", "unknown"),
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
expected_cluster_shared = [
|
|
|
|
|
("coordinator_1", "127.0.0.1:10111", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_2", "127.0.0.1:10112", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", "unknown", "coordinator"),
|
|
|
|
|
("instance_1", "", "", "unknown", "main"),
|
|
|
|
|
("instance_2", "", "", "unknown", "replica"),
|
|
|
|
|
("instance_3", "", "", "unknown", "main"),
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_shared, check_coordinator1)
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_shared, check_coordinator2)
|
|
|
|
|
mg_sleep_and_assert(expected_cluster, check_coordinator3)
|
|
|
|
|
|
|
|
|
|
execute_and_fetch_all(coordinator3_cursor, "UNREGISTER INSTANCE instance_3")
|
|
|
|
|
|
|
|
|
|
expected_cluster = [
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", True, "coordinator"),
|
|
|
|
|
("instance_1", "", "127.0.0.1:10011", True, "main"),
|
|
|
|
|
("instance_2", "", "127.0.0.1:10012", True, "replica"),
|
|
|
|
|
("coordinator_1", "127.0.0.1:10111", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_2", "127.0.0.1:10112", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", "unknown", "coordinator"),
|
|
|
|
|
("instance_1", "", "127.0.0.1:10011", "up", "main"),
|
|
|
|
|
("instance_2", "", "127.0.0.1:10012", "up", "replica"),
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
expected_cluster_shared = [
|
|
|
|
|
("coordinator_1", "127.0.0.1:10111", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_2", "127.0.0.1:10112", "", "unknown", "coordinator"),
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", "unknown", "coordinator"),
|
|
|
|
|
("instance_1", "", "", "unknown", "main"),
|
|
|
|
|
("instance_2", "", "", "unknown", "replica"),
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
expected_replicas = [
|
|
|
|
@ -441,259 +538,8 @@ def test_unregister_main():
|
|
|
|
|
def check_main():
|
|
|
|
|
return sorted(list(execute_and_fetch_all(main_cursor, "SHOW REPLICAS")))
|
|
|
|
|
|
|
|
|
|
mg_sleep_and_assert(expected_cluster, check_coordinator3)
|
|
|
|
|
mg_sleep_and_assert(expected_replicas, check_main)
|
|
|
|
|
|
|
|
|
|
def test_register_coordinator_then_repl_instances():
|
|
|
|
|
safe_execute(shutil.rmtree, TEMP_DIR)
|
|
|
|
|
interactive_mg_runner.start_all(MEMGRAPH_INSTANCES_DESCRIPTION)
|
|
|
|
|
|
|
|
|
|
coordinator3_cursor = connect(host="localhost", port=7692).cursor()
|
|
|
|
|
|
|
|
|
|
assert add_coordinator(coordinator3_cursor, "ADD COORDINATOR 1 ON '127.0.0.1:10111'")
|
|
|
|
|
assert add_coordinator(coordinator3_cursor, "ADD COORDINATOR 2 ON '127.0.0.1:10112'")
|
|
|
|
|
execute_and_fetch_all(
|
|
|
|
|
coordinator3_cursor, "REGISTER INSTANCE instance_1 ON '127.0.0.1:10011' WITH '127.0.0.1:10001'"
|
|
|
|
|
)
|
|
|
|
|
execute_and_fetch_all(
|
|
|
|
|
coordinator3_cursor, "REGISTER INSTANCE instance_2 ON '127.0.0.1:10012' WITH '127.0.0.1:10002'"
|
|
|
|
|
)
|
|
|
|
|
execute_and_fetch_all(
|
|
|
|
|
coordinator3_cursor, "REGISTER INSTANCE instance_3 ON '127.0.0.1:10013' WITH '127.0.0.1:10003'"
|
|
|
|
|
)
|
|
|
|
|
execute_and_fetch_all(coordinator3_cursor, "SET INSTANCE instance_3 TO MAIN")
|
|
|
|
|
|
|
|
|
|
def check_coordinator3():
|
|
|
|
|
return sorted(list(execute_and_fetch_all(coordinator3_cursor, "SHOW INSTANCES")))
|
|
|
|
|
|
|
|
|
|
expected_cluster_coord3 = [
|
|
|
|
|
("coordinator_1", "127.0.0.1:10111", "", True, "coordinator"),
|
|
|
|
|
("coordinator_2", "127.0.0.1:10112", "", True, "coordinator"),
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", True, "coordinator"),
|
|
|
|
|
("instance_1", "", "127.0.0.1:10011", True, "replica"),
|
|
|
|
|
("instance_2", "", "127.0.0.1:10012", True, "replica"),
|
|
|
|
|
("instance_3", "", "127.0.0.1:10013", True, "main"),
|
|
|
|
|
]
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_coord3, check_coordinator3)
|
|
|
|
|
|
|
|
|
|
coordinator1_cursor = connect(host="localhost", port=7690).cursor()
|
|
|
|
|
|
|
|
|
|
def check_coordinator1():
|
|
|
|
|
return sorted(list(execute_and_fetch_all(coordinator1_cursor, "SHOW INSTANCES")))
|
|
|
|
|
|
|
|
|
|
# TODO: (andi) This should be solved eventually
|
|
|
|
|
expected_cluster_not_shared = [
|
|
|
|
|
("coordinator_1", "127.0.0.1:10111", "", True, "coordinator"),
|
|
|
|
|
("coordinator_2", "127.0.0.1:10112", "", True, "coordinator"),
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", True, "coordinator"),
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_not_shared, check_coordinator1)
|
|
|
|
|
|
|
|
|
|
coordinator2_cursor = connect(host="localhost", port=7691).cursor()
|
|
|
|
|
|
|
|
|
|
def check_coordinator2():
|
|
|
|
|
return sorted(list(execute_and_fetch_all(coordinator2_cursor, "SHOW INSTANCES")))
|
|
|
|
|
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_not_shared, check_coordinator2)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
def test_coordinators_communication_with_restarts():
|
|
|
|
|
safe_execute(shutil.rmtree, TEMP_DIR)
|
|
|
|
|
interactive_mg_runner.start_all(MEMGRAPH_INSTANCES_DESCRIPTION)
|
|
|
|
|
|
|
|
|
|
coordinator3_cursor = connect(host="localhost", port=7692).cursor()
|
|
|
|
|
|
|
|
|
|
assert add_coordinator(coordinator3_cursor, "ADD COORDINATOR 1 ON '127.0.0.1:10111'")
|
|
|
|
|
assert add_coordinator(coordinator3_cursor, "ADD COORDINATOR 2 ON '127.0.0.1:10112'")
|
|
|
|
|
execute_and_fetch_all(
|
|
|
|
|
coordinator3_cursor, "REGISTER INSTANCE instance_1 ON '127.0.0.1:10011' WITH '127.0.0.1:10001'"
|
|
|
|
|
)
|
|
|
|
|
execute_and_fetch_all(
|
|
|
|
|
coordinator3_cursor, "REGISTER INSTANCE instance_2 ON '127.0.0.1:10012' WITH '127.0.0.1:10002'"
|
|
|
|
|
)
|
|
|
|
|
execute_and_fetch_all(
|
|
|
|
|
coordinator3_cursor, "REGISTER INSTANCE instance_3 ON '127.0.0.1:10013' WITH '127.0.0.1:10003'"
|
|
|
|
|
)
|
|
|
|
|
execute_and_fetch_all(coordinator3_cursor, "SET INSTANCE instance_3 TO MAIN")
|
|
|
|
|
|
|
|
|
|
expected_cluster_not_shared = [
|
|
|
|
|
("coordinator_1", "127.0.0.1:10111", "", True, "coordinator"),
|
|
|
|
|
("coordinator_2", "127.0.0.1:10112", "", True, "coordinator"),
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", True, "coordinator"),
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
coordinator1_cursor = connect(host="localhost", port=7690).cursor()
|
|
|
|
|
|
|
|
|
|
def check_coordinator1():
|
|
|
|
|
return sorted(list(execute_and_fetch_all(coordinator1_cursor, "SHOW INSTANCES")))
|
|
|
|
|
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_not_shared, check_coordinator1)
|
|
|
|
|
|
|
|
|
|
coordinator2_cursor = connect(host="localhost", port=7691).cursor()
|
|
|
|
|
|
|
|
|
|
def check_coordinator2():
|
|
|
|
|
return sorted(list(execute_and_fetch_all(coordinator2_cursor, "SHOW INSTANCES")))
|
|
|
|
|
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_not_shared, check_coordinator2)
|
|
|
|
|
|
|
|
|
|
interactive_mg_runner.kill(MEMGRAPH_INSTANCES_DESCRIPTION, "coordinator_1")
|
|
|
|
|
interactive_mg_runner.start(MEMGRAPH_INSTANCES_DESCRIPTION, "coordinator_1")
|
|
|
|
|
coordinator1_cursor = connect(host="localhost", port=7690).cursor()
|
|
|
|
|
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_not_shared, check_coordinator1)
|
|
|
|
|
|
|
|
|
|
interactive_mg_runner.kill(MEMGRAPH_INSTANCES_DESCRIPTION, "coordinator_1")
|
|
|
|
|
interactive_mg_runner.kill(MEMGRAPH_INSTANCES_DESCRIPTION, "coordinator_2")
|
|
|
|
|
|
|
|
|
|
interactive_mg_runner.start(MEMGRAPH_INSTANCES_DESCRIPTION, "coordinator_1")
|
|
|
|
|
interactive_mg_runner.start(MEMGRAPH_INSTANCES_DESCRIPTION, "coordinator_2")
|
|
|
|
|
coordinator1_cursor = connect(host="localhost", port=7690).cursor()
|
|
|
|
|
coordinator2_cursor = connect(host="localhost", port=7691).cursor()
|
|
|
|
|
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_not_shared, check_coordinator1)
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_not_shared, check_coordinator2)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
# TODO: (andi) Test when dealing with distributed coordinators that you can register on one coordinator and unregister from any other coordinator
|
|
|
|
|
@pytest.mark.parametrize(
|
|
|
|
|
"kill_instance",
|
|
|
|
|
[True, False],
|
|
|
|
|
)
|
|
|
|
|
def test_unregister_replicas(kill_instance):
|
|
|
|
|
safe_execute(shutil.rmtree, TEMP_DIR)
|
|
|
|
|
interactive_mg_runner.start_all(MEMGRAPH_INSTANCES_DESCRIPTION)
|
|
|
|
|
|
|
|
|
|
coordinator3_cursor = connect(host="localhost", port=7692).cursor()
|
|
|
|
|
execute_and_fetch_all(
|
|
|
|
|
coordinator3_cursor, "REGISTER INSTANCE instance_1 ON '127.0.0.1:10011' WITH '127.0.0.1:10001'"
|
|
|
|
|
)
|
|
|
|
|
execute_and_fetch_all(
|
|
|
|
|
coordinator3_cursor, "REGISTER INSTANCE instance_2 ON '127.0.0.1:10012' WITH '127.0.0.1:10002'"
|
|
|
|
|
)
|
|
|
|
|
execute_and_fetch_all(
|
|
|
|
|
coordinator3_cursor, "REGISTER INSTANCE instance_3 ON '127.0.0.1:10013' WITH '127.0.0.1:10003'"
|
|
|
|
|
)
|
|
|
|
|
execute_and_fetch_all(coordinator3_cursor, "SET INSTANCE instance_3 TO MAIN")
|
|
|
|
|
|
|
|
|
|
def check_coordinator3():
|
|
|
|
|
return sorted(list(execute_and_fetch_all(coordinator3_cursor, "SHOW INSTANCES")))
|
|
|
|
|
|
|
|
|
|
main_cursor = connect(host="localhost", port=7689).cursor()
|
|
|
|
|
|
|
|
|
|
def check_main():
|
|
|
|
|
return sorted(list(execute_and_fetch_all(main_cursor, "SHOW REPLICAS")))
|
|
|
|
|
|
|
|
|
|
expected_cluster = [
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", True, "coordinator"),
|
|
|
|
|
("instance_1", "", "127.0.0.1:10011", True, "replica"),
|
|
|
|
|
("instance_2", "", "127.0.0.1:10012", True, "replica"),
|
|
|
|
|
("instance_3", "", "127.0.0.1:10013", True, "main"),
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
expected_replicas = [
|
|
|
|
|
("instance_1", "127.0.0.1:10001", "sync", 0, 0, "ready"),
|
|
|
|
|
("instance_2", "127.0.0.1:10002", "sync", 0, 0, "ready"),
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
mg_sleep_and_assert(expected_cluster, check_coordinator3)
|
|
|
|
|
mg_sleep_and_assert(expected_replicas, check_main)
|
|
|
|
|
|
|
|
|
|
if kill_instance:
|
|
|
|
|
interactive_mg_runner.kill(MEMGRAPH_INSTANCES_DESCRIPTION, "instance_1")
|
|
|
|
|
execute_and_fetch_all(coordinator3_cursor, "UNREGISTER INSTANCE instance_1")
|
|
|
|
|
|
|
|
|
|
expected_cluster = [
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", True, "coordinator"),
|
|
|
|
|
("instance_2", "", "127.0.0.1:10012", True, "replica"),
|
|
|
|
|
("instance_3", "", "127.0.0.1:10013", True, "main"),
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
expected_replicas = [
|
|
|
|
|
("instance_2", "127.0.0.1:10002", "sync", 0, 0, "ready"),
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
mg_sleep_and_assert(expected_cluster, check_coordinator3)
|
|
|
|
|
mg_sleep_and_assert(expected_replicas, check_main)
|
|
|
|
|
|
|
|
|
|
if kill_instance:
|
|
|
|
|
interactive_mg_runner.kill(MEMGRAPH_INSTANCES_DESCRIPTION, "instance_2")
|
|
|
|
|
execute_and_fetch_all(coordinator3_cursor, "UNREGISTER INSTANCE instance_2")
|
|
|
|
|
|
|
|
|
|
expected_cluster = [
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", True, "coordinator"),
|
|
|
|
|
("instance_3", "", "127.0.0.1:10013", True, "main"),
|
|
|
|
|
]
|
|
|
|
|
expected_replicas = []
|
|
|
|
|
|
|
|
|
|
mg_sleep_and_assert(expected_cluster, check_coordinator3)
|
|
|
|
|
mg_sleep_and_assert(expected_replicas, check_main)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
def test_unregister_main():
|
|
|
|
|
safe_execute(shutil.rmtree, TEMP_DIR)
|
|
|
|
|
interactive_mg_runner.start_all(MEMGRAPH_INSTANCES_DESCRIPTION)
|
|
|
|
|
|
|
|
|
|
coordinator3_cursor = connect(host="localhost", port=7692).cursor()
|
|
|
|
|
execute_and_fetch_all(
|
|
|
|
|
coordinator3_cursor, "REGISTER INSTANCE instance_1 ON '127.0.0.1:10011' WITH '127.0.0.1:10001'"
|
|
|
|
|
)
|
|
|
|
|
execute_and_fetch_all(
|
|
|
|
|
coordinator3_cursor, "REGISTER INSTANCE instance_2 ON '127.0.0.1:10012' WITH '127.0.0.1:10002'"
|
|
|
|
|
)
|
|
|
|
|
execute_and_fetch_all(
|
|
|
|
|
coordinator3_cursor, "REGISTER INSTANCE instance_3 ON '127.0.0.1:10013' WITH '127.0.0.1:10003'"
|
|
|
|
|
)
|
|
|
|
|
execute_and_fetch_all(coordinator3_cursor, "SET INSTANCE instance_3 TO MAIN")
|
|
|
|
|
|
|
|
|
|
def check_coordinator3():
|
|
|
|
|
return sorted(list(execute_and_fetch_all(coordinator3_cursor, "SHOW INSTANCES")))
|
|
|
|
|
|
|
|
|
|
expected_cluster = [
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", True, "coordinator"),
|
|
|
|
|
("instance_1", "", "127.0.0.1:10011", True, "replica"),
|
|
|
|
|
("instance_2", "", "127.0.0.1:10012", True, "replica"),
|
|
|
|
|
("instance_3", "", "127.0.0.1:10013", True, "main"),
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
mg_sleep_and_assert(expected_cluster, check_coordinator3)
|
|
|
|
|
|
|
|
|
|
try:
|
|
|
|
|
execute_and_fetch_all(coordinator3_cursor, "UNREGISTER INSTANCE instance_3")
|
|
|
|
|
except Exception as e:
|
|
|
|
|
assert (
|
|
|
|
|
str(e)
|
|
|
|
|
== "Alive main instance can't be unregistered! Shut it down to trigger failover and then unregister it!"
|
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
interactive_mg_runner.kill(MEMGRAPH_INSTANCES_DESCRIPTION, "instance_3")
|
|
|
|
|
|
|
|
|
|
expected_cluster = [
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", True, "coordinator"),
|
|
|
|
|
("instance_1", "", "127.0.0.1:10011", True, "main"),
|
|
|
|
|
("instance_2", "", "127.0.0.1:10012", True, "replica"),
|
|
|
|
|
("instance_3", "", "127.0.0.1:10013", False, "unknown"),
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
mg_sleep_and_assert(expected_cluster, check_coordinator3)
|
|
|
|
|
|
|
|
|
|
execute_and_fetch_all(coordinator3_cursor, "UNREGISTER INSTANCE instance_3")
|
|
|
|
|
|
|
|
|
|
expected_cluster = [
|
|
|
|
|
("coordinator_3", "127.0.0.1:10113", "", True, "coordinator"),
|
|
|
|
|
("instance_1", "", "127.0.0.1:10011", True, "main"),
|
|
|
|
|
("instance_2", "", "127.0.0.1:10012", True, "replica"),
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
expected_replicas = [
|
|
|
|
|
("instance_2", "127.0.0.1:10002", "sync", 0, 0, "ready"),
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
main_cursor = connect(host="localhost", port=7687).cursor()
|
|
|
|
|
|
|
|
|
|
def check_main():
|
|
|
|
|
return sorted(list(execute_and_fetch_all(main_cursor, "SHOW REPLICAS")))
|
|
|
|
|
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_shared, check_coordinator1)
|
|
|
|
|
mg_sleep_and_assert(expected_cluster_shared, check_coordinator2)
|
|
|
|
|
mg_sleep_and_assert(expected_cluster, check_coordinator3)
|
|
|
|
|
mg_sleep_and_assert(expected_replicas, check_main)
|
|
|
|
|
|
|
|
|
|