Aerospike version:C-7.0.0.10 Server OS: Ubuntu22.04 There are 2 node in the cluster, and they can find each other. Node1 config:/etc/aerospike/aerospike.conf: #####################################################################
# Aerospike database configuration file for use with systemd.
service {
user root
group root
pidfile /var/run/aerospike/asd.pid
service-threads 20 # Should be 5 times number of vCPUs for 4.7+ and
# at least one SSD namespace, otherwise number of vCPUs
proto-fd-max 15000
node-id-interface ens34f0np0
cluster-name cakery
}
logging {
file /var/log/aerospike/aerospike.log {
context any info
}
}
network {
service {
address any
access-address 109.105.118.39
port 3000
}
heartbeat {
mode mesh
address 109.105.118.39
port 3002
mesh-seed-address-port 109.105.118.129 3002
mesh-seed-address-port 109.105.118.39 3002
interval 150
timeout 10
}
fabric {
port 3001
}
info {
port 3003
}
}
namespace wyf_test_39 {
replication-factor 2
storage-engine device {
# Use one or more lines like those below with actual device paths.
device /dev/nvme2n1
max-write-cache 2048M
write-block-size 1M
}
}
##################################################################################
Node2 config:/etc/aerospike/aerospike.conf:
# Aerospike database configuration file for use with systemd.
service {
user root
group root
pidfile /var/run/aerospike/asd.pid
service-threads 20 # Should be 5 times number of vCPUs for 4.7+ and
# at least one SSD namespace, otherwise number of vCPUs
proto-fd-max 15000
node-id-interface eno8303
cluster-name cakery
}
logging {
file /var/log/aerospike/aerospike.log {
context any info
}
}
network {
service {
address any
access-address 109.105.118.129
port 3000
}
heartbeat {
mode mesh
address 109.105.118.129
port 3002
mesh-seed-address-port 109.105.118.39 3002
mesh-seed-address-port 109.105.118.129 3002
interval 150
timeout 10
}
fabric {
port 3001
}
info {
port 3003
}
}
namespace test_129 {
replication-factor 2
storage-engine device {
# Use one or more lines like those below with actual device paths.
device /dev/nvme0n1
max-write-cache 2048M
write-block-size 1M
}
}
I load some data with YCSB on node1: ./bin/ycsb.sh load aerospike -s -P /myworkloadconfig -threads 16。 The data can be load successfully on node1, but NOT replica on node 2. Below is the “info” command result through asadm on node 1:
base) root@testpc11839:/home/czm/YCSB# asadm
Seed: [('127.0.0.1', 3000, None)]
Config_file: /root/.aerospike/astools.conf, /etc/aerospike/astools.conf
Aerospike Interactive Shell, version 2.22.0
Found 2 nodes
Online: 109.105.118.39:3000, 109.105.118.129:3000
Admin> info
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~Network Information (2024-06-20 06:10:36 UTC)~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Node| Node ID| IP| Build|Migrations|~~~~~~~~~~~~~~~~~Cluster~~~~~~~~~~~~~~~~~~|Client| Uptime
| | | | |Size| Key|Integrity| Principal| Conns|
testpc118129:3000 |*BB9EC404C722AEC|109.105.118.129:3000|C-7.0.0.10| 0.000 | 2|5C7231310C1|True |BB9EC404C722AEC| 7|00:37:52
testpc11839.samsungds.net:3000| BB9508EAFF22314|109.105.118.39:3000 |C-7.0.0.10| 0.000 | 2|5C7231310C1|True |BB9EC404C722AEC| 7|00:31:50
Number of rows: 2
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~Namespace Usage Information (2024-06-20 06:10:36 UTC)~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Namespace| Node|Evictions| Stop|~System Memory~|~Primary Index~|~~Secondary~~|~~~~~~~~~~~~~~~~~~~Storage Engine~~~~~~~~~~~~~~~~~~
| | |Writes| Avail%| Evict%| Type| Used|~~~~Index~~~~| Type| Used| Used%|Evict%| Used|Avail%|Avail
| | | | | | | |Type| Used| | | | | Stop%| |Stop%
test_129 |testpc118129:3000 | 0.000 |False | 92| 0|mem | 0.000 B |mem |0.000 B |device| 0.000 B | 0.0 %| 0.0 %|70.0 %|99.0 %|5.0 %
test_129 | | 0.000 | | | | | 0.000 B | |0.000 B | | 0.000 B | 0.0 %| | | |
wyf_test_39|testpc11839.samsungds.net:3000| 0.000 |False | 54| 0|mem | 2.980 GB|mem |0.000 B |device|193.715 GB|9.46 %| 0.0 %|70.0 %|90.0 %|5.0 %
wyf_test_39| | 0.000 | | | | | 2.980 GB| |0.000 B | |193.715 GB|9.46 %| | | |
Number of rows: 2
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~Namespace Object Information (2024-06-20 06:10:36 UTC)~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Namespace| Node|Rack| Repl|Expirations| Total|~~~~~~~~~~Objects~~~~~~~~~~~|~~~~~~~~~Tombstones~~~~~~~~|~~~~Pending~~~~
| | ID|Factor| | Records| Master| Prole|Non-Replica| Master| Prole|Non-Replica|~~~~Migrates~~~
| | | | | | | | | | | | Tx| Rx
test_129 |testpc118129:3000 | 0| 1| 0.000 | 0.000 | 0.000 |0.000 | 0.000 |0.000 |0.000 | 0.000 |0.000 |0.000
test_129 | | | | 0.000 | 0.000 | 0.000 |0.000 | 0.000 |0.000 |0.000 | 0.000 |0.000 |0.000
wyf_test_39|testpc11839.samsungds.net:3000| 0| 1| 0.000 |50.000 M|50.000 M|0.000 | 0.000 |0.000 |0.000 | 0.000 |0.000 |0.000
wyf_test_39| | | | 0.000 |50.000 M|50.000 M|0.000 | 0.000 |0.000 |0.000 | 0.000 |0.000 |0.000
Number of rows: 2
It seems like no replica did. Because the size of storage device on node2 didn’t change any. Why the “Rel Factor” here is 1? I got the log from /var/log/aerospike/aerospike.log, it shows the replication factor is 1 too. Can anyone tell me why?
Jun 20 2024 05:38:48 GMT: INFO (partition): (partition_balance.c:800) {wyf_test_39} is on 1 of 2 nodes
Jun 20 2024 05:38:48 GMT: INFO (partition): (partition_balance.c:1038) {wyf_test_39} replication factor is 1
Jun 20 2024 05:38:48 GMT: INFO (partition): (partition_balance.c:1011) {wyf_test_39} rebalanced: expected-migrations (0,0,0) fresh-partitions 4096