Files
SugarDB/echovault/echovault_test.go
2024-07-01 06:05:46 +08:00

1093 lines
26 KiB
Go

// Copyright 2024 Kelvin Clement Mwinuka
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
package echovault
import (
"bufio"
"context"
"crypto/tls"
"crypto/x509"
"fmt"
"github.com/echovault/echovault/internal"
"github.com/echovault/echovault/internal/clock"
"github.com/echovault/echovault/internal/config"
"github.com/echovault/echovault/internal/constants"
"github.com/tidwall/resp"
"io"
"math"
"net"
"os"
"path"
"strings"
"sync"
"testing"
"time"
)
type ClientServerPair struct {
dataDir string
serverId string
bindAddr string
port int
discoveryPort int
bootstrapCluster bool
forwardCommand bool
joinAddr string
raw net.Conn
client *resp.Conn
server *EchoVault
}
var bindLock sync.Mutex
var bindNum byte = 10
func getBindAddrNet(network byte) net.IP {
bindLock.Lock()
defer bindLock.Unlock()
result := net.IPv4(127, 0, network, bindNum)
bindNum++
if bindNum > 255 {
bindNum = 10
}
return result
}
func getBindAddr() net.IP {
return getBindAddrNet(0)
}
func setupServer(
dataDir string,
serverId string,
bootstrapCluster bool,
forwardCommand bool,
bindAddr,
joinAddr string,
port,
discoveryPort int,
) (*EchoVault, error) {
conf := DefaultConfig()
conf.DataDir = dataDir
conf.ForwardCommand = forwardCommand
conf.BindAddr = bindAddr
conf.JoinAddr = joinAddr
conf.Port = uint16(port)
conf.ServerID = serverId
conf.DiscoveryPort = uint16(discoveryPort)
conf.BootstrapCluster = bootstrapCluster
conf.EvictionPolicy = constants.NoEviction
return NewEchoVault(
WithContext(context.Background()),
WithConfig(conf),
)
}
func setupNode(node *ClientServerPair, isLeader bool, errChan *chan error) {
server, err := setupServer(
node.dataDir,
node.serverId,
node.bootstrapCluster,
node.forwardCommand,
node.bindAddr,
node.joinAddr,
node.port,
node.discoveryPort,
)
if err != nil {
*errChan <- fmt.Errorf("could not start server; %v", err)
}
// Start the server.
go func() {
server.Start()
}()
if isLeader {
// If node is a leader, wait until it's established itself as a leader of the raft cluster.
for {
if server.raft.IsRaftLeader() {
break
}
}
} else {
// If the node is a follower, wait until it's joined the raft cluster before moving forward.
for {
if server.raft.HasJoinedCluster() {
break
}
}
}
// Setup client connection.
conn, err := internal.GetConnection(node.bindAddr, node.port)
if err != nil {
*errChan <- fmt.Errorf("could not open tcp connection: %v", err)
}
client := resp.NewConn(conn)
node.raw = conn
node.client = client
node.server = server
}
func makeCluster(size int) ([]ClientServerPair, error) {
pairs := make([]ClientServerPair, size)
// Set up node metadata.
for i := 0; i < len(pairs); i++ {
dataDir := ""
serverId := fmt.Sprintf("SERVER-%d", i)
bindAddr := getBindAddr().String()
bootstrapCluster := i == 0
forwardCommand := i < len(pairs)-1 // The last node will not forward commands to the cluster leader.
joinAddr := ""
if !bootstrapCluster {
joinAddr = fmt.Sprintf("%s/%s:%d", pairs[0].serverId, pairs[0].bindAddr, pairs[0].discoveryPort)
}
port, err := internal.GetFreePort()
if err != nil {
return nil, fmt.Errorf("could not get free port: %v", err)
}
discoveryPort, err := internal.GetFreePort()
if err != nil {
return nil, fmt.Errorf("could not get free memberlist port: %v", err)
}
pairs[i] = ClientServerPair{
dataDir: dataDir,
serverId: serverId,
bindAddr: bindAddr,
port: port,
discoveryPort: discoveryPort,
bootstrapCluster: bootstrapCluster,
forwardCommand: forwardCommand,
joinAddr: joinAddr,
}
}
errChan := make(chan error)
doneChan := make(chan struct{})
// Set up nodes.
wg := sync.WaitGroup{}
for i := 0; i < len(pairs); i++ {
if i == 0 {
setupNode(&pairs[i], pairs[i].bootstrapCluster, &errChan)
continue
}
wg.Add(1)
go func(idx int) {
setupNode(&pairs[idx], pairs[idx].bootstrapCluster, &errChan)
wg.Done()
}(i)
}
go func() {
wg.Wait()
doneChan <- struct{}{}
}()
select {
case err := <-errChan:
return nil, err
case <-doneChan:
}
return pairs, nil
}
func Test_Cluster(t *testing.T) {
nodes, err := makeCluster(5)
if err != nil {
t.Error(err)
return
}
t.Cleanup(func() {
for i := len(nodes) - 1; i > -1; i-- {
_ = nodes[i].raw.Close()
nodes[i].server.ShutDown()
}
})
// Prepare the write data for the cluster.
tests := map[string][]struct {
key string
value string
}{
"replication": {
{key: "key1", value: "value1"},
{key: "key2", value: "value2"},
{key: "key3", value: "value3"},
},
"deletion": {
{key: "key4", value: "value4"},
{key: "key5", value: "value4"},
{key: "key6", value: "value5"},
},
"raft-apply-delete": {
{key: "key7", value: "value7"},
{key: "key8", value: "value8"},
{key: "key9", value: "value9"},
},
"forward": {
{key: "key10", value: "value10"},
{key: "key11", value: "value11"},
{key: "key12", value: "value12"},
},
}
t.Run("Test_Replication", func(t *testing.T) {
tests := tests["replication"]
// Write all the data to the cluster leader.
for i, test := range tests {
node := nodes[0]
if err := node.client.WriteArray([]resp.Value{
resp.StringValue("SET"), resp.StringValue(test.key), resp.StringValue(test.value),
}); err != nil {
t.Errorf("could not write data to leader node (test %d): %v", i, err)
}
// Read response and make sure we received "ok" response.
rd, _, err := node.client.ReadValue()
if err != nil {
t.Errorf("could not read response from leader node (test %d): %v", i, err)
}
if !strings.EqualFold(rd.String(), "ok") {
t.Errorf("expected response for test %d to be \"OK\", got %s", i, rd.String())
}
}
// Yield
ticker := time.NewTicker(200 * time.Millisecond)
defer func() {
ticker.Stop()
}()
<-ticker.C
// Check if the data has been replicated on a quorum (majority of the cluster).
quorum := int(math.Ceil(float64(len(nodes)/2)) + 1)
for i, test := range tests {
count := 0
for j := 0; j < len(nodes); j++ {
node := nodes[j]
if err := node.client.WriteArray([]resp.Value{
resp.StringValue("GET"),
resp.StringValue(test.key),
}); err != nil {
t.Errorf("could not write data to follower node %d (test %d): %v", j, i, err)
}
rd, _, err := node.client.ReadValue()
if err != nil {
t.Errorf("could not read data from follower node %d (test %d): %v", j, i, err)
}
if rd.String() == test.value {
count += 1 // If the expected value is found, increment the count.
}
}
// Fail if count is less than quorum.
if count < quorum {
t.Errorf("could not find value %s at key %s in cluster quorum", test.value, test.key)
}
}
})
t.Run("Test_DeleteKey", func(t *testing.T) {
tests := tests["deletion"]
// Write all the data to the cluster leader.
for i, test := range tests {
node := nodes[0]
_, ok, err := node.server.Set(test.key, test.value, SetOptions{})
if err != nil {
t.Errorf("could not write command to leader node (test %d): %v", i, err)
}
if !ok {
t.Errorf("expected set for test %d ok = true, got ok = false", i)
}
}
// Yield
ticker := time.NewTicker(200 * time.Millisecond)
defer func() {
ticker.Stop()
}()
<-ticker.C
// Check if the data has been replicated on a quorum (majority of the cluster).
quorum := int(math.Ceil(float64(len(nodes)/2)) + 1)
for i, test := range tests {
count := 0
for j := 0; j < len(nodes); j++ {
node := nodes[j]
if err := node.client.WriteArray([]resp.Value{
resp.StringValue("GET"),
resp.StringValue(test.key),
}); err != nil {
t.Errorf("could not write command to follower node %d (test %d): %v", j, i, err)
}
rd, _, err := node.client.ReadValue()
if err != nil {
t.Errorf("could not read data from follower node %d (test %d): %v", j, i, err)
}
if rd.String() == test.value {
count += 1 // If the expected value is found, increment the count.
}
}
// Fail if count is less than quorum.
if count < quorum {
t.Errorf("could not find value %s at key %s in cluster quorum", test.value, test.key)
return
}
}
// Delete the key on the leader node
// 1. Prepare delete command.
command := []resp.Value{resp.StringValue("DEL")}
for _, test := range tests {
command = append(command, resp.StringValue(test.key))
}
// 2. Send delete command.
if err := nodes[0].client.WriteArray(command); err != nil {
t.Error(err)
return
}
res, _, err := nodes[0].client.ReadValue()
if err != nil {
t.Error(err)
return
}
// 3. Check the delete count is equal to length of tests.
if res.Integer() != len(tests) {
t.Errorf("expected delete response to be %d, got %d", len(tests), res.Integer())
}
// Yield
ticker.Reset(200 * time.Millisecond)
<-ticker.C
// 4. Check if the data is absent in quorum (majority of the cluster).
for i, test := range tests {
count := 0
for j := 0; j < len(nodes); j++ {
node := nodes[j]
if err := node.client.WriteArray([]resp.Value{
resp.StringValue("GET"),
resp.StringValue(test.key),
}); err != nil {
t.Errorf("could not write command to follower node %d (test %d): %v", j, i, err)
}
rd, _, err := node.client.ReadValue()
if err != nil {
t.Errorf("could not read data from follower node %d (test %d): %v", j, i, err)
}
if rd.IsNull() {
count += 1 // If the expected value is found, increment the count.
}
}
// 5. Fail if count is less than quorum.
if count < quorum {
t.Errorf("could not find value %s at key %s in cluster quorum", test.value, test.key)
}
}
})
t.Run("Test_raftApplyDeleteKey", func(t *testing.T) {
tests := tests["raft-apply-delete"]
// Write all the data to the cluster leader.
for i, test := range tests {
node := nodes[0]
_, ok, err := node.server.Set(test.key, test.value, SetOptions{})
if err != nil {
t.Errorf("could not write command to leader node (test %d): %v", i, err)
}
if !ok {
t.Errorf("expected set for test %d ok = true, got ok = false", i)
}
}
// Yield
ticker := time.NewTicker(200 * time.Millisecond)
defer func() {
ticker.Stop()
}()
<-ticker.C
// Check if the data has been replicated on a quorum (majority of the cluster).
quorum := int(math.Ceil(float64(len(nodes)/2)) + 1)
for i, test := range tests {
count := 0
for j := 0; j < len(nodes); j++ {
node := nodes[j]
if err := node.client.WriteArray([]resp.Value{
resp.StringValue("GET"),
resp.StringValue(test.key),
}); err != nil {
t.Errorf("could not write command to follower node %d (test %d): %v", j, i, err)
}
rd, _, err := node.client.ReadValue()
if err != nil {
t.Errorf("could not read data from follower node %d (test %d): %v", j, i, err)
}
if rd.String() == test.value {
count += 1 // If the expected value is found, increment the count.
}
}
// Fail if count is less than quorum.
if count < quorum {
t.Errorf("could not find value %s at key %s in cluster quorum", test.value, test.key)
return
}
}
// Delete the keys using raftApplyDelete method.
for _, test := range tests {
if err := nodes[0].server.raftApplyDeleteKey(nodes[0].server.context, test.key); err != nil {
t.Error(err)
}
}
// Yield to give key deletion time to take effect across cluster.
ticker.Reset(200 * time.Millisecond)
<-ticker.C
// Check if the data is absent in quorum (majority of the cluster).
for i, test := range tests {
count := 0
for j := 0; j < len(nodes); j++ {
node := nodes[j]
if err := node.client.WriteArray([]resp.Value{
resp.StringValue("GET"),
resp.StringValue(test.key),
}); err != nil {
t.Errorf("could not write command to follower node %d (test %d): %v", j, i, err)
}
rd, _, err := node.client.ReadValue()
if err != nil {
t.Errorf("could not read data from follower node %d (test %d): %v", j, i, err)
}
if rd.IsNull() {
count += 1 // If the expected value is found, increment the count.
}
}
// Fail if count is less than quorum.
if count < quorum {
t.Errorf("found value %s at key %s in cluster quorum", test.value, test.key)
}
}
})
t.Run("Test_ForwardCommand", func(t *testing.T) {
tests := tests["forward"]
// Write all the data a random cluster follower.
for i, test := range tests {
// Send write command to follower node.
node := nodes[1]
if err := node.client.WriteArray([]resp.Value{
resp.StringValue("SET"),
resp.StringValue(test.key),
resp.StringValue(test.value),
}); err != nil {
t.Errorf("could not write data to follower node (test %d): %v", i, err)
}
// Read response and make sure we received "ok" response.
rd, _, err := node.client.ReadValue()
if err != nil {
t.Errorf("could not read response from follower node (test %d): %v", i, err)
}
if !strings.EqualFold(rd.String(), "ok") {
t.Errorf("expected response for test %d to be \"OK\", got %s", i, rd.String())
}
}
ticker := time.NewTicker(1 * time.Second)
defer func() {
ticker.Stop()
}()
<-ticker.C
// Check if the data has been replicated on a quorum (majority of the cluster).
var forwardError error
doneChan := make(chan struct{})
go func() {
quorum := int(math.Ceil(float64(len(nodes)/2)) + 1)
for i := 0; i < len(tests); i++ {
test := tests[i]
count := 0
for j := 0; j < len(nodes); j++ {
node := nodes[j]
if err := node.client.WriteArray([]resp.Value{
resp.StringValue("GET"),
resp.StringValue(test.key),
}); err != nil {
forwardError = fmt.Errorf("could not write data to follower node %d (test %d): %v", j, i, err)
i = 0
continue
}
rd, _, err := node.client.ReadValue()
if err != nil {
forwardError = fmt.Errorf("could not read data from follower node %d (test %d): %v", j, i, err)
i = 0
continue
}
if rd.String() == test.value {
count += 1 // If the expected value is found, increment the count.
}
}
// Fail if count is less than quorum.
if count < quorum {
forwardError = fmt.Errorf("could not find value %s at key %s in cluster quorum", test.value, test.key)
i = 0
continue
}
}
doneChan <- struct{}{}
}()
ticker.Reset(5 * time.Second)
select {
case <-ticker.C:
if forwardError != nil {
t.Errorf("timeout error: %v\n", forwardError)
}
return
case <-doneChan:
}
})
t.Run("Test_NotLeaderError", func(t *testing.T) {
node := nodes[len(nodes)-1]
err := node.client.WriteArray([]resp.Value{
resp.StringValue("SET"),
resp.StringValue("key"),
resp.StringValue("value"),
})
if err != nil {
t.Error(err)
return
}
res, _, err := node.client.ReadValue()
if err != nil {
t.Error(err)
return
}
expected := "not cluster leader, cannot carry out command"
if !strings.Contains(res.Error().Error(), expected) {
t.Errorf("expected response to contain \"%s\", got \"%s\"", expected, res.Error().Error())
}
})
}
func Test_Standalone(t *testing.T) {
port, err := internal.GetFreePort()
if err != nil {
t.Error(err)
return
}
mockServer, err := NewEchoVault(
WithConfig(config.Config{
BindAddr: "localhost",
Port: uint16(port),
DataDir: "",
EvictionPolicy: constants.NoEviction,
}),
)
if err != nil {
t.Error(err)
return
}
go func() {
mockServer.Start()
}()
t.Cleanup(func() {
mockServer.ShutDown()
})
t.Run("Test_EmptyCommand", func(t *testing.T) {
conn, err := internal.GetConnection("localhost", port)
if err != nil {
t.Error(err)
return
}
defer func() {
_ = conn.Close()
}()
client := resp.NewConn(conn)
if err := client.WriteArray([]resp.Value{}); err != nil {
t.Error(err)
return
}
res, _, err := client.ReadValue()
if err != nil {
t.Error(err)
}
expected := "empty command"
if !strings.Contains(res.Error().Error(), expected) {
t.Errorf("expcted response to contain \"%s\", got \"%s\"", expected, res.Error().Error())
}
})
t.Run("Test_TLS", func(t *testing.T) {
t.Parallel()
port, err := internal.GetFreePort()
if err != nil {
t.Error(err)
return
}
conf := DefaultConfig()
conf.DataDir = ""
conf.BindAddr = "localhost"
conf.Port = uint16(port)
conf.TLS = true
conf.CertKeyPairs = [][]string{
{
path.Join("..", "openssl", "server", "server1.crt"),
path.Join("..", "openssl", "server", "server1.key"),
},
{
path.Join("..", "openssl", "server", "server2.crt"),
path.Join("..", "openssl", "server", "server2.key"),
},
}
server, err := NewEchoVault(WithConfig(conf))
if err != nil {
t.Error(err)
return
}
wg := sync.WaitGroup{}
wg.Add(1)
go func() {
wg.Done()
server.Start()
}()
wg.Wait()
// Dial with ServerCAs
serverCAs := x509.NewCertPool()
f, err := os.Open(path.Join("..", "openssl", "server", "rootCA.crt"))
if err != nil {
t.Error(err)
}
cert, err := io.ReadAll(bufio.NewReader(f))
if err != nil {
t.Error(err)
}
ok := serverCAs.AppendCertsFromPEM(cert)
if !ok {
t.Error("could not load server CA")
}
conn, err := internal.GetTLSConnection("localhost", port, &tls.Config{
RootCAs: serverCAs,
})
if err != nil {
t.Error(err)
return
}
defer func() {
_ = conn.Close()
server.ShutDown()
}()
client := resp.NewConn(conn)
// Test that we can set and get a value from the server.
key := "key1"
value := "value1"
err = client.WriteArray([]resp.Value{
resp.StringValue("SET"), resp.StringValue(key), resp.StringValue(value),
})
if err != nil {
t.Error(err)
}
res, _, err := client.ReadValue()
if err != nil {
t.Error(err)
}
if !strings.EqualFold(res.String(), "ok") {
t.Errorf("expected response OK, got \"%s\"", res.String())
}
err = client.WriteArray([]resp.Value{resp.StringValue("GET"), resp.StringValue(key)})
if err != nil {
t.Error(err)
}
res, _, err = client.ReadValue()
if err != nil {
t.Error(err)
}
if res.String() != value {
t.Errorf("expected response at key \"%s\" to be \"%s\", got \"%s\"", key, value, res.String())
}
})
t.Run("Test_MTLS", func(t *testing.T) {
t.Parallel()
port, err := internal.GetFreePort()
if err != nil {
t.Error(err)
return
}
conf := DefaultConfig()
conf.DataDir = ""
conf.BindAddr = "localhost"
conf.Port = uint16(port)
conf.TLS = true
conf.MTLS = true
conf.ClientCAs = []string{
path.Join("..", "openssl", "client", "rootCA.crt"),
}
conf.CertKeyPairs = [][]string{
{
path.Join("..", "openssl", "server", "server1.crt"),
path.Join("..", "openssl", "server", "server1.key"),
},
{
path.Join("..", "openssl", "server", "server2.crt"),
path.Join("..", "openssl", "server", "server2.key"),
},
}
server, err := NewEchoVault(WithConfig(conf))
if err != nil {
t.Error(err)
return
}
wg := sync.WaitGroup{}
wg.Add(1)
go func() {
wg.Done()
server.Start()
}()
wg.Wait()
// Dial with ServerCAs and client certificates
clientCertKeyPairs := [][]string{
{
path.Join("..", "openssl", "client", "client1.crt"),
path.Join("..", "openssl", "client", "client1.key"),
},
{
path.Join("..", "openssl", "client", "client2.crt"),
path.Join("..", "openssl", "client", "client2.key"),
},
}
var certificates []tls.Certificate
for _, pair := range clientCertKeyPairs {
c, err := tls.LoadX509KeyPair(pair[0], pair[1])
if err != nil {
t.Error(err)
}
certificates = append(certificates, c)
}
serverCAs := x509.NewCertPool()
f, err := os.Open(path.Join("..", "openssl", "server", "rootCA.crt"))
if err != nil {
t.Error(err)
}
cert, err := io.ReadAll(bufio.NewReader(f))
if err != nil {
t.Error(err)
}
ok := serverCAs.AppendCertsFromPEM(cert)
if !ok {
t.Error("could not load server CA")
}
conn, err := internal.GetTLSConnection("localhost", port, &tls.Config{
RootCAs: serverCAs,
Certificates: certificates,
})
if err != nil {
t.Error(err)
return
}
defer func() {
_ = conn.Close()
server.ShutDown()
}()
client := resp.NewConn(conn)
// Test that we can set and get a value from the server.
key := "key1"
value := "value1"
err = client.WriteArray([]resp.Value{
resp.StringValue("SET"), resp.StringValue(key), resp.StringValue(value),
})
if err != nil {
t.Error(err)
}
res, _, err := client.ReadValue()
if err != nil {
t.Error(err)
}
if !strings.EqualFold(res.String(), "ok") {
t.Errorf("expected response OK, got \"%s\"", res.String())
}
err = client.WriteArray([]resp.Value{resp.StringValue("GET"), resp.StringValue(key)})
if err != nil {
t.Error(err)
}
res, _, err = client.ReadValue()
if err != nil {
t.Error(err)
}
if res.String() != value {
t.Errorf("expected response at key \"%s\" to be \"%s\", got \"%s\"", key, value, res.String())
}
})
t.Run("Test_SnapshotRestore", func(t *testing.T) {
t.Parallel()
dataDir := path.Join(".", "testdata", "test_snapshot")
t.Cleanup(func() {
_ = os.RemoveAll(dataDir)
})
tests := []struct {
name string
dataDir string
values map[int]map[string]string
snapshotFunc func(mockServer *EchoVault) error
lastSaveFunc func(mockServer *EchoVault) (int, error)
wantLastSave int
}{
{
name: "1. Snapshot in embedded instance",
dataDir: path.Join(dataDir, "embedded_instance"),
values: map[int]map[string]string{
0: {"key5": "value-05", "key6": "value-06", "key7": "value-07", "key8": "value-08"},
1: {"key5": "value-15", "key6": "value-16", "key7": "value-17", "key8": "value-18"},
},
snapshotFunc: func(mockServer *EchoVault) error {
if _, err := mockServer.Save(); err != nil {
return err
}
return nil
},
lastSaveFunc: func(mockServer *EchoVault) (int, error) {
return mockServer.LastSave()
},
wantLastSave: int(clock.NewClock().Now().UnixMilli()),
},
}
for _, test := range tests {
t.Run(test.name, func(t *testing.T) {
t.Parallel()
port, err := internal.GetFreePort()
if err != nil {
t.Error(err)
return
}
conf := DefaultConfig()
conf.DataDir = test.dataDir
conf.BindAddr = "localhost"
conf.Port = uint16(port)
conf.RestoreSnapshot = true
mockServer, err := NewEchoVault(WithConfig(conf))
if err != nil {
t.Error(err)
return
}
defer func() {
// Shutdown
mockServer.ShutDown()
}()
// Trigger some write commands
for database, data := range test.values {
_ = mockServer.SelectDB(database)
for key, value := range data {
if _, _, err = mockServer.Set(key, value, SetOptions{}); err != nil {
t.Error(err)
return
}
}
}
// Function to trigger snapshot save
if err = test.snapshotFunc(mockServer); err != nil {
t.Error(err)
}
// Yield to allow snapshot to complete sync.
ticker := time.NewTicker(20 * time.Millisecond)
<-ticker.C
ticker.Stop()
// Restart server with the same config. This should restore the snapshot
mockServer, err = NewEchoVault(WithConfig(conf))
if err != nil {
t.Error(err)
return
}
// Check that all the key/value pairs have been restored into the store.
for database, data := range test.values {
_ = mockServer.SelectDB(database)
for key, value := range data {
res, err := mockServer.Get(key)
if err != nil {
t.Error(err)
return
}
if res != value {
t.Errorf("expected value at key \"%s\" to be \"%s\", got \"%s\"", key, value, res)
return
}
}
}
// Check that the lastsave is the time the last snapshot was taken.
lastSave, err := test.lastSaveFunc(mockServer)
if err != nil {
t.Error(err)
return
}
if lastSave != test.wantLastSave {
t.Errorf("expected lastsave to be %d, got %d", test.wantLastSave, lastSave)
}
})
}
})
t.Run("Test_AOFRestore", func(t *testing.T) {
t.Parallel()
ticker := time.NewTicker(50 * time.Millisecond)
dataDir := path.Join(".", "testdata", "test_aof")
t.Cleanup(func() {
_ = os.RemoveAll(dataDir)
ticker.Stop()
})
// Prepare data for testing.
data := map[string]map[string]string{
"before-rewrite": {
"key1": "value1",
"key2": "value2",
"key3": "value3",
"key4": "value4",
},
"after-rewrite": {
"key3": "value3-updated",
"key4": "value4-updated",
"key5": "value5",
"key6": "value6",
},
"expected-values": {
"key1": "value1",
"key2": "value2",
"key3": "value3-updated",
"key4": "value4-updated",
"key5": "value5",
"key6": "value6",
},
}
conf := DefaultConfig()
conf.RestoreAOF = true
conf.DataDir = dataDir
conf.AOFSyncStrategy = "always"
mockServer, err := NewEchoVault(WithConfig(conf))
if err != nil {
t.Error(err)
return
}
// Perform write commands from "before-rewrite"
for key, value := range data["before-rewrite"] {
if _, _, err := mockServer.Set(key, value, SetOptions{}); err != nil {
t.Error(err)
return
}
}
// Yield
<-ticker.C
// Rewrite AOF
if _, err := mockServer.RewriteAOF(); err != nil {
t.Error(err)
return
}
// Perform write commands from "after-rewrite"
for key, value := range data["after-rewrite"] {
if _, _, err := mockServer.Set(key, value, SetOptions{}); err != nil {
t.Error(err)
return
}
}
// Yield
<-ticker.C
// Shutdown the EchoVault instance
mockServer.ShutDown()
// Start another instance of EchoVault
mockServer, err = NewEchoVault(WithConfig(conf))
if err != nil {
t.Error(err)
return
}
// Check if the servers contains the keys and values from "expected-values"
for key, value := range data["expected-values"] {
res, err := mockServer.Get(key)
if err != nil {
t.Error(err)
return
}
if res != value {
t.Errorf("expected value at key \"%s\" to be \"%s\", got \"%s\"", key, value, res)
return
}
}
})
}