timeout fix
Some checks failed
continuous-integration/drone/push Build is failing

This commit is contained in:
ycc
2026-03-01 14:23:07 +01:00
parent b722a916a9
commit 7d06f0ff3e
2 changed files with 125 additions and 116 deletions

View File

@@ -10,9 +10,11 @@ import (
"testing"
"time"
"forge.redroom.link/yves/meowlib"
"forge.redroom.link/yves/meowlib/client"
"github.com/stretchr/testify/assert"
"github.com/stretchr/testify/require"
"google.golang.org/protobuf/proto"
_ "github.com/mattn/go-sqlite3"
)
@@ -37,14 +39,35 @@ func closedServerURL(t *testing.T) string {
return srv.URL
}
// writeMsgFile writes dummy bytes to a temp file and returns the path.
// writeMsgFile writes a valid serialised empty PackedUserMessage to a temp file
// and returns the path. The file content satisfies proto.Unmarshal inside
// attemptSendJob; the httptest endpoints ignore the encrypted payload.
func writeMsgFile(t *testing.T, dir, name string) string {
t.Helper()
p := filepath.Join(dir, name)
require.NoError(t, os.WriteFile(p, []byte("packed-server-message"), 0600))
data, err := proto.Marshal(&meowlib.PackedUserMessage{})
require.NoError(t, err)
require.NoError(t, os.WriteFile(p, data, 0600))
return p
}
// newTestServer creates a client.Server for the given URL, generates a
// throwaway keypair so that AsymEncryptMessage succeeds, and stores the server
// in the current identity's MessageServers so that PackMessageForServer can
// look it up via LoadServer. Returns the registered server.
//
// Call setupMsgHelperConfig before this so an identity is in place.
func newTestServer(t *testing.T, url string) client.Server {
t.Helper()
srv, err := client.CreateServerFromUrl(url)
require.NoError(t, err)
kp, err := meowlib.NewKeyPair()
require.NoError(t, err)
srv.PublicKey = kp.Public
require.NoError(t, client.GetConfig().GetIdentity().MessageServers.StoreServer(srv))
return *srv
}
// pushJob is a convenience wrapper around client.PushSendJob.
func pushJob(t *testing.T, dir, queue, file string, servers []client.Server, timeout int) {
t.Helper()
@@ -67,50 +90,23 @@ func serverSlice(urls ...string) []client.Server {
// --- unit tests ---------------------------------------------------------
func TestAllServersExhausted_NoServers(t *testing.T) {
job := &client.SendJob{}
assert.True(t, allServersExhausted(job))
}
func TestAllServersExhausted_NoneExhausted(t *testing.T) {
job := &client.SendJob{
Servers: serverSlice("http://s1", "http://s2"),
Retries: []int{0, 0},
}
assert.False(t, allServersExhausted(job))
}
func TestAllServersExhausted_PartiallyExhausted(t *testing.T) {
job := &client.SendJob{
Servers: serverSlice("http://s1", "http://s2"),
Retries: []int{maxRetriesPerServer, 0},
}
assert.False(t, allServersExhausted(job))
}
func TestAllServersExhausted_AllExhausted(t *testing.T) {
job := &client.SendJob{
Servers: serverSlice("http://s1", "http://s2"),
Retries: []int{maxRetriesPerServer, maxRetriesPerServer},
}
assert.True(t, allServersExhausted(job))
}
// TestAttemptSendJob_Success verifies a successful POST to the first server.
func TestAttemptSendJob_Success(t *testing.T) {
dir := t.TempDir()
dir, _ := setupMsgHelperConfig(t)
var received int64
srv := acceptServer(t, &received)
defer srv.Close()
newTestServer(t, srv.URL)
job := &client.SendJob{
File: writeMsgFile(t, dir, "msg"),
Servers: serverSlice(srv.URL),
Timeout: 5,
Retries: []int{0},
}
retries := make([]int, len(job.Servers))
idx, err := attemptSendJob(job)
idx, err := attemptSendJob(job, retries)
require.NoError(t, err)
assert.Equal(t, 0, idx)
assert.Equal(t, int64(1), atomic.LoadInt64(&received))
@@ -119,62 +115,74 @@ func TestAttemptSendJob_Success(t *testing.T) {
// TestAttemptSendJob_Fallback verifies that when the first server refuses the
// connection, the second server is tried and succeeds.
func TestAttemptSendJob_Fallback(t *testing.T) {
dir := t.TempDir()
dir, _ := setupMsgHelperConfig(t)
var received int64
good := acceptServer(t, &received)
defer good.Close()
deadURL := closedServerURL(t)
newTestServer(t, deadURL)
newTestServer(t, good.URL)
job := &client.SendJob{
File: writeMsgFile(t, dir, "msg"),
Servers: serverSlice(closedServerURL(t), good.URL),
Servers: serverSlice(deadURL, good.URL),
Timeout: 5,
Retries: []int{0, 0},
}
retries := make([]int, len(job.Servers))
idx, err := attemptSendJob(job)
idx, err := attemptSendJob(job, retries)
require.NoError(t, err)
assert.Equal(t, 1, idx, "second server should have been used")
assert.Equal(t, int64(1), atomic.LoadInt64(&received))
assert.Equal(t, 1, job.Retries[0], "first server retry should be incremented")
assert.Equal(t, 0, job.Retries[1], "second server retry must stay at zero")
assert.Equal(t, 1, retries[0], "first server retry should be incremented")
assert.Equal(t, 0, retries[1], "second server retry must stay at zero")
}
// TestAttemptSendJob_AllFail verifies that all retry counts are incremented
// and an error is returned when every server refuses connections.
func TestAttemptSendJob_AllFail(t *testing.T) {
dir := t.TempDir()
dir, _ := setupMsgHelperConfig(t)
dead1 := closedServerURL(t)
dead2 := closedServerURL(t)
newTestServer(t, dead1)
newTestServer(t, dead2)
job := &client.SendJob{
File: writeMsgFile(t, dir, "msg"),
Servers: serverSlice(closedServerURL(t), closedServerURL(t)),
Servers: serverSlice(dead1, dead2),
Timeout: 5,
Retries: []int{0, 0},
}
retries := make([]int, len(job.Servers))
_, err := attemptSendJob(job)
_, err := attemptSendJob(job, retries)
assert.Error(t, err)
assert.Equal(t, 1, job.Retries[0])
assert.Equal(t, 1, job.Retries[1])
assert.Equal(t, 1, retries[0])
assert.Equal(t, 1, retries[1])
}
// TestAttemptSendJob_SkipsExhaustedServer verifies that a server already at
// maxRetriesPerServer is not contacted.
func TestAttemptSendJob_SkipsExhaustedServer(t *testing.T) {
dir := t.TempDir()
dir, _ := setupMsgHelperConfig(t)
var received int64
good := acceptServer(t, &received)
defer good.Close()
deadURL := closedServerURL(t)
newTestServer(t, good.URL) // only good server needs to be reachable
job := &client.SendJob{
File: writeMsgFile(t, dir, "msg"),
Servers: serverSlice(
closedServerURL(t), // exhausted must be skipped
deadURL, // exhausted must be skipped (no need to store in identity)
good.URL,
),
Timeout: 5,
Retries: []int{maxRetriesPerServer, 0},
}
retries := []int{maxRetriesPerServer, 0} // first server already exhausted this run
idx, err := attemptSendJob(job)
idx, err := attemptSendJob(job, retries)
require.NoError(t, err)
assert.Equal(t, 1, idx)
assert.Equal(t, int64(1), atomic.LoadInt64(&received))
@@ -201,11 +209,13 @@ func TestWriteSendJob(t *testing.T) {
// TestProcessSendQueues_Success verifies that a pending job is delivered and
// marked as sent when the server accepts it.
func TestProcessSendQueues_Success(t *testing.T) {
dir := t.TempDir()
dir, _ := setupMsgHelperConfig(t)
var received int64
srv := acceptServer(t, &received)
defer srv.Close()
newTestServer(t, srv.URL)
msgPath := writeMsgFile(t, dir, "msg")
pushJob(t, dir, "q1", msgPath, serverSlice(srv.URL), 10)
@@ -229,13 +239,17 @@ func TestProcessSendQueues_Success(t *testing.T) {
// TestProcessSendQueues_ServerFallback verifies that when the first server is
// unreachable, the second server is tried successfully in the same pass.
func TestProcessSendQueues_ServerFallback(t *testing.T) {
dir := t.TempDir()
dir, _ := setupMsgHelperConfig(t)
var received int64
good := acceptServer(t, &received)
defer good.Close()
deadURL := closedServerURL(t)
newTestServer(t, deadURL)
newTestServer(t, good.URL)
msgPath := writeMsgFile(t, dir, "msg")
pushJob(t, dir, "q1", msgPath, serverSlice(closedServerURL(t), good.URL), 10)
pushJob(t, dir, "q1", msgPath, serverSlice(deadURL, good.URL), 10)
_, id, err := client.PeekSendJob(dir, "q1")
require.NoError(t, err)
@@ -252,39 +266,43 @@ func TestProcessSendQueues_ServerFallback(t *testing.T) {
assert.Equal(t, 1, *job.SuccessfulServer, "second server should be recorded as successful")
}
// TestProcessSendQueues_AllServersExhausted verifies that after maxRetriesPerServer
// failed attempts per server the job is marked as failed.
func TestProcessSendQueues_AllServersExhausted(t *testing.T) {
dir := t.TempDir()
// TestProcessSendQueues_FailedRunsStayPending verifies that repeated delivery
// failures do NOT mark a job as permanently failed. Only a TTL timeout can do
// that; retry exhaustion merely stops the current run.
func TestProcessSendQueues_FailedRunsStayPending(t *testing.T) {
dir, _ := setupMsgHelperConfig(t)
deadURL := closedServerURL(t)
newTestServer(t, deadURL)
msgPath := writeMsgFile(t, dir, "msg")
// timeout=0 → uses defaultSendTimeout (24 h), so the job won't expire here.
pushJob(t, dir, "q1", msgPath, serverSlice(deadURL), 0)
_, id, err := client.PeekSendJob(dir, "q1")
require.NoError(t, err)
// Each call to ProcessSendQueues increments the retry counter by 1.
// After maxRetriesPerServer calls, all servers are exhausted → failed.
for i := 0; i < maxRetriesPerServer; i++ {
// Run several times per-server retry counts reset each run, so the job
// must remain pending no matter how many runs fail.
for i := 0; i < maxRetriesPerServer+2; i++ {
ProcessSendQueues(dir)
}
job, err := client.GetSendJob(dir, "q1", id)
require.NoError(t, err)
require.NotNil(t, job)
assert.Equal(t, client.SendStatusFailed, job.Status)
assert.Equal(t, maxRetriesPerServer, job.Retries[0])
assert.Equal(t, client.SendStatusPending, job.Status, "repeated failures must not cause permanent failure only timeout does")
}
// TestProcessSendQueues_JobTimeout verifies that a job whose timeout has elapsed
// is immediately marked as failed without any send attempt.
func TestProcessSendQueues_JobTimeout(t *testing.T) {
dir := t.TempDir()
dir, _ := setupMsgHelperConfig(t)
var received int64
srv := acceptServer(t, &received)
defer srv.Close()
newTestServer(t, srv.URL)
msgPath := writeMsgFile(t, dir, "msg")
// Timeout of 1 second; we will backdate inserted_at so the job looks expired.
pushJob(t, dir, "q1", msgPath, serverSlice(srv.URL), 1)
@@ -309,17 +327,25 @@ func TestProcessSendQueues_JobTimeout(t *testing.T) {
// TestProcessSendQueues_MultipleQueues verifies that jobs in different queue
// files are processed concurrently and independently.
func TestProcessSendQueues_MultipleQueues(t *testing.T) {
dir := t.TempDir()
dir, _ := setupMsgHelperConfig(t)
var received int64
srv := acceptServer(t, &received)
defer srv.Close()
newTestServer(t, srv.URL)
for _, q := range []string{"qa", "qb", "qc"} {
msgPath := writeMsgFile(t, dir, "msg_"+q)
pushJob(t, dir, q, msgPath, serverSlice(srv.URL), 10)
}
ProcessSendQueues(dir)
// Concurrent goroutines for each queue all try to open the same BadgerDB for
// server lookup; only one can hold the lock at a time. Jobs that lose the
// race stay pending and are retried on the next call. Three passes guarantee
// every queue gets at least one uncontested turn.
for i := 0; i < 3; i++ {
ProcessSendQueues(dir)
}
assert.Equal(t, int64(3), atomic.LoadInt64(&received), "all three queues should have delivered their message")
}