You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
353 lines
9.1 KiB
353 lines
9.1 KiB
package log_buffer
|
|
|
|
import (
|
|
"sync"
|
|
"sync/atomic"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/seaweedfs/seaweedfs/weed/pb/filer_pb"
|
|
)
|
|
|
|
// TestConcurrentProducerConsumer simulates the integration test scenario:
|
|
// - One producer writing messages continuously
|
|
// - Multiple consumers reading from different offsets
|
|
// - Consumers reading sequentially (like Kafka consumers)
|
|
func TestConcurrentProducerConsumer(t *testing.T) {
|
|
lb := NewLogBuffer("integration-test", time.Hour, nil, nil, func() {})
|
|
lb.hasOffsets = true
|
|
|
|
const numMessages = 1000
|
|
const numConsumers = 2
|
|
const messagesPerConsumer = numMessages / numConsumers
|
|
|
|
// Start producer
|
|
producerDone := make(chan bool)
|
|
go func() {
|
|
for i := 0; i < numMessages; i++ {
|
|
entry := &filer_pb.LogEntry{
|
|
TsNs: time.Now().UnixNano(),
|
|
Key: []byte("key"),
|
|
Data: []byte("value"),
|
|
Offset: int64(i),
|
|
}
|
|
lb.AddLogEntryToBuffer(entry)
|
|
time.Sleep(1 * time.Millisecond) // Simulate production rate
|
|
}
|
|
producerDone <- true
|
|
}()
|
|
|
|
// Start consumers
|
|
consumerWg := sync.WaitGroup{}
|
|
consumerErrors := make(chan error, numConsumers)
|
|
consumedCounts := make([]int64, numConsumers)
|
|
|
|
for consumerID := 0; consumerID < numConsumers; consumerID++ {
|
|
consumerWg.Add(1)
|
|
go func(id int, startOffset int64, endOffset int64) {
|
|
defer consumerWg.Done()
|
|
|
|
currentOffset := startOffset
|
|
for currentOffset < endOffset {
|
|
// Read 10 messages at a time (like integration test)
|
|
messages, nextOffset, _, _, err := lb.ReadMessagesAtOffset(currentOffset, 10, 10240)
|
|
if err != nil {
|
|
consumerErrors <- err
|
|
return
|
|
}
|
|
|
|
if len(messages) == 0 {
|
|
// No data yet, wait a bit
|
|
time.Sleep(5 * time.Millisecond)
|
|
continue
|
|
}
|
|
|
|
// Count only messages in this consumer's assigned range
|
|
messagesInRange := 0
|
|
for i, msg := range messages {
|
|
if msg.Offset >= startOffset && msg.Offset < endOffset {
|
|
messagesInRange++
|
|
expectedOffset := currentOffset + int64(i)
|
|
if msg.Offset != expectedOffset {
|
|
t.Errorf("Consumer %d: Expected offset %d, got %d", id, expectedOffset, msg.Offset)
|
|
}
|
|
}
|
|
}
|
|
|
|
atomic.AddInt64(&consumedCounts[id], int64(messagesInRange))
|
|
currentOffset = nextOffset
|
|
}
|
|
}(consumerID, int64(consumerID*messagesPerConsumer), int64((consumerID+1)*messagesPerConsumer))
|
|
}
|
|
|
|
// Wait for producer to finish
|
|
<-producerDone
|
|
|
|
// Wait for consumers (with timeout)
|
|
done := make(chan bool)
|
|
go func() {
|
|
consumerWg.Wait()
|
|
done <- true
|
|
}()
|
|
|
|
select {
|
|
case <-done:
|
|
// Success
|
|
case err := <-consumerErrors:
|
|
t.Fatalf("Consumer error: %v", err)
|
|
case <-time.After(10 * time.Second):
|
|
t.Fatal("Timeout waiting for consumers to finish")
|
|
}
|
|
|
|
// Verify all messages were consumed
|
|
totalConsumed := int64(0)
|
|
for i, count := range consumedCounts {
|
|
t.Logf("Consumer %d consumed %d messages", i, count)
|
|
totalConsumed += count
|
|
}
|
|
|
|
if totalConsumed != numMessages {
|
|
t.Errorf("Expected to consume %d messages, but consumed %d", numMessages, totalConsumed)
|
|
}
|
|
}
|
|
|
|
// TestBackwardSeeksWhileProducing simulates consumer rebalancing where
|
|
// consumers seek backward to earlier offsets while producer is still writing
|
|
func TestBackwardSeeksWhileProducing(t *testing.T) {
|
|
lb := NewLogBuffer("backward-seek-test", time.Hour, nil, nil, func() {})
|
|
lb.hasOffsets = true
|
|
|
|
const numMessages = 500
|
|
const numSeeks = 10
|
|
|
|
// Start producer
|
|
producerDone := make(chan bool)
|
|
go func() {
|
|
for i := 0; i < numMessages; i++ {
|
|
entry := &filer_pb.LogEntry{
|
|
TsNs: time.Now().UnixNano(),
|
|
Key: []byte("key"),
|
|
Data: []byte("value"),
|
|
Offset: int64(i),
|
|
}
|
|
lb.AddLogEntryToBuffer(entry)
|
|
time.Sleep(1 * time.Millisecond)
|
|
}
|
|
producerDone <- true
|
|
}()
|
|
|
|
// Consumer that seeks backward periodically
|
|
consumerDone := make(chan bool)
|
|
readOffsets := make(map[int64]int) // Track how many times each offset was read
|
|
|
|
go func() {
|
|
currentOffset := int64(0)
|
|
seeksRemaining := numSeeks
|
|
|
|
for currentOffset < numMessages {
|
|
// Read some messages
|
|
messages, nextOffset, _, endOfPartition, err := lb.ReadMessagesAtOffset(currentOffset, 10, 10240)
|
|
if err != nil {
|
|
// For stateless reads, "offset out of range" means data not in memory yet
|
|
// This is expected when reading historical data or before production starts
|
|
time.Sleep(5 * time.Millisecond)
|
|
continue
|
|
}
|
|
|
|
if len(messages) == 0 {
|
|
// No data available yet or caught up to producer
|
|
if !endOfPartition {
|
|
// Data might be coming, wait
|
|
time.Sleep(5 * time.Millisecond)
|
|
} else {
|
|
// At end of partition, wait for more production
|
|
time.Sleep(5 * time.Millisecond)
|
|
}
|
|
continue
|
|
}
|
|
|
|
// Track read offsets
|
|
for _, msg := range messages {
|
|
readOffsets[msg.Offset]++
|
|
}
|
|
|
|
// Periodically seek backward (simulating rebalancing)
|
|
if seeksRemaining > 0 && nextOffset > 50 && nextOffset%100 == 0 {
|
|
seekOffset := nextOffset - 20
|
|
t.Logf("Seeking backward from %d to %d", nextOffset, seekOffset)
|
|
currentOffset = seekOffset
|
|
seeksRemaining--
|
|
} else {
|
|
currentOffset = nextOffset
|
|
}
|
|
}
|
|
|
|
consumerDone <- true
|
|
}()
|
|
|
|
// Wait for both
|
|
<-producerDone
|
|
<-consumerDone
|
|
|
|
// Verify each offset was read at least once
|
|
for i := int64(0); i < numMessages; i++ {
|
|
if readOffsets[i] == 0 {
|
|
t.Errorf("Offset %d was never read", i)
|
|
}
|
|
}
|
|
|
|
t.Logf("Total unique offsets read: %d out of %d", len(readOffsets), numMessages)
|
|
}
|
|
|
|
// TestHighConcurrencyReads simulates multiple consumers reading from
|
|
// different offsets simultaneously (stress test)
|
|
func TestHighConcurrencyReads(t *testing.T) {
|
|
lb := NewLogBuffer("high-concurrency-test", time.Hour, nil, nil, func() {})
|
|
lb.hasOffsets = true
|
|
|
|
const numMessages = 1000
|
|
const numReaders = 10
|
|
|
|
// Pre-populate buffer
|
|
for i := 0; i < numMessages; i++ {
|
|
entry := &filer_pb.LogEntry{
|
|
TsNs: time.Now().UnixNano(),
|
|
Key: []byte("key"),
|
|
Data: []byte("value"),
|
|
Offset: int64(i),
|
|
}
|
|
lb.AddLogEntryToBuffer(entry)
|
|
}
|
|
|
|
// Start many concurrent readers at different offsets
|
|
wg := sync.WaitGroup{}
|
|
errors := make(chan error, numReaders)
|
|
|
|
for reader := 0; reader < numReaders; reader++ {
|
|
wg.Add(1)
|
|
go func(startOffset int64) {
|
|
defer wg.Done()
|
|
|
|
// Read 100 messages from this offset
|
|
currentOffset := startOffset
|
|
readCount := 0
|
|
|
|
for readCount < 100 && currentOffset < numMessages {
|
|
messages, nextOffset, _, _, err := lb.ReadMessagesAtOffset(currentOffset, 10, 10240)
|
|
if err != nil {
|
|
errors <- err
|
|
return
|
|
}
|
|
|
|
// Verify offsets are sequential
|
|
for i, msg := range messages {
|
|
expected := currentOffset + int64(i)
|
|
if msg.Offset != expected {
|
|
t.Errorf("Reader at %d: expected offset %d, got %d", startOffset, expected, msg.Offset)
|
|
}
|
|
}
|
|
|
|
readCount += len(messages)
|
|
currentOffset = nextOffset
|
|
}
|
|
}(int64(reader * 10))
|
|
}
|
|
|
|
// Wait with timeout
|
|
done := make(chan bool)
|
|
go func() {
|
|
wg.Wait()
|
|
done <- true
|
|
}()
|
|
|
|
select {
|
|
case <-done:
|
|
// Success
|
|
case err := <-errors:
|
|
t.Fatalf("Reader error: %v", err)
|
|
case <-time.After(10 * time.Second):
|
|
t.Fatal("Timeout waiting for readers")
|
|
}
|
|
}
|
|
|
|
// TestRepeatedReadsAtSameOffset simulates what happens when Kafka
|
|
// consumer re-fetches the same offset multiple times (due to timeouts or retries)
|
|
func TestRepeatedReadsAtSameOffset(t *testing.T) {
|
|
lb := NewLogBuffer("repeated-reads-test", time.Hour, nil, nil, func() {})
|
|
lb.hasOffsets = true
|
|
|
|
const numMessages = 100
|
|
|
|
// Pre-populate buffer
|
|
for i := 0; i < numMessages; i++ {
|
|
entry := &filer_pb.LogEntry{
|
|
TsNs: time.Now().UnixNano(),
|
|
Key: []byte("key"),
|
|
Data: []byte("value"),
|
|
Offset: int64(i),
|
|
}
|
|
lb.AddLogEntryToBuffer(entry)
|
|
}
|
|
|
|
// Read the same offset multiple times concurrently
|
|
const numReads = 10
|
|
const testOffset = int64(50)
|
|
|
|
wg := sync.WaitGroup{}
|
|
results := make([][]*filer_pb.LogEntry, numReads)
|
|
|
|
for i := 0; i < numReads; i++ {
|
|
wg.Add(1)
|
|
go func(idx int) {
|
|
defer wg.Done()
|
|
messages, _, _, _, err := lb.ReadMessagesAtOffset(testOffset, 10, 10240)
|
|
if err != nil {
|
|
t.Errorf("Read %d error: %v", idx, err)
|
|
return
|
|
}
|
|
results[idx] = messages
|
|
}(i)
|
|
}
|
|
|
|
wg.Wait()
|
|
|
|
// Verify all reads returned the same data
|
|
firstRead := results[0]
|
|
for i := 1; i < numReads; i++ {
|
|
if len(results[i]) != len(firstRead) {
|
|
t.Errorf("Read %d returned %d messages, expected %d", i, len(results[i]), len(firstRead))
|
|
}
|
|
|
|
for j := range results[i] {
|
|
if results[i][j].Offset != firstRead[j].Offset {
|
|
t.Errorf("Read %d message %d has offset %d, expected %d",
|
|
i, j, results[i][j].Offset, firstRead[j].Offset)
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// TestEmptyPartitionPolling simulates consumers polling empty partitions
|
|
// waiting for data (common in Kafka)
|
|
func TestEmptyPartitionPolling(t *testing.T) {
|
|
lb := NewLogBuffer("empty-partition-test", time.Hour, nil, nil, func() {})
|
|
lb.hasOffsets = true
|
|
lb.bufferStartOffset = 0
|
|
lb.offset = 0
|
|
|
|
// Try to read from empty partition
|
|
messages, nextOffset, _, endOfPartition, err := lb.ReadMessagesAtOffset(0, 10, 10240)
|
|
|
|
if err != nil {
|
|
t.Errorf("Unexpected error: %v", err)
|
|
}
|
|
if len(messages) != 0 {
|
|
t.Errorf("Expected 0 messages, got %d", len(messages))
|
|
}
|
|
if nextOffset != 0 {
|
|
t.Errorf("Expected nextOffset=0, got %d", nextOffset)
|
|
}
|
|
if !endOfPartition {
|
|
t.Error("Expected endOfPartition=true for future offset")
|
|
}
|
|
}
|