You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
 
 
 
 
 
 

372 lines
9.3 KiB

package log_buffer
import (
"testing"
"time"
"github.com/seaweedfs/seaweedfs/weed/pb/filer_pb"
)
func TestReadMessagesAtOffset_EmptyBuffer(t *testing.T) {
lb := NewLogBuffer("test", time.Hour, nil, nil, func() {})
lb.hasOffsets = true
lb.bufferStartOffset = 0
lb.offset = 0 // Empty buffer
messages, nextOffset, hwm, endOfPartition, err := lb.ReadMessagesAtOffset(100, 10, 1024)
// Reading from future offset (100) when buffer is at 0
// Should return empty, no error
if err != nil {
t.Errorf("Expected no error for future offset, got %v", err)
}
if len(messages) != 0 {
t.Errorf("Expected 0 messages, got %d", len(messages))
}
if nextOffset != 100 {
t.Errorf("Expected nextOffset=100, got %d", nextOffset)
}
if !endOfPartition {
t.Error("Expected endOfPartition=true for future offset")
}
if hwm != 0 {
t.Errorf("Expected highWaterMark=0, got %d", hwm)
}
}
func TestReadMessagesAtOffset_SingleMessage(t *testing.T) {
lb := NewLogBuffer("test", time.Hour, nil, nil, func() {})
lb.hasOffsets = true
// Add a message
entry := &filer_pb.LogEntry{
TsNs: time.Now().UnixNano(),
Key: []byte("key1"),
Data: []byte("value1"),
Offset: 0,
}
lb.AddLogEntryToBuffer(entry)
// Read from offset 0
messages, nextOffset, _, endOfPartition, err := lb.ReadMessagesAtOffset(0, 10, 1024)
if err != nil {
t.Errorf("Expected no error, got %v", err)
}
if len(messages) != 1 {
t.Errorf("Expected 1 message, got %d", len(messages))
}
if nextOffset != 1 {
t.Errorf("Expected nextOffset=1, got %d", nextOffset)
}
if !endOfPartition {
t.Error("Expected endOfPartition=true after reading all messages")
}
if messages[0].Offset != 0 {
t.Errorf("Expected message offset=0, got %d", messages[0].Offset)
}
if string(messages[0].Key) != "key1" {
t.Errorf("Expected key='key1', got '%s'", string(messages[0].Key))
}
}
func TestReadMessagesAtOffset_MultipleMessages(t *testing.T) {
lb := NewLogBuffer("test", time.Hour, nil, nil, func() {})
lb.hasOffsets = true
// Add 5 messages
for i := 0; i < 5; i++ {
entry := &filer_pb.LogEntry{
TsNs: time.Now().UnixNano(),
Key: []byte("key"),
Data: []byte("value"),
Offset: int64(i),
}
lb.AddLogEntryToBuffer(entry)
}
// Read from offset 0, max 3 messages
messages, nextOffset, _, _, err := lb.ReadMessagesAtOffset(0, 3, 10240)
if err != nil {
t.Errorf("Expected no error, got %v", err)
}
if len(messages) != 3 {
t.Errorf("Expected 3 messages, got %d", len(messages))
}
if nextOffset != 3 {
t.Errorf("Expected nextOffset=3, got %d", nextOffset)
}
// Verify offsets are sequential
for i, msg := range messages {
if msg.Offset != int64(i) {
t.Errorf("Message %d: expected offset=%d, got %d", i, i, msg.Offset)
}
}
}
func TestReadMessagesAtOffset_StartFromMiddle(t *testing.T) {
lb := NewLogBuffer("test", time.Hour, nil, nil, func() {})
lb.hasOffsets = true
// Add 10 messages (0-9)
for i := 0; i < 10; i++ {
entry := &filer_pb.LogEntry{
TsNs: time.Now().UnixNano(),
Key: []byte("key"),
Data: []byte("value"),
Offset: int64(i),
}
lb.AddLogEntryToBuffer(entry)
}
// Read from offset 5
messages, nextOffset, _, _, err := lb.ReadMessagesAtOffset(5, 3, 10240)
if err != nil {
t.Errorf("Expected no error, got %v", err)
}
if len(messages) != 3 {
t.Errorf("Expected 3 messages, got %d", len(messages))
}
if nextOffset != 8 {
t.Errorf("Expected nextOffset=8, got %d", nextOffset)
}
// Verify we got messages 5, 6, 7
expectedOffsets := []int64{5, 6, 7}
for i, msg := range messages {
if msg.Offset != expectedOffsets[i] {
t.Errorf("Message %d: expected offset=%d, got %d", i, expectedOffsets[i], msg.Offset)
}
}
}
func TestReadMessagesAtOffset_MaxBytesLimit(t *testing.T) {
lb := NewLogBuffer("test", time.Hour, nil, nil, func() {})
lb.hasOffsets = true
// Add messages with 100 bytes each
for i := 0; i < 10; i++ {
entry := &filer_pb.LogEntry{
TsNs: time.Now().UnixNano(),
Key: []byte("key"),
Data: make([]byte, 100), // 100 bytes
Offset: int64(i),
}
lb.AddLogEntryToBuffer(entry)
}
// Request with max 250 bytes (should get ~2 messages)
messages, _, _, _, err := lb.ReadMessagesAtOffset(0, 100, 250)
if err != nil {
t.Errorf("Expected no error, got %v", err)
}
// Should get at least 1 message, but likely 2
if len(messages) == 0 {
t.Error("Expected at least 1 message")
}
if len(messages) > 3 {
t.Errorf("Expected max 3 messages with 250 byte limit, got %d", len(messages))
}
}
func TestReadMessagesAtOffset_ConcurrentReads(t *testing.T) {
lb := NewLogBuffer("test", time.Hour, nil, nil, func() {})
lb.hasOffsets = true
// Add 100 messages
for i := 0; i < 100; i++ {
entry := &filer_pb.LogEntry{
TsNs: time.Now().UnixNano(),
Key: []byte("key"),
Data: []byte("value"),
Offset: int64(i),
}
lb.AddLogEntryToBuffer(entry)
}
// Start 10 concurrent readers at different offsets
done := make(chan bool, 10)
for reader := 0; reader < 10; reader++ {
startOffset := int64(reader * 10)
go func(offset int64) {
messages, nextOffset, _, _, err := lb.ReadMessagesAtOffset(offset, 5, 10240)
if err != nil {
t.Errorf("Reader at offset %d: unexpected error: %v", offset, err)
}
if len(messages) != 5 {
t.Errorf("Reader at offset %d: expected 5 messages, got %d", offset, len(messages))
}
if nextOffset != offset+5 {
t.Errorf("Reader at offset %d: expected nextOffset=%d, got %d", offset, offset+5, nextOffset)
}
// Verify sequential offsets
for i, msg := range messages {
expectedOffset := offset + int64(i)
if msg.Offset != expectedOffset {
t.Errorf("Reader at offset %d: message %d has offset %d, expected %d",
offset, i, msg.Offset, expectedOffset)
}
}
done <- true
}(startOffset)
}
// Wait for all readers
for i := 0; i < 10; i++ {
<-done
}
}
func TestReadMessagesAtOffset_FutureOffset(t *testing.T) {
lb := NewLogBuffer("test", time.Hour, nil, nil, func() {})
lb.hasOffsets = true
// Add 5 messages (0-4)
for i := 0; i < 5; i++ {
entry := &filer_pb.LogEntry{
TsNs: time.Now().UnixNano(),
Key: []byte("key"),
Data: []byte("value"),
Offset: int64(i),
}
lb.AddLogEntryToBuffer(entry)
}
// Try to read from offset 10 (future)
messages, nextOffset, _, endOfPartition, err := lb.ReadMessagesAtOffset(10, 10, 10240)
if err != nil {
t.Errorf("Expected no error for future offset, got %v", err)
}
if len(messages) != 0 {
t.Errorf("Expected 0 messages for future offset, got %d", len(messages))
}
if nextOffset != 10 {
t.Errorf("Expected nextOffset=10, got %d", nextOffset)
}
if !endOfPartition {
t.Error("Expected endOfPartition=true for future offset")
}
}
func TestWaitForDataWithTimeout_DataAvailable(t *testing.T) {
lb := NewLogBuffer("test", time.Hour, nil, nil, func() {})
lb.hasOffsets = true
// Add message at offset 0
entry := &filer_pb.LogEntry{
TsNs: time.Now().UnixNano(),
Key: []byte("key"),
Data: []byte("value"),
Offset: 0,
}
lb.AddLogEntryToBuffer(entry)
// Wait for data at offset 0 (should return immediately)
dataAvailable := lb.WaitForDataWithTimeout(0, 100)
if !dataAvailable {
t.Error("Expected data to be available at offset 0")
}
}
func TestWaitForDataWithTimeout_NoData(t *testing.T) {
lb := NewLogBuffer("test", time.Hour, nil, nil, func() {})
lb.hasOffsets = true
lb.bufferStartOffset = 0
lb.offset = 0
// Don't add any messages, wait for offset 10
// Wait for data at offset 10 with short timeout
start := time.Now()
dataAvailable := lb.WaitForDataWithTimeout(10, 50)
elapsed := time.Since(start)
if dataAvailable {
t.Error("Expected no data to be available")
}
// Note: Actual wait time may be shorter if subscriber mechanism
// returns immediately. Just verify no data was returned.
t.Logf("Waited %v for timeout", elapsed)
}
func TestWaitForDataWithTimeout_DataArrives(t *testing.T) {
lb := NewLogBuffer("test", time.Hour, nil, nil, func() {})
lb.hasOffsets = true
// Start waiting in background
done := make(chan bool)
var dataAvailable bool
go func() {
dataAvailable = lb.WaitForDataWithTimeout(0, 500)
done <- true
}()
// Add data after 50ms
time.Sleep(50 * time.Millisecond)
entry := &filer_pb.LogEntry{
TsNs: time.Now().UnixNano(),
Key: []byte("key"),
Data: []byte("value"),
Offset: 0,
}
lb.AddLogEntryToBuffer(entry)
// Wait for result
<-done
if !dataAvailable {
t.Error("Expected data to become available after being added")
}
}
func TestGetHighWaterMark(t *testing.T) {
lb := NewLogBuffer("test", time.Hour, nil, nil, func() {})
lb.hasOffsets = true
// Initially should be 0
hwm := lb.GetHighWaterMark()
if hwm != 0 {
t.Errorf("Expected initial HWM=0, got %d", hwm)
}
// Add messages (offsets 0-4)
for i := 0; i < 5; i++ {
entry := &filer_pb.LogEntry{
TsNs: time.Now().UnixNano(),
Key: []byte("key"),
Data: []byte("value"),
Offset: int64(i),
}
lb.AddLogEntryToBuffer(entry)
}
// HWM should be 5 (next offset to write, not last written offset)
// This matches Kafka semantics where HWM = last offset + 1
hwm = lb.GetHighWaterMark()
if hwm != 5 {
t.Errorf("Expected HWM=5 after adding 5 messages (0-4), got %d", hwm)
}
}
func TestGetLogStartOffset(t *testing.T) {
lb := NewLogBuffer("test", time.Hour, nil, nil, func() {})
lb.hasOffsets = true
lb.bufferStartOffset = 10
lso := lb.GetLogStartOffset()
if lso != 10 {
t.Errorf("Expected LSO=10, got %d", lso)
}
}