1
0
Fork 0
telegraf/plugins/outputs/kinesis/kinesis_test.go
Daniel Baumann 4978089aab
Adding upstream version 1.34.4.
Signed-off-by: Daniel Baumann <daniel@debian.org>
2025-05-24 07:26:29 +02:00

596 lines
14 KiB
Go

package kinesis
import (
"context"
"fmt"
"testing"
"github.com/aws/aws-sdk-go-v2/aws"
"github.com/aws/aws-sdk-go-v2/service/kinesis"
"github.com/aws/aws-sdk-go-v2/service/kinesis/types"
"github.com/gofrs/uuid/v5"
"github.com/stretchr/testify/require"
"github.com/influxdata/telegraf"
"github.com/influxdata/telegraf/plugins/serializers/influx"
"github.com/influxdata/telegraf/testutil"
)
const testPartitionKey = "partitionKey"
const testShardID = "shardId-000000000003"
const testSequenceNumber = "49543463076570308322303623326179887152428262250726293588"
const testStreamName = "streamName"
const zero int64 = 0
func TestPartitionKey(t *testing.T) {
testPoint := testutil.TestMetric(1)
k := KinesisOutput{
Log: testutil.Logger{},
Partition: &Partition{
Method: "static",
Key: "-",
},
}
require.Equal(t, "-", k.getPartitionKey(testPoint), "PartitionKey should be '-'")
k = KinesisOutput{
Log: testutil.Logger{},
Partition: &Partition{
Method: "tag",
Key: "tag1",
},
}
require.Equal(t, testPoint.Tags()["tag1"], k.getPartitionKey(testPoint), "PartitionKey should be value of 'tag1'")
k = KinesisOutput{
Log: testutil.Logger{},
Partition: &Partition{
Method: "tag",
Key: "doesnotexist",
Default: "somedefault",
},
}
require.Equal(t, "somedefault", k.getPartitionKey(testPoint), "PartitionKey should use default")
k = KinesisOutput{
Log: testutil.Logger{},
Partition: &Partition{
Method: "tag",
Key: "doesnotexist",
},
}
require.Equal(t, "telegraf", k.getPartitionKey(testPoint), "PartitionKey should be telegraf")
k = KinesisOutput{
Log: testutil.Logger{},
Partition: &Partition{
Method: "not supported",
},
}
require.Empty(t, k.getPartitionKey(testPoint), "PartitionKey should be value of ''")
k = KinesisOutput{
Log: testutil.Logger{},
Partition: &Partition{
Method: "measurement",
},
}
require.Equal(t, testPoint.Name(), k.getPartitionKey(testPoint), "PartitionKey should be value of measurement name")
k = KinesisOutput{
Log: testutil.Logger{},
Partition: &Partition{
Method: "random",
},
}
partitionKey := k.getPartitionKey(testPoint)
u, err := uuid.FromString(partitionKey)
require.NoError(t, err, "Issue parsing UUID")
require.Equal(t, byte(4), u.Version(), "PartitionKey should be UUIDv4")
k = KinesisOutput{
Log: testutil.Logger{},
PartitionKey: "-",
}
require.Equal(t, "-", k.getPartitionKey(testPoint), "PartitionKey should be '-'")
k = KinesisOutput{
Log: testutil.Logger{},
RandomPartitionKey: true,
}
partitionKey = k.getPartitionKey(testPoint)
u, err = uuid.FromString(partitionKey)
require.NoError(t, err, "Issue parsing UUID")
require.Equal(t, byte(4), u.Version(), "PartitionKey should be UUIDv4")
}
func TestWriteKinesis_WhenSuccess(t *testing.T) {
records := []types.PutRecordsRequestEntry{
{
PartitionKey: aws.String(testPartitionKey),
Data: []byte{0x65},
},
}
svc := &mockKinesisPutRecords{}
svc.SetupResponse(
0,
[]types.PutRecordsResultEntry{
{
SequenceNumber: aws.String(testSequenceNumber),
ShardId: aws.String(testShardID),
},
},
)
k := KinesisOutput{
Log: testutil.Logger{},
StreamName: testStreamName,
svc: svc,
}
elapsed := k.writeKinesis(records)
require.GreaterOrEqual(t, elapsed.Nanoseconds(), zero)
svc.AssertRequests(t, []*kinesis.PutRecordsInput{
{
StreamName: aws.String(testStreamName),
Records: records,
},
})
}
func TestWriteKinesis_WhenRecordErrors(t *testing.T) {
records := []types.PutRecordsRequestEntry{
{
PartitionKey: aws.String(testPartitionKey),
Data: []byte{0x66},
},
}
svc := &mockKinesisPutRecords{}
svc.SetupResponse(
1,
[]types.PutRecordsResultEntry{
{
ErrorCode: aws.String("InternalFailure"),
ErrorMessage: aws.String("Internal Service Failure"),
},
},
)
k := KinesisOutput{
Log: testutil.Logger{},
StreamName: testStreamName,
svc: svc,
}
elapsed := k.writeKinesis(records)
require.GreaterOrEqual(t, elapsed.Nanoseconds(), zero)
svc.AssertRequests(t, []*kinesis.PutRecordsInput{
{
StreamName: aws.String(testStreamName),
Records: records,
},
})
}
func TestWriteKinesis_WhenServiceError(t *testing.T) {
records := []types.PutRecordsRequestEntry{
{
PartitionKey: aws.String(testPartitionKey),
},
}
svc := &mockKinesisPutRecords{}
svc.SetupErrorResponse(
&types.InvalidArgumentException{Message: aws.String("Invalid record")},
)
k := KinesisOutput{
Log: testutil.Logger{},
StreamName: testStreamName,
svc: svc,
}
elapsed := k.writeKinesis(records)
require.GreaterOrEqual(t, elapsed.Nanoseconds(), zero)
svc.AssertRequests(t, []*kinesis.PutRecordsInput{
{
StreamName: aws.String(testStreamName),
Records: records,
},
})
}
func TestWrite_NoMetrics(t *testing.T) {
serializer := &influx.Serializer{}
require.NoError(t, serializer.Init())
svc := &mockKinesisPutRecords{}
k := KinesisOutput{
Log: testutil.Logger{},
Partition: &Partition{
Method: "static",
Key: "partitionKey",
},
StreamName: "stream",
serializer: serializer,
svc: svc,
}
err := k.Write(nil)
require.NoError(t, err, "Should not return error")
svc.AssertRequests(t, make([]*kinesis.PutRecordsInput, 0))
}
func TestWrite_SingleMetric(t *testing.T) {
serializer := &influx.Serializer{}
require.NoError(t, serializer.Init())
svc := &mockKinesisPutRecords{}
svc.SetupGenericResponse(1, 0)
k := KinesisOutput{
Log: testutil.Logger{},
Partition: &Partition{
Method: "static",
Key: testPartitionKey,
},
StreamName: testStreamName,
serializer: serializer,
svc: svc,
}
metric, metricData := createTestMetric(t, "metric1", serializer)
err := k.Write([]telegraf.Metric{metric})
require.NoError(t, err, "Should not return error")
svc.AssertRequests(t, []*kinesis.PutRecordsInput{
{
StreamName: aws.String(testStreamName),
Records: []types.PutRecordsRequestEntry{
{
PartitionKey: aws.String(testPartitionKey),
Data: metricData,
},
},
},
})
}
func TestWrite_MultipleMetrics_SinglePartialRequest(t *testing.T) {
serializer := &influx.Serializer{}
require.NoError(t, serializer.Init())
svc := &mockKinesisPutRecords{}
svc.SetupGenericResponse(3, 0)
k := KinesisOutput{
Log: testutil.Logger{},
Partition: &Partition{
Method: "static",
Key: testPartitionKey,
},
StreamName: testStreamName,
serializer: serializer,
svc: svc,
}
metrics, metricsData := createTestMetrics(t, 3, serializer)
err := k.Write(metrics)
require.NoError(t, err, "Should not return error")
svc.AssertRequests(t, []*kinesis.PutRecordsInput{
{
StreamName: aws.String(testStreamName),
Records: createPutRecordsRequestEntries(
metricsData,
),
},
})
}
func TestWrite_MultipleMetrics_SingleFullRequest(t *testing.T) {
serializer := &influx.Serializer{}
require.NoError(t, serializer.Init())
svc := &mockKinesisPutRecords{}
svc.SetupGenericResponse(maxRecordsPerRequest, 0)
k := KinesisOutput{
Log: testutil.Logger{},
Partition: &Partition{
Method: "static",
Key: testPartitionKey,
},
StreamName: testStreamName,
serializer: serializer,
svc: svc,
}
metrics, metricsData := createTestMetrics(t, maxRecordsPerRequest, serializer)
err := k.Write(metrics)
require.NoError(t, err, "Should not return error")
svc.AssertRequests(t, []*kinesis.PutRecordsInput{
{
StreamName: aws.String(testStreamName),
Records: createPutRecordsRequestEntries(
metricsData,
),
},
})
}
func TestWrite_MultipleMetrics_MultipleRequests(t *testing.T) {
serializer := &influx.Serializer{}
require.NoError(t, serializer.Init())
svc := &mockKinesisPutRecords{}
svc.SetupGenericResponse(maxRecordsPerRequest, 0)
svc.SetupGenericResponse(1, 0)
k := KinesisOutput{
Log: testutil.Logger{},
Partition: &Partition{
Method: "static",
Key: testPartitionKey,
},
StreamName: testStreamName,
serializer: serializer,
svc: svc,
}
metrics, metricsData := createTestMetrics(t, maxRecordsPerRequest+1, serializer)
err := k.Write(metrics)
require.NoError(t, err, "Should not return error")
svc.AssertRequests(t, []*kinesis.PutRecordsInput{
{
StreamName: aws.String(testStreamName),
Records: createPutRecordsRequestEntries(
metricsData[0:maxRecordsPerRequest],
),
},
{
StreamName: aws.String(testStreamName),
Records: createPutRecordsRequestEntries(
metricsData[maxRecordsPerRequest:],
),
},
})
}
func TestWrite_MultipleMetrics_MultipleFullRequests(t *testing.T) {
serializer := &influx.Serializer{}
require.NoError(t, serializer.Init())
svc := &mockKinesisPutRecords{}
svc.SetupGenericResponse(maxRecordsPerRequest, 0)
svc.SetupGenericResponse(maxRecordsPerRequest, 0)
k := KinesisOutput{
Log: testutil.Logger{},
Partition: &Partition{
Method: "static",
Key: testPartitionKey,
},
StreamName: testStreamName,
serializer: serializer,
svc: svc,
}
metrics, metricsData := createTestMetrics(t, maxRecordsPerRequest*2, serializer)
err := k.Write(metrics)
require.NoError(t, err, "Should not return error")
svc.AssertRequests(t, []*kinesis.PutRecordsInput{
{
StreamName: aws.String(testStreamName),
Records: createPutRecordsRequestEntries(
metricsData[0:maxRecordsPerRequest],
),
},
{
StreamName: aws.String(testStreamName),
Records: createPutRecordsRequestEntries(
metricsData[maxRecordsPerRequest:],
),
},
})
}
func TestWrite_SerializerError(t *testing.T) {
serializer := &influx.Serializer{}
require.NoError(t, serializer.Init())
svc := &mockKinesisPutRecords{}
svc.SetupGenericResponse(2, 0)
k := KinesisOutput{
Log: testutil.Logger{},
Partition: &Partition{
Method: "static",
Key: testPartitionKey,
},
StreamName: testStreamName,
serializer: serializer,
svc: svc,
}
metric1, metric1Data := createTestMetric(t, "metric1", serializer)
metric2, metric2Data := createTestMetric(t, "metric2", serializer)
// metric is invalid because of empty name
invalidMetric := testutil.TestMetric(3, "")
err := k.Write([]telegraf.Metric{
metric1,
invalidMetric,
metric2,
})
require.NoError(t, err, "Should not return error")
// remaining valid metrics should still get written
svc.AssertRequests(t, []*kinesis.PutRecordsInput{
{
StreamName: aws.String(testStreamName),
Records: []types.PutRecordsRequestEntry{
{
PartitionKey: aws.String(testPartitionKey),
Data: metric1Data,
},
{
PartitionKey: aws.String(testPartitionKey),
Data: metric2Data,
},
},
},
})
}
type mockKinesisPutRecordsResponse struct {
Output *kinesis.PutRecordsOutput
Err error
}
type mockKinesisPutRecords struct {
requests []*kinesis.PutRecordsInput
responses []*mockKinesisPutRecordsResponse
}
func (m *mockKinesisPutRecords) SetupResponse(
failedRecordCount int32,
records []types.PutRecordsResultEntry,
) {
m.responses = append(m.responses, &mockKinesisPutRecordsResponse{
Err: nil,
Output: &kinesis.PutRecordsOutput{
FailedRecordCount: aws.Int32(failedRecordCount),
Records: records,
},
})
}
func (m *mockKinesisPutRecords) SetupGenericResponse(successfulRecordCount uint32, failedRecordCount int32) {
records := make([]types.PutRecordsResultEntry, 0, int32(successfulRecordCount)+failedRecordCount)
for i := uint32(0); i < successfulRecordCount; i++ {
records = append(records, types.PutRecordsResultEntry{
SequenceNumber: aws.String(testSequenceNumber),
ShardId: aws.String(testShardID),
})
}
for i := int32(0); i < failedRecordCount; i++ {
records = append(records, types.PutRecordsResultEntry{
ErrorCode: aws.String("InternalFailure"),
ErrorMessage: aws.String("Internal Service Failure"),
})
}
m.SetupResponse(failedRecordCount, records)
}
func (m *mockKinesisPutRecords) SetupErrorResponse(err error) {
m.responses = append(m.responses, &mockKinesisPutRecordsResponse{
Err: err,
Output: nil,
})
}
func (m *mockKinesisPutRecords) PutRecords(_ context.Context, input *kinesis.PutRecordsInput, _ ...func(*kinesis.Options)) (*kinesis.PutRecordsOutput, error) {
reqNum := len(m.requests)
if reqNum > len(m.responses) {
return nil, fmt.Errorf("response for request %+v not setup", reqNum)
}
m.requests = append(m.requests, input)
resp := m.responses[reqNum]
return resp.Output, resp.Err
}
func (m *mockKinesisPutRecords) AssertRequests(
t *testing.T,
expected []*kinesis.PutRecordsInput,
) {
require.Lenf(t, m.requests, len(expected), "Expected %v requests", len(expected))
for i, expectedInput := range expected {
actualInput := m.requests[i]
require.Equalf(t,
expectedInput.StreamName,
actualInput.StreamName,
"Expected request %v to have correct StreamName", i,
)
require.Lenf(t, actualInput.Records, len(expectedInput.Records), "Expected request %v to have %v Records", i, len(expectedInput.Records))
for r, expectedRecord := range expectedInput.Records {
actualRecord := actualInput.Records[r]
require.Equalf(t,
expectedRecord.PartitionKey,
actualRecord.PartitionKey,
"Expected (request %v, record %v) to have correct PartitionKey", i, r,
)
require.Equalf(t,
expectedRecord.ExplicitHashKey,
actualRecord.ExplicitHashKey,
"Expected (request %v, record %v) to have correct ExplicitHashKey", i, r,
)
require.Equalf(t,
expectedRecord.Data,
actualRecord.Data,
"Expected (request %v, record %v) to have correct Data", i, r,
)
}
}
}
func createTestMetric(t *testing.T, name string, serializer telegraf.Serializer) (telegraf.Metric, []byte) {
metric := testutil.TestMetric(1, name)
data, err := serializer.Serialize(metric)
require.NoError(t, err)
return metric, data
}
func createTestMetrics(t *testing.T, count uint32, serializer telegraf.Serializer) ([]telegraf.Metric, [][]byte) {
metrics := make([]telegraf.Metric, 0, count)
metricsData := make([][]byte, 0, count)
for i := uint32(0); i < count; i++ {
name := fmt.Sprintf("metric%d", i)
metric, data := createTestMetric(t, name, serializer)
metrics = append(metrics, metric)
metricsData = append(metricsData, data)
}
return metrics, metricsData
}
func createPutRecordsRequestEntries(
metricsData [][]byte,
) []types.PutRecordsRequestEntry {
records := make([]types.PutRecordsRequestEntry, 0, len(metricsData))
for _, data := range metricsData {
records = append(records, types.PutRecordsRequestEntry{
PartitionKey: aws.String(testPartitionKey),
Data: data,
})
}
return records
}