// Copyright 2017 The Prometheus Authors
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
package wlog
import (
"bytes"
"fmt"
"io"
"math/rand"
"os"
"path/filepath"
"testing"
client_testutil "github.com/prometheus/client_golang/prometheus/testutil"
"github.com/stretchr/testify/require"
"go.uber.org/goleak"
"github.com/prometheus/prometheus/tsdb/fileutil"
"github.com/prometheus/prometheus/util/testutil"
)
func TestMain ( m * testing . M ) {
goleak . VerifyTestMain ( m )
}
// TestWALRepair_ReadingError ensures that a repair is run for an error
// when reading a record.
func TestWALRepair_ReadingError ( t * testing . T ) {
for name , test := range map [ string ] struct {
corrSgm int // Which segment to corrupt.
corrFunc func ( f * os . File ) // Func that applies the corruption.
intactRecs int // Total expected records left after the repair.
} {
"torn_last_record" : {
2 ,
func ( f * os . File ) {
_ , err := f . Seek ( pageSize * 2 , 0 )
require . NoError ( t , err )
_ , err = f . Write ( [ ] byte { byte ( recFirst ) } )
require . NoError ( t , err )
} ,
8 ,
} ,
// Ensures that the page buffer is big enough to fit
// an entire page size without panicking.
// https://github.com/prometheus/tsdb/pull/414
"bad_header" : {
1 ,
func ( f * os . File ) {
_ , err := f . Seek ( pageSize , 0 )
require . NoError ( t , err )
_ , err = f . Write ( [ ] byte { byte ( recPageTerm ) } )
require . NoError ( t , err )
} ,
4 ,
} ,
"bad_fragment_sequence" : {
1 ,
func ( f * os . File ) {
_ , err := f . Seek ( pageSize , 0 )
require . NoError ( t , err )
_ , err = f . Write ( [ ] byte { byte ( recLast ) } )
require . NoError ( t , err )
} ,
4 ,
} ,
"bad_fragment_flag" : {
1 ,
func ( f * os . File ) {
_ , err := f . Seek ( pageSize , 0 )
require . NoError ( t , err )
_ , err = f . Write ( [ ] byte { 123 } )
require . NoError ( t , err )
} ,
4 ,
} ,
"bad_checksum" : {
1 ,
func ( f * os . File ) {
_ , err := f . Seek ( pageSize + 4 , 0 )
require . NoError ( t , err )
_ , err = f . Write ( [ ] byte { 0 } )
require . NoError ( t , err )
} ,
4 ,
} ,
"bad_length" : {
1 ,
func ( f * os . File ) {
_ , err := f . Seek ( pageSize + 2 , 0 )
require . NoError ( t , err )
_ , err = f . Write ( [ ] byte { 0 } )
require . NoError ( t , err )
} ,
4 ,
} ,
"bad_content" : {
1 ,
func ( f * os . File ) {
_ , err := f . Seek ( pageSize + 100 , 0 )
require . NoError ( t , err )
_ , err = f . Write ( [ ] byte ( "beef" ) )
require . NoError ( t , err )
} ,
4 ,
} ,
} {
t . Run ( name , func ( t * testing . T ) {
dir := t . TempDir ( )
// We create 3 segments with 3 records each and
// then corrupt a given record in a given segment.
// As a result we want a repaired WAL with given intact records.
segSize := 3 * pageSize
w , err := NewSize ( nil , nil , dir , segSize , CompressionNone )
require . NoError ( t , err )
var records [ ] [ ] byte
for i := 1 ; i <= 9 ; i ++ {
b := make ( [ ] byte , pageSize - recordHeaderSize )
b [ 0 ] = byte ( i )
records = append ( records , b )
require . NoError ( t , w . Log ( b ) )
}
first , last , err := Segments ( w . Dir ( ) )
require . NoError ( t , err )
require . Equal ( t , 3 , 1 + last - first , "wlog creation didn't result in expected number of segments" )
require . NoError ( t , w . Close ( ) )
f , err := os . OpenFile ( SegmentName ( dir , test . corrSgm ) , os . O_RDWR , 0 o666 )
require . NoError ( t , err )
// Apply corruption function.
test . corrFunc ( f )
require . NoError ( t , f . Close ( ) )
w , err = NewSize ( nil , nil , dir , segSize , CompressionNone )
require . NoError ( t , err )
defer w . Close ( )
first , last , err = Segments ( w . Dir ( ) )
require . NoError ( t , err )
// Backfill segments from the most recent checkpoint onwards.
for i := first ; i <= last ; i ++ {
s , err := OpenReadSegment ( SegmentName ( w . Dir ( ) , i ) )
require . NoError ( t , err )
sr := NewSegmentBufReader ( s )
require . NoError ( t , err )
r := NewReader ( sr )
for r . Next ( ) { // nolint:revive
}
// Close the segment so we don't break things on Windows.
s . Close ( )
// No corruption in this segment.
if r . Err ( ) == nil {
continue
}
require . NoError ( t , w . Repair ( r . Err ( ) ) )
break
}
sr , err := NewSegmentsReader ( dir )
require . NoError ( t , err )
defer sr . Close ( )
r := NewReader ( sr )
var result [ ] [ ] byte
for r . Next ( ) {
var b [ ] byte
result = append ( result , append ( b , r . Record ( ) ... ) )
}
require . NoError ( t , r . Err ( ) )
require . Equal ( t , test . intactRecs , len ( result ) , "Wrong number of intact records" )
for i , r := range result {
if ! bytes . Equal ( records [ i ] , r ) {
t . Fatalf ( "record %d diverges: want %x, got %x" , i , records [ i ] [ : 10 ] , r [ : 10 ] )
}
}
// Make sure there is a new 0 size Segment after the corrupted Segment.
_ , last , err = Segments ( w . Dir ( ) )
require . NoError ( t , err )
require . Equal ( t , test . corrSgm + 1 , last )
fi , err := os . Stat ( SegmentName ( dir , last ) )
require . NoError ( t , err )
require . Equal ( t , int64 ( 0 ) , fi . Size ( ) )
} )
}
}
// TestCorruptAndCarryOn writes a multi-segment WAL; corrupts the first segment and
// ensures that an error during reading that segment are correctly repaired before
// moving to write more records to the WAL.
func TestCorruptAndCarryOn ( t * testing . T ) {
dir := t . TempDir ( )
var (
logger = testutil . NewLogger ( t )
segmentSize = pageSize * 3
recordSize = ( pageSize / 3 ) - recordHeaderSize
)
// Produce a WAL with a two segments of 3 pages with 3 records each,
// so when we truncate the file we're guaranteed to split a record.
{
w , err := NewSize ( logger , nil , dir , segmentSize , CompressionNone )
require . NoError ( t , err )
for i := 0 ; i < 18 ; i ++ {
buf := make ( [ ] byte , recordSize )
_ , err := rand . Read ( buf )
require . NoError ( t , err )
err = w . Log ( buf )
require . NoError ( t , err )
}
err = w . Close ( )
require . NoError ( t , err )
}
// Check all the segments are the correct size.
{
segments , err := listSegments ( dir )
require . NoError ( t , err )
for _ , segment := range segments {
f , err := os . OpenFile ( filepath . Join ( dir , fmt . Sprintf ( "%08d" , segment . index ) ) , os . O_RDONLY , 0 o666 )
require . NoError ( t , err )
fi , err := f . Stat ( )
require . NoError ( t , err )
t . Log ( "segment" , segment . index , "size" , fi . Size ( ) )
require . Equal ( t , int64 ( segmentSize ) , fi . Size ( ) )
err = f . Close ( )
require . NoError ( t , err )
}
}
// Truncate the first file, splitting the middle record in the second
// page in half, leaving 4 valid records.
{
f , err := os . OpenFile ( filepath . Join ( dir , fmt . Sprintf ( "%08d" , 0 ) ) , os . O_RDWR , 0 o666 )
require . NoError ( t , err )
fi , err := f . Stat ( )
require . NoError ( t , err )
require . Equal ( t , int64 ( segmentSize ) , fi . Size ( ) )
err = f . Truncate ( int64 ( segmentSize / 2 ) )
require . NoError ( t , err )
err = f . Close ( )
require . NoError ( t , err )
}
// Now try and repair this WAL, and write 5 more records to it.
{
sr , err := NewSegmentsReader ( dir )
require . NoError ( t , err )
reader := NewReader ( sr )
i := 0
for ; i < 4 && reader . Next ( ) ; i ++ {
require . Equal ( t , recordSize , len ( reader . Record ( ) ) )
}
require . Equal ( t , 4 , i , "not enough records" )
require . False ( t , reader . Next ( ) , "unexpected record" )
corruptionErr := reader . Err ( )
require . Error ( t , corruptionErr )
err = sr . Close ( )
require . NoError ( t , err )
w , err := NewSize ( logger , nil , dir , segmentSize , CompressionNone )
require . NoError ( t , err )
err = w . Repair ( corruptionErr )
require . NoError ( t , err )
// Ensure that we have a completely clean slate after repairing.
require . Equal ( t , w . segment . Index ( ) , 1 ) // We corrupted segment 0.
require . Equal ( t , w . donePages , 0 )
for i := 0 ; i < 5 ; i ++ {
buf := make ( [ ] byte , recordSize )
_ , err := rand . Read ( buf )
require . NoError ( t , err )
err = w . Log ( buf )
require . NoError ( t , err )
}
err = w . Close ( )
require . NoError ( t , err )
}
// Replay the WAL. Should get 9 records.
{
sr , err := NewSegmentsReader ( dir )
require . NoError ( t , err )
reader := NewReader ( sr )
i := 0
for ; i < 9 && reader . Next ( ) ; i ++ {
require . Equal ( t , recordSize , len ( reader . Record ( ) ) )
}
require . Equal ( t , 9 , i , "wrong number of records" )
require . False ( t , reader . Next ( ) , "unexpected record" )
require . Equal ( t , nil , reader . Err ( ) )
sr . Close ( )
}
}
// TestClose ensures that calling Close more than once doesn't panic and doesn't block.
func TestClose ( t * testing . T ) {
dir := t . TempDir ( )
w , err := NewSize ( nil , nil , dir , pageSize , CompressionNone )
require . NoError ( t , err )
require . NoError ( t , w . Close ( ) )
require . Error ( t , w . Close ( ) )
}
func TestSegmentMetric ( t * testing . T ) {
var (
segmentSize = pageSize
recordSize = ( pageSize / 2 ) - recordHeaderSize
)
dir := t . TempDir ( )
w , err := NewSize ( nil , nil , dir , segmentSize , CompressionNone )
require . NoError ( t , err )
initialSegment := client_testutil . ToFloat64 ( w . metrics . currentSegment )
// Write 3 records, each of which is half the segment size, meaning we should rotate to the next segment.
for i := 0 ; i < 3 ; i ++ {
buf := make ( [ ] byte , recordSize )
_ , err := rand . Read ( buf )
require . NoError ( t , err )
err = w . Log ( buf )
require . NoError ( t , err )
}
require . Equal ( t , initialSegment + 1 , client_testutil . ToFloat64 ( w . metrics . currentSegment ) , "segment metric did not increment after segment rotation" )
require . NoError ( t , w . Close ( ) )
}
func TestCompression ( t * testing . T ) {
bootstrap := func ( compressed CompressionType ) string {
const (
segmentSize = pageSize
recordSize = ( pageSize / 2 ) - recordHeaderSize
records = 100
)
dirPath := t . TempDir ( )
w , err := NewSize ( nil , nil , dirPath , segmentSize , compressed )
require . NoError ( t , err )
buf := make ( [ ] byte , recordSize )
for i := 0 ; i < records ; i ++ {
require . NoError ( t , w . Log ( buf ) )
}
require . NoError ( t , w . Close ( ) )
return dirPath
}
tmpDirs := make ( [ ] string , 0 , 3 )
defer func ( ) {
for _ , dir := range tmpDirs {
require . NoError ( t , os . RemoveAll ( dir ) )
}
} ( )
dirUnCompressed := bootstrap ( CompressionNone )
tmpDirs = append ( tmpDirs , dirUnCompressed )
for _ , compressionType := range [ ] CompressionType { CompressionSnappy , CompressionZstd } {
dirCompressed := bootstrap ( compressionType )
tmpDirs = append ( tmpDirs , dirCompressed )
uncompressedSize , err := fileutil . DirSize ( dirUnCompressed )
require . NoError ( t , err )
compressedSize , err := fileutil . DirSize ( dirCompressed )
require . NoError ( t , err )
require . Greater ( t , float64 ( uncompressedSize ) * 0.75 , float64 ( compressedSize ) , "Compressing zeroes should save at least 25%% space - uncompressedSize: %d, compressedSize: %d" , uncompressedSize , compressedSize )
}
}
func TestLogPartialWrite ( t * testing . T ) {
const segmentSize = pageSize * 2
record := [ ] byte { 1 , 2 , 3 , 4 , 5 }
tests := map [ string ] struct {
numRecords int
faultyRecord int
} {
"partial write when logging first record in a page" : {
numRecords : 10 ,
faultyRecord : 1 ,
} ,
"partial write when logging record in the middle of a page" : {
numRecords : 10 ,
faultyRecord : 3 ,
} ,
"partial write when logging last record of a page" : {
numRecords : ( pageSize / ( recordHeaderSize + len ( record ) ) ) + 10 ,
faultyRecord : pageSize / ( recordHeaderSize + len ( record ) ) ,
} ,
// TODO the current implementation suffers this:
// "partial write when logging a record overlapping two pages": {
// numRecords: (pageSize / (recordHeaderSize + len(record))) + 10,
// faultyRecord: pageSize/(recordHeaderSize+len(record)) + 1,
// },
}
for testName , testData := range tests {
t . Run ( testName , func ( t * testing . T ) {
dirPath := t . TempDir ( )
w , err := NewSize ( nil , nil , dirPath , segmentSize , CompressionNone )
require . NoError ( t , err )
// Replace the underlying segment file with a mocked one that injects a failure.
w . segment . SegmentFile = & faultySegmentFile {
SegmentFile : w . segment . SegmentFile ,
writeFailureAfter : ( ( recordHeaderSize + len ( record ) ) * ( testData . faultyRecord - 1 ) ) + 2 ,
writeFailureErr : io . ErrShortWrite ,
}
for i := 1 ; i <= testData . numRecords ; i ++ {
if err := w . Log ( record ) ; i == testData . faultyRecord {
require . Error ( t , io . ErrShortWrite , err )
} else {
require . NoError ( t , err )
}
}
require . NoError ( t , w . Close ( ) )
// Read it back. We expect no corruption.
s , err := OpenReadSegment ( SegmentName ( dirPath , 0 ) )
require . NoError ( t , err )
defer func ( ) { require . NoError ( t , s . Close ( ) ) } ( )
r := NewReader ( NewSegmentBufReader ( s ) )
for i := 0 ; i < testData . numRecords ; i ++ {
require . True ( t , r . Next ( ) )
require . NoError ( t , r . Err ( ) )
require . Equal ( t , record , r . Record ( ) )
}
require . False ( t , r . Next ( ) )
require . NoError ( t , r . Err ( ) )
} )
}
}
type faultySegmentFile struct {
SegmentFile
written int
writeFailureAfter int
writeFailureErr error
}
func ( f * faultySegmentFile ) Write ( p [ ] byte ) ( int , error ) {
if f . writeFailureAfter >= 0 && f . writeFailureAfter < f . written + len ( p ) {
partialLen := f . writeFailureAfter - f . written
if partialLen <= 0 || partialLen >= len ( p ) {
partialLen = 1
}
// Inject failure.
n , _ := f . SegmentFile . Write ( p [ : partialLen ] )
f . written += n
f . writeFailureAfter = - 1
return n , f . writeFailureErr
}
// Proxy the write to the underlying file.
n , err := f . SegmentFile . Write ( p )
f . written += n
return n , err
}
func BenchmarkWAL_LogBatched ( b * testing . B ) {
for _ , compress := range [ ] CompressionType { CompressionNone , CompressionSnappy , CompressionZstd } {
b . Run ( fmt . Sprintf ( "compress=%s" , compress ) , func ( b * testing . B ) {
dir := b . TempDir ( )
w , err := New ( nil , nil , dir , compress )
require . NoError ( b , err )
defer w . Close ( )
var buf [ 2048 ] byte
var recs [ ] [ ] byte
b . SetBytes ( 2048 )
for i := 0 ; i < b . N ; i ++ {
recs = append ( recs , buf [ : ] )
if len ( recs ) < 1000 {
continue
}
err := w . Log ( recs ... )
require . NoError ( b , err )
recs = recs [ : 0 ]
}
// Stop timer to not count fsync time on close.
// If it's counted batched vs. single benchmarks are very similar but
// do not show burst throughput well.
b . StopTimer ( )
} )
}
}
func BenchmarkWAL_Log ( b * testing . B ) {
for _ , compress := range [ ] CompressionType { CompressionNone , CompressionSnappy , CompressionZstd } {
b . Run ( fmt . Sprintf ( "compress=%s" , compress ) , func ( b * testing . B ) {
dir := b . TempDir ( )
w , err := New ( nil , nil , dir , compress )
require . NoError ( b , err )
defer w . Close ( )
var buf [ 2048 ] byte
b . SetBytes ( 2048 )
for i := 0 ; i < b . N ; i ++ {
err := w . Log ( buf [ : ] )
require . NoError ( b , err )
}
// Stop timer to not count fsync time on close.
// If it's counted batched vs. single benchmarks are very similar but
// do not show burst throughput well.
b . StopTimer ( )
} )
}
}