mirror of https://github.com/prometheus/prometheus
Browse Source
Deprecated since 2018. Signed-off-by: Bryan Boreham <bjboreham@gmail.com>pull/13764/head
Bryan Boreham
8 months ago
4 changed files with 1 additions and 1861 deletions
@ -1,553 +0,0 @@
|
||||
// Copyright 2017 The Prometheus Authors
|
||||
// Licensed under the Apache License, Version 2.0 (the "License");
|
||||
// you may not use this file except in compliance with the License.
|
||||
// You may obtain a copy of the License at
|
||||
//
|
||||
// http://www.apache.org/licenses/LICENSE-2.0
|
||||
//
|
||||
// Unless required by applicable law or agreed to in writing, software
|
||||
// distributed under the License is distributed on an "AS IS" BASIS,
|
||||
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
// See the License for the specific language governing permissions and
|
||||
// limitations under the License.
|
||||
|
||||
//go:build !windows
|
||||
|
||||
package tsdb |
||||
|
||||
import ( |
||||
"encoding/binary" |
||||
"io" |
||||
"math/rand" |
||||
"os" |
||||
"path" |
||||
"path/filepath" |
||||
"testing" |
||||
"time" |
||||
|
||||
"github.com/go-kit/log" |
||||
"github.com/stretchr/testify/require" |
||||
|
||||
"github.com/prometheus/prometheus/model/labels" |
||||
"github.com/prometheus/prometheus/storage" |
||||
"github.com/prometheus/prometheus/tsdb/chunks" |
||||
"github.com/prometheus/prometheus/tsdb/record" |
||||
"github.com/prometheus/prometheus/tsdb/tombstones" |
||||
"github.com/prometheus/prometheus/tsdb/wlog" |
||||
"github.com/prometheus/prometheus/util/testutil" |
||||
) |
||||
|
||||
func TestSegmentWAL_cut(t *testing.T) { |
||||
tmpdir := t.TempDir() |
||||
|
||||
// This calls cut() implicitly the first time without a previous tail.
|
||||
w, err := OpenSegmentWAL(tmpdir, nil, 0, nil) |
||||
require.NoError(t, err) |
||||
|
||||
require.NoError(t, w.write(WALEntrySeries, 1, []byte("Hello World!!"))) |
||||
|
||||
require.NoError(t, w.cut()) |
||||
|
||||
// Cutting creates a new file.
|
||||
require.Len(t, w.files, 2) |
||||
|
||||
require.NoError(t, w.write(WALEntrySeries, 1, []byte("Hello World!!"))) |
||||
|
||||
require.NoError(t, w.Close()) |
||||
|
||||
for _, of := range w.files { |
||||
f, err := os.Open(of.Name()) |
||||
require.NoError(t, err) |
||||
|
||||
// Verify header data.
|
||||
metab := make([]byte, 8) |
||||
_, err = f.Read(metab) |
||||
require.NoError(t, err) |
||||
require.Equal(t, WALMagic, binary.BigEndian.Uint32(metab[:4])) |
||||
require.Equal(t, WALFormatDefault, metab[4]) |
||||
|
||||
// We cannot actually check for correct pre-allocation as it is
|
||||
// optional per filesystem and handled transparently.
|
||||
et, flag, b, err := newWALReader(nil, nil).entry(f) |
||||
require.NoError(t, err) |
||||
require.Equal(t, WALEntrySeries, et) |
||||
require.Equal(t, byte(walSeriesSimple), flag) |
||||
require.Equal(t, []byte("Hello World!!"), b) |
||||
} |
||||
} |
||||
|
||||
func TestSegmentWAL_Truncate(t *testing.T) { |
||||
const ( |
||||
numMetrics = 20000 |
||||
batch = 100 |
||||
) |
||||
series, err := labels.ReadLabels(filepath.Join("testdata", "20kseries.json"), numMetrics) |
||||
require.NoError(t, err) |
||||
|
||||
dir := t.TempDir() |
||||
|
||||
w, err := OpenSegmentWAL(dir, nil, 0, nil) |
||||
require.NoError(t, err) |
||||
defer func(wal *SegmentWAL) { require.NoError(t, wal.Close()) }(w) |
||||
w.segmentSize = 10000 |
||||
|
||||
for i := 0; i < numMetrics; i += batch { |
||||
var rs []record.RefSeries |
||||
|
||||
for j, s := range series[i : i+batch] { |
||||
rs = append(rs, record.RefSeries{Labels: s, Ref: chunks.HeadSeriesRef(i+j) + 1}) |
||||
} |
||||
err := w.LogSeries(rs) |
||||
require.NoError(t, err) |
||||
} |
||||
|
||||
// We mark the 2nd half of the files with a min timestamp that should discard
|
||||
// them from the selection of compactable files.
|
||||
for i, f := range w.files[len(w.files)/2:] { |
||||
f.maxTime = int64(1000 + i) |
||||
} |
||||
// All series in those files must be preserved regarding of the provided postings list.
|
||||
boundarySeries := w.files[len(w.files)/2].minSeries |
||||
|
||||
// We truncate while keeping every 2nd series.
|
||||
keep := map[chunks.HeadSeriesRef]struct{}{} |
||||
for i := 1; i <= numMetrics; i += 2 { |
||||
keep[chunks.HeadSeriesRef(i)] = struct{}{} |
||||
} |
||||
keepf := func(id chunks.HeadSeriesRef) bool { |
||||
_, ok := keep[id] |
||||
return ok |
||||
} |
||||
|
||||
err = w.Truncate(1000, keepf) |
||||
require.NoError(t, err) |
||||
|
||||
var expected []record.RefSeries |
||||
|
||||
for i := 1; i <= numMetrics; i++ { |
||||
if i%2 == 1 || chunks.HeadSeriesRef(i) >= boundarySeries { |
||||
expected = append(expected, record.RefSeries{Ref: chunks.HeadSeriesRef(i), Labels: series[i-1]}) |
||||
} |
||||
} |
||||
|
||||
// Call Truncate once again to see whether we can read the written file without
|
||||
// creating a new WAL.
|
||||
err = w.Truncate(1000, keepf) |
||||
require.NoError(t, err) |
||||
require.NoError(t, w.Close()) |
||||
|
||||
// The same again with a new WAL.
|
||||
w, err = OpenSegmentWAL(dir, nil, 0, nil) |
||||
require.NoError(t, err) |
||||
defer func(wal *SegmentWAL) { require.NoError(t, wal.Close()) }(w) |
||||
|
||||
var readSeries []record.RefSeries |
||||
r := w.Reader() |
||||
|
||||
require.NoError(t, r.Read(func(s []record.RefSeries) { |
||||
readSeries = append(readSeries, s...) |
||||
}, nil, nil)) |
||||
|
||||
testutil.RequireEqual(t, expected, readSeries) |
||||
} |
||||
|
||||
// Symmetrical test of reading and writing to the WAL via its main interface.
|
||||
func TestSegmentWAL_Log_Restore(t *testing.T) { |
||||
const ( |
||||
numMetrics = 50 |
||||
iterations = 5 |
||||
stepSize = 5 |
||||
) |
||||
// Generate testing data. It does not make semantic sense but
|
||||
// for the purpose of this test.
|
||||
series, err := labels.ReadLabels(filepath.Join("testdata", "20kseries.json"), numMetrics) |
||||
require.NoError(t, err) |
||||
|
||||
dir := t.TempDir() |
||||
|
||||
var ( |
||||
recordedSeries [][]record.RefSeries |
||||
recordedSamples [][]record.RefSample |
||||
recordedDeletes [][]tombstones.Stone |
||||
) |
||||
var totalSamples int |
||||
|
||||
// Open WAL a bunch of times, validate all previous data can be read,
|
||||
// write more data to it, close it.
|
||||
for k := 0; k < numMetrics; k += numMetrics / iterations { |
||||
w, err := OpenSegmentWAL(dir, nil, 0, nil) |
||||
require.NoError(t, err) |
||||
|
||||
// Set smaller segment size so we can actually write several files.
|
||||
w.segmentSize = 1000 * 1000 |
||||
|
||||
r := w.Reader() |
||||
|
||||
var ( |
||||
resultSeries [][]record.RefSeries |
||||
resultSamples [][]record.RefSample |
||||
resultDeletes [][]tombstones.Stone |
||||
) |
||||
|
||||
serf := func(series []record.RefSeries) { |
||||
if len(series) > 0 { |
||||
clsets := make([]record.RefSeries, len(series)) |
||||
copy(clsets, series) |
||||
resultSeries = append(resultSeries, clsets) |
||||
} |
||||
} |
||||
smplf := func(smpls []record.RefSample) { |
||||
if len(smpls) > 0 { |
||||
csmpls := make([]record.RefSample, len(smpls)) |
||||
copy(csmpls, smpls) |
||||
resultSamples = append(resultSamples, csmpls) |
||||
} |
||||
} |
||||
|
||||
delf := func(stones []tombstones.Stone) { |
||||
if len(stones) > 0 { |
||||
cst := make([]tombstones.Stone, len(stones)) |
||||
copy(cst, stones) |
||||
resultDeletes = append(resultDeletes, cst) |
||||
} |
||||
} |
||||
|
||||
require.NoError(t, r.Read(serf, smplf, delf)) |
||||
|
||||
testutil.RequireEqual(t, recordedSamples, resultSamples) |
||||
testutil.RequireEqual(t, recordedSeries, resultSeries) |
||||
testutil.RequireEqual(t, recordedDeletes, resultDeletes) |
||||
|
||||
series := series[k : k+(numMetrics/iterations)] |
||||
|
||||
// Insert in batches and generate different amounts of samples for each.
|
||||
for i := 0; i < len(series); i += stepSize { |
||||
var samples []record.RefSample |
||||
var stones []tombstones.Stone |
||||
|
||||
for j := 0; j < i*10; j++ { |
||||
samples = append(samples, record.RefSample{ |
||||
Ref: chunks.HeadSeriesRef(j % 10000), |
||||
T: int64(j * 2), |
||||
V: rand.Float64(), |
||||
}) |
||||
} |
||||
|
||||
for j := 0; j < i*20; j++ { |
||||
ts := rand.Int63() |
||||
stones = append(stones, tombstones.Stone{Ref: storage.SeriesRef(rand.Uint64()), Intervals: tombstones.Intervals{{Mint: ts, Maxt: ts + rand.Int63n(10000)}}}) |
||||
} |
||||
|
||||
lbls := series[i : i+stepSize] |
||||
series := make([]record.RefSeries, 0, len(series)) |
||||
for j, l := range lbls { |
||||
series = append(series, record.RefSeries{ |
||||
Ref: chunks.HeadSeriesRef(i + j), |
||||
Labels: l, |
||||
}) |
||||
} |
||||
|
||||
require.NoError(t, w.LogSeries(series)) |
||||
require.NoError(t, w.LogSamples(samples)) |
||||
require.NoError(t, w.LogDeletes(stones)) |
||||
|
||||
if len(lbls) > 0 { |
||||
recordedSeries = append(recordedSeries, series) |
||||
} |
||||
if len(samples) > 0 { |
||||
recordedSamples = append(recordedSamples, samples) |
||||
totalSamples += len(samples) |
||||
} |
||||
if len(stones) > 0 { |
||||
recordedDeletes = append(recordedDeletes, stones) |
||||
} |
||||
} |
||||
|
||||
require.NoError(t, w.Close()) |
||||
} |
||||
} |
||||
|
||||
func TestWALRestoreCorrupted_invalidSegment(t *testing.T) { |
||||
dir := t.TempDir() |
||||
|
||||
wal, err := OpenSegmentWAL(dir, nil, 0, nil) |
||||
require.NoError(t, err) |
||||
defer func(wal *SegmentWAL) { require.NoError(t, wal.Close()) }(wal) |
||||
|
||||
_, err = wal.createSegmentFile(filepath.Join(dir, "000000")) |
||||
require.NoError(t, err) |
||||
f, err := wal.createSegmentFile(filepath.Join(dir, "000001")) |
||||
require.NoError(t, err) |
||||
f2, err := wal.createSegmentFile(filepath.Join(dir, "000002")) |
||||
require.NoError(t, err) |
||||
require.NoError(t, f2.Close()) |
||||
|
||||
// Make header of second segment invalid.
|
||||
_, err = f.WriteAt([]byte{1, 2, 3, 4}, 0) |
||||
require.NoError(t, err) |
||||
require.NoError(t, f.Close()) |
||||
|
||||
require.NoError(t, wal.Close()) |
||||
|
||||
wal, err = OpenSegmentWAL(dir, log.NewLogfmtLogger(os.Stderr), 0, nil) |
||||
require.NoError(t, err) |
||||
defer func(wal *SegmentWAL) { require.NoError(t, wal.Close()) }(wal) |
||||
|
||||
files, err := os.ReadDir(dir) |
||||
require.NoError(t, err) |
||||
fns := []string{} |
||||
for _, f := range files { |
||||
fns = append(fns, f.Name()) |
||||
} |
||||
require.Equal(t, []string{"000000"}, fns) |
||||
} |
||||
|
||||
// Test reading from a WAL that has been corrupted through various means.
|
||||
func TestWALRestoreCorrupted(t *testing.T) { |
||||
cases := []struct { |
||||
name string |
||||
f func(*testing.T, *SegmentWAL) |
||||
}{ |
||||
{ |
||||
name: "truncate_checksum", |
||||
f: func(t *testing.T, w *SegmentWAL) { |
||||
f, err := os.OpenFile(w.files[0].Name(), os.O_WRONLY, 0o666) |
||||
require.NoError(t, err) |
||||
defer f.Close() |
||||
|
||||
off, err := f.Seek(0, io.SeekEnd) |
||||
require.NoError(t, err) |
||||
|
||||
require.NoError(t, f.Truncate(off-1)) |
||||
}, |
||||
}, |
||||
{ |
||||
name: "truncate_body", |
||||
f: func(t *testing.T, w *SegmentWAL) { |
||||
f, err := os.OpenFile(w.files[0].Name(), os.O_WRONLY, 0o666) |
||||
require.NoError(t, err) |
||||
defer f.Close() |
||||
|
||||
off, err := f.Seek(0, io.SeekEnd) |
||||
require.NoError(t, err) |
||||
|
||||
require.NoError(t, f.Truncate(off-8)) |
||||
}, |
||||
}, |
||||
{ |
||||
name: "body_content", |
||||
f: func(t *testing.T, w *SegmentWAL) { |
||||
f, err := os.OpenFile(w.files[0].Name(), os.O_WRONLY, 0o666) |
||||
require.NoError(t, err) |
||||
defer f.Close() |
||||
|
||||
off, err := f.Seek(0, io.SeekEnd) |
||||
require.NoError(t, err) |
||||
|
||||
// Write junk before checksum starts.
|
||||
_, err = f.WriteAt([]byte{1, 2, 3, 4}, off-8) |
||||
require.NoError(t, err) |
||||
}, |
||||
}, |
||||
{ |
||||
name: "checksum", |
||||
f: func(t *testing.T, w *SegmentWAL) { |
||||
f, err := os.OpenFile(w.files[0].Name(), os.O_WRONLY, 0o666) |
||||
require.NoError(t, err) |
||||
defer f.Close() |
||||
|
||||
off, err := f.Seek(0, io.SeekEnd) |
||||
require.NoError(t, err) |
||||
|
||||
// Write junk into checksum
|
||||
_, err = f.WriteAt([]byte{1, 2, 3, 4}, off-4) |
||||
require.NoError(t, err) |
||||
}, |
||||
}, |
||||
} |
||||
for _, c := range cases { |
||||
t.Run(c.name, func(t *testing.T) { |
||||
// Generate testing data. It does not make semantic sense but
|
||||
// for the purpose of this test.
|
||||
dir := t.TempDir() |
||||
|
||||
w, err := OpenSegmentWAL(dir, nil, 0, nil) |
||||
require.NoError(t, err) |
||||
defer func(wal *SegmentWAL) { require.NoError(t, wal.Close()) }(w) |
||||
|
||||
require.NoError(t, w.LogSamples([]record.RefSample{{T: 1, V: 2}})) |
||||
require.NoError(t, w.LogSamples([]record.RefSample{{T: 2, V: 3}})) |
||||
|
||||
require.NoError(t, w.cut()) |
||||
|
||||
// Sleep 2 seconds to avoid error where cut and test "cases" function may write or
|
||||
// truncate the file out of orders as "cases" are not synchronized with cut.
|
||||
// Hopefully cut will complete by 2 seconds.
|
||||
time.Sleep(2 * time.Second) |
||||
|
||||
require.NoError(t, w.LogSamples([]record.RefSample{{T: 3, V: 4}})) |
||||
require.NoError(t, w.LogSamples([]record.RefSample{{T: 5, V: 6}})) |
||||
|
||||
require.NoError(t, w.Close()) |
||||
|
||||
// cut() truncates and fsyncs the first segment async. If it happens after
|
||||
// the corruption we apply below, the corruption will be overwritten again.
|
||||
// Fire and forget a sync to avoid flakiness.
|
||||
w.files[0].Sync() |
||||
// Corrupt the second entry in the first file.
|
||||
// After re-opening we must be able to read the first entry
|
||||
// and the rest, including the second file, must be truncated for clean further
|
||||
// writes.
|
||||
c.f(t, w) |
||||
|
||||
logger := log.NewLogfmtLogger(os.Stderr) |
||||
|
||||
w2, err := OpenSegmentWAL(dir, logger, 0, nil) |
||||
require.NoError(t, err) |
||||
defer func(wal *SegmentWAL) { require.NoError(t, wal.Close()) }(w2) |
||||
|
||||
r := w2.Reader() |
||||
|
||||
serf := func(l []record.RefSeries) { |
||||
require.Empty(t, l) |
||||
} |
||||
|
||||
// Weird hack to check order of reads.
|
||||
i := 0 |
||||
samplef := func(s []record.RefSample) { |
||||
if i == 0 { |
||||
require.Equal(t, []record.RefSample{{T: 1, V: 2}}, s) |
||||
i++ |
||||
} else { |
||||
require.Equal(t, []record.RefSample{{T: 99, V: 100}}, s) |
||||
} |
||||
} |
||||
|
||||
require.NoError(t, r.Read(serf, samplef, nil)) |
||||
|
||||
require.NoError(t, w2.LogSamples([]record.RefSample{{T: 99, V: 100}})) |
||||
require.NoError(t, w2.Close()) |
||||
|
||||
// We should see the first valid entry and the new one, everything after
|
||||
// is truncated.
|
||||
w3, err := OpenSegmentWAL(dir, logger, 0, nil) |
||||
require.NoError(t, err) |
||||
defer func(wal *SegmentWAL) { require.NoError(t, wal.Close()) }(w3) |
||||
|
||||
r = w3.Reader() |
||||
|
||||
i = 0 |
||||
require.NoError(t, r.Read(serf, samplef, nil)) |
||||
}) |
||||
} |
||||
} |
||||
|
||||
func TestMigrateWAL_Empty(t *testing.T) { |
||||
// The migration procedure must properly deal with a zero-length segment,
|
||||
// which is valid in the new format.
|
||||
dir := t.TempDir() |
||||
|
||||
wdir := path.Join(dir, "wal") |
||||
|
||||
// Initialize empty WAL.
|
||||
w, err := wlog.New(nil, nil, wdir, wlog.CompressionNone) |
||||
require.NoError(t, err) |
||||
require.NoError(t, w.Close()) |
||||
|
||||
require.NoError(t, MigrateWAL(nil, wdir)) |
||||
} |
||||
|
||||
func TestMigrateWAL_Fuzz(t *testing.T) { |
||||
dir := t.TempDir() |
||||
|
||||
wdir := path.Join(dir, "wal") |
||||
|
||||
// Should pass if no WAL exists yet.
|
||||
require.NoError(t, MigrateWAL(nil, wdir)) |
||||
|
||||
oldWAL, err := OpenSegmentWAL(wdir, nil, time.Minute, nil) |
||||
require.NoError(t, err) |
||||
|
||||
// Write some data.
|
||||
require.NoError(t, oldWAL.LogSeries([]record.RefSeries{ |
||||
{Ref: 100, Labels: labels.FromStrings("abc", "def", "123", "456")}, |
||||
{Ref: 1, Labels: labels.FromStrings("abc", "def2", "1234", "4567")}, |
||||
})) |
||||
require.NoError(t, oldWAL.LogSamples([]record.RefSample{ |
||||
{Ref: 1, T: 100, V: 200}, |
||||
{Ref: 2, T: 300, V: 400}, |
||||
})) |
||||
require.NoError(t, oldWAL.LogSeries([]record.RefSeries{ |
||||
{Ref: 200, Labels: labels.FromStrings("xyz", "def", "foo", "bar")}, |
||||
})) |
||||
require.NoError(t, oldWAL.LogSamples([]record.RefSample{ |
||||
{Ref: 3, T: 100, V: 200}, |
||||
{Ref: 4, T: 300, V: 400}, |
||||
})) |
||||
require.NoError(t, oldWAL.LogDeletes([]tombstones.Stone{ |
||||
{Ref: 1, Intervals: []tombstones.Interval{{Mint: 100, Maxt: 200}}}, |
||||
})) |
||||
|
||||
require.NoError(t, oldWAL.Close()) |
||||
|
||||
// Perform migration.
|
||||
require.NoError(t, MigrateWAL(nil, wdir)) |
||||
|
||||
w, err := wlog.New(nil, nil, wdir, wlog.CompressionNone) |
||||
require.NoError(t, err) |
||||
|
||||
// We can properly write some new data after migration.
|
||||
var enc record.Encoder |
||||
require.NoError(t, w.Log(enc.Samples([]record.RefSample{ |
||||
{Ref: 500, T: 1, V: 1}, |
||||
}, nil))) |
||||
|
||||
require.NoError(t, w.Close()) |
||||
|
||||
// Read back all data.
|
||||
sr, err := wlog.NewSegmentsReader(wdir) |
||||
require.NoError(t, err) |
||||
|
||||
r := wlog.NewReader(sr) |
||||
var res []interface{} |
||||
dec := record.NewDecoder(labels.NewSymbolTable()) |
||||
|
||||
for r.Next() { |
||||
rec := r.Record() |
||||
|
||||
switch dec.Type(rec) { |
||||
case record.Series: |
||||
s, err := dec.Series(rec, nil) |
||||
require.NoError(t, err) |
||||
res = append(res, s) |
||||
case record.Samples: |
||||
s, err := dec.Samples(rec, nil) |
||||
require.NoError(t, err) |
||||
res = append(res, s) |
||||
case record.Tombstones: |
||||
s, err := dec.Tombstones(rec, nil) |
||||
require.NoError(t, err) |
||||
res = append(res, s) |
||||
default: |
||||
require.Fail(t, "unknown record type %d", dec.Type(rec)) |
||||
} |
||||
} |
||||
require.NoError(t, r.Err()) |
||||
|
||||
testutil.RequireEqual(t, []interface{}{ |
||||
[]record.RefSeries{ |
||||
{Ref: 100, Labels: labels.FromStrings("abc", "def", "123", "456")}, |
||||
{Ref: 1, Labels: labels.FromStrings("abc", "def2", "1234", "4567")}, |
||||
}, |
||||
[]record.RefSample{{Ref: 1, T: 100, V: 200}, {Ref: 2, T: 300, V: 400}}, |
||||
[]record.RefSeries{ |
||||
{Ref: 200, Labels: labels.FromStrings("xyz", "def", "foo", "bar")}, |
||||
}, |
||||
[]record.RefSample{{Ref: 3, T: 100, V: 200}, {Ref: 4, T: 300, V: 400}}, |
||||
[]tombstones.Stone{{Ref: 1, Intervals: []tombstones.Interval{{Mint: 100, Maxt: 200}}}}, |
||||
[]record.RefSample{{Ref: 500, T: 1, V: 1}}, |
||||
}, res) |
||||
|
||||
// Migrating an already migrated WAL shouldn't do anything.
|
||||
require.NoError(t, MigrateWAL(nil, wdir)) |
||||
} |
Loading…
Reference in new issue