mirror of
https://github.com/dolthub/dolt.git
synced 2026-01-24 10:30:48 -06:00
489 lines
10 KiB
Go
489 lines
10 KiB
Go
// Copyright 2019 Dolthub, Inc.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
//
|
|
// This file incorporates work covered by the following copyright and
|
|
// permission notice:
|
|
//
|
|
// Copyright 2016 Attic Labs, Inc. All rights reserved.
|
|
// Licensed under the Apache License, version 2.0:
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
package nbs
|
|
|
|
import (
|
|
"bytes"
|
|
"context"
|
|
"errors"
|
|
"fmt"
|
|
"sort"
|
|
"sync"
|
|
|
|
lru "github.com/hashicorp/golang-lru/v2"
|
|
"golang.org/x/sync/errgroup"
|
|
|
|
"github.com/dolthub/dolt/go/store/chunks"
|
|
)
|
|
|
|
// Returned when a chunk with a reference to a non-existence chunk is
|
|
// persisted into the ChunkStore. The sanity check is done when we
|
|
// flush the memtable, which means that a ChunkStore interaction which
|
|
// sees this error is not necessarily responsible for the dangling ref.
|
|
// Regardless, all pending writes in the memtable are thrown away when
|
|
// any chunk in the memtable has a dangling ref.
|
|
var ErrDanglingRef = errors.New("dangling ref")
|
|
|
|
const concurrentCompactions = 5
|
|
|
|
func newTableSet(p tablePersister, q MemoryQuotaProvider) tableSet {
|
|
return tableSet{p: p, q: q, rl: make(chan struct{}, concurrentCompactions)}
|
|
}
|
|
|
|
// tableSet is an immutable set of persistable chunkSources.
|
|
type tableSet struct {
|
|
novel, upstream chunkSourceSet
|
|
p tablePersister
|
|
q MemoryQuotaProvider
|
|
rl chan struct{}
|
|
}
|
|
|
|
func (ts tableSet) has(h addr) (bool, error) {
|
|
f := func(css chunkSourceSet) (bool, error) {
|
|
for _, haver := range css {
|
|
has, err := haver.has(h)
|
|
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
|
|
if has {
|
|
return true, nil
|
|
}
|
|
}
|
|
return false, nil
|
|
}
|
|
|
|
novelHas, err := f(ts.novel)
|
|
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
|
|
if novelHas {
|
|
return true, nil
|
|
}
|
|
|
|
return f(ts.upstream)
|
|
}
|
|
|
|
func (ts tableSet) hasMany(addrs []hasRecord) (bool, error) {
|
|
f := func(css chunkSourceSet) (bool, error) {
|
|
for _, haver := range css {
|
|
has, err := haver.hasMany(addrs)
|
|
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
|
|
if !has {
|
|
return false, nil
|
|
}
|
|
}
|
|
return true, nil
|
|
}
|
|
remaining, err := f(ts.novel)
|
|
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
|
|
if !remaining {
|
|
return false, nil
|
|
}
|
|
|
|
return f(ts.upstream)
|
|
}
|
|
|
|
func (ts tableSet) get(ctx context.Context, h addr, stats *Stats) ([]byte, error) {
|
|
if err := ctx.Err(); err != nil {
|
|
return nil, err
|
|
}
|
|
f := func(css chunkSourceSet) ([]byte, error) {
|
|
for _, haver := range css {
|
|
data, err := haver.get(ctx, h, stats)
|
|
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
if data != nil {
|
|
return data, nil
|
|
}
|
|
}
|
|
|
|
return nil, nil
|
|
}
|
|
|
|
data, err := f(ts.novel)
|
|
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
if data != nil {
|
|
return data, nil
|
|
}
|
|
|
|
return f(ts.upstream)
|
|
}
|
|
|
|
func (ts tableSet) getMany(ctx context.Context, eg *errgroup.Group, reqs []getRecord, found func(context.Context, *chunks.Chunk), stats *Stats) (remaining bool, err error) {
|
|
f := func(css chunkSourceSet) bool {
|
|
for _, haver := range css {
|
|
remaining, err = haver.getMany(ctx, eg, reqs, found, stats)
|
|
if err != nil {
|
|
return true
|
|
}
|
|
if !remaining {
|
|
return false
|
|
}
|
|
}
|
|
return true
|
|
}
|
|
|
|
return f(ts.novel) && err == nil && f(ts.upstream), err
|
|
}
|
|
|
|
func (ts tableSet) getManyCompressed(ctx context.Context, eg *errgroup.Group, reqs []getRecord, found func(context.Context, CompressedChunk), stats *Stats) (remaining bool, err error) {
|
|
f := func(css chunkSourceSet) bool {
|
|
for _, haver := range css {
|
|
remaining, err = haver.getManyCompressed(ctx, eg, reqs, found, stats)
|
|
if err != nil {
|
|
return true
|
|
}
|
|
if !remaining {
|
|
return false
|
|
}
|
|
}
|
|
|
|
return true
|
|
}
|
|
|
|
return f(ts.novel) && err == nil && f(ts.upstream), err
|
|
}
|
|
|
|
func (ts tableSet) count() (uint32, error) {
|
|
f := func(css chunkSourceSet) (count uint32, err error) {
|
|
for _, haver := range css {
|
|
thisCount, err := haver.count()
|
|
|
|
if err != nil {
|
|
return 0, err
|
|
}
|
|
|
|
count += thisCount
|
|
}
|
|
return
|
|
}
|
|
|
|
novelCount, err := f(ts.novel)
|
|
|
|
if err != nil {
|
|
return 0, err
|
|
}
|
|
|
|
upCount, err := f(ts.upstream)
|
|
|
|
if err != nil {
|
|
return 0, err
|
|
}
|
|
|
|
return novelCount + upCount, nil
|
|
}
|
|
|
|
func (ts tableSet) uncompressedLen() (uint64, error) {
|
|
f := func(css chunkSourceSet) (data uint64, err error) {
|
|
for _, haver := range css {
|
|
uncmpLen, err := haver.uncompressedLen()
|
|
|
|
if err != nil {
|
|
return 0, err
|
|
}
|
|
|
|
data += uncmpLen
|
|
}
|
|
return
|
|
}
|
|
|
|
novelCount, err := f(ts.novel)
|
|
|
|
if err != nil {
|
|
return 0, err
|
|
}
|
|
|
|
upCount, err := f(ts.upstream)
|
|
|
|
if err != nil {
|
|
return 0, err
|
|
}
|
|
|
|
return novelCount + upCount, nil
|
|
}
|
|
|
|
func (ts tableSet) physicalLen() (uint64, error) {
|
|
f := func(css chunkSourceSet) (data uint64, err error) {
|
|
for _, haver := range css {
|
|
data += haver.currentSize()
|
|
}
|
|
return
|
|
}
|
|
|
|
lenNovel, err := f(ts.novel)
|
|
if err != nil {
|
|
return 0, err
|
|
}
|
|
|
|
lenUp, err := f(ts.upstream)
|
|
if err != nil {
|
|
return 0, err
|
|
}
|
|
|
|
return lenNovel + lenUp, nil
|
|
}
|
|
|
|
func (ts tableSet) close() error {
|
|
var firstErr error
|
|
setErr := func(err error) {
|
|
if err != nil && firstErr == nil {
|
|
firstErr = err
|
|
}
|
|
}
|
|
|
|
for _, t := range ts.novel {
|
|
err := t.close()
|
|
setErr(err)
|
|
}
|
|
for _, t := range ts.upstream {
|
|
err := t.close()
|
|
setErr(err)
|
|
}
|
|
return firstErr
|
|
}
|
|
|
|
// Size returns the number of tables in this tableSet.
|
|
func (ts tableSet) Size() int {
|
|
return len(ts.novel) + len(ts.upstream)
|
|
}
|
|
|
|
// append adds a memTable to an existing tableSet, compacting |mt| and
|
|
// returning a new tableSet with newly compacted table added.
|
|
func (ts tableSet) append(ctx context.Context, mt *memTable, checker refCheck, hasCache *lru.TwoQueueCache[addr, struct{}], stats *Stats) (tableSet, error) {
|
|
for i := range mt.pendingRefs {
|
|
if !mt.pendingRefs[i].has && hasCache.Contains(*mt.pendingRefs[i].a) {
|
|
mt.pendingRefs[i].has = true
|
|
}
|
|
}
|
|
|
|
sort.Sort(hasRecordByPrefix(mt.pendingRefs))
|
|
absent, err := checker(mt.pendingRefs)
|
|
if err != nil {
|
|
return tableSet{}, err
|
|
} else if absent.Size() > 0 {
|
|
return tableSet{}, fmt.Errorf("%w: found dangling references to %s", ErrDanglingRef, absent.String())
|
|
}
|
|
|
|
for _, e := range mt.pendingRefs {
|
|
hasCache.Add(*e.a, struct{}{})
|
|
}
|
|
|
|
cs, err := ts.p.Persist(ctx, mt, ts, stats)
|
|
if err != nil {
|
|
return tableSet{}, err
|
|
}
|
|
|
|
newTs := tableSet{
|
|
novel: copyChunkSourceSet(ts.novel),
|
|
upstream: copyChunkSourceSet(ts.upstream),
|
|
p: ts.p,
|
|
q: ts.q,
|
|
rl: ts.rl,
|
|
}
|
|
newTs.novel[cs.hash()] = cs
|
|
return newTs, nil
|
|
}
|
|
|
|
// flatten returns a new tableSet with |upstream| set to the union of ts.novel
|
|
// and ts.upstream.
|
|
func (ts tableSet) flatten(ctx context.Context) (tableSet, error) {
|
|
flattened := tableSet{
|
|
upstream: copyChunkSourceSet(ts.upstream),
|
|
p: ts.p,
|
|
q: ts.q,
|
|
rl: ts.rl,
|
|
}
|
|
|
|
for _, src := range ts.novel {
|
|
cnt, err := src.count()
|
|
if err != nil {
|
|
return tableSet{}, err
|
|
} else if cnt > 0 {
|
|
flattened.upstream[src.hash()] = src
|
|
}
|
|
}
|
|
return flattened, nil
|
|
}
|
|
|
|
// rebase returns a new tableSet holding the novel tables managed by |ts| and
|
|
// those specified by |specs|.
|
|
func (ts tableSet) rebase(ctx context.Context, specs []tableSpec, stats *Stats) (tableSet, error) {
|
|
// deduplicate |specs|
|
|
orig := specs
|
|
specs = make([]tableSpec, 0, len(orig))
|
|
seen := map[addr]struct{}{}
|
|
for _, spec := range orig {
|
|
if _, ok := seen[spec.name]; ok {
|
|
continue
|
|
}
|
|
seen[spec.name] = struct{}{}
|
|
// keep specs in order to play nicely with
|
|
// manifest appendix optimization
|
|
specs = append(specs, spec)
|
|
}
|
|
|
|
// copy |ts.novel|, skipping empty chunkSources
|
|
// (usually due to de-duping during table compaction)
|
|
novel := make(chunkSourceSet, len(ts.novel))
|
|
for _, t := range ts.novel {
|
|
cnt, err := t.count()
|
|
if err != nil {
|
|
return tableSet{}, err
|
|
} else if cnt == 0 {
|
|
continue
|
|
}
|
|
t2, err := t.clone()
|
|
if err != nil {
|
|
return tableSet{}, err
|
|
}
|
|
novel[t2.hash()] = t2
|
|
}
|
|
|
|
eg, ctx := errgroup.WithContext(ctx)
|
|
mu := new(sync.Mutex)
|
|
upstream := make(chunkSourceSet, len(specs))
|
|
for _, s := range specs {
|
|
// clone tables that we have already opened
|
|
if cs, ok := ts.upstream[s.name]; ok {
|
|
cl, err := cs.clone()
|
|
if err != nil {
|
|
_ = eg.Wait()
|
|
for _, cs := range upstream {
|
|
// close any opened chunkSources
|
|
_ = cs.close()
|
|
}
|
|
return tableSet{}, err
|
|
}
|
|
mu.Lock()
|
|
upstream[cl.hash()] = cl
|
|
mu.Unlock()
|
|
continue
|
|
}
|
|
// open missing tables in parallel
|
|
spec := s
|
|
eg.Go(func() error {
|
|
cs, err := ts.p.Open(ctx, spec.name, spec.chunkCount, stats)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
mu.Lock()
|
|
upstream[cs.hash()] = cs
|
|
mu.Unlock()
|
|
return nil
|
|
})
|
|
}
|
|
|
|
if err := eg.Wait(); err != nil {
|
|
for _, cs := range upstream {
|
|
// close any opened chunkSources
|
|
_ = cs.close()
|
|
}
|
|
return tableSet{}, err
|
|
}
|
|
|
|
return tableSet{
|
|
novel: novel,
|
|
upstream: upstream,
|
|
p: ts.p,
|
|
q: ts.q,
|
|
rl: ts.rl,
|
|
}, nil
|
|
}
|
|
|
|
func (ts tableSet) toSpecs() ([]tableSpec, error) {
|
|
tableSpecs := make([]tableSpec, 0, ts.Size())
|
|
for a, src := range ts.novel {
|
|
if _, ok := ts.upstream[a]; ok {
|
|
continue
|
|
}
|
|
|
|
cnt, err := src.count()
|
|
if err != nil {
|
|
return nil, err
|
|
} else if cnt > 0 {
|
|
h := src.hash()
|
|
tableSpecs = append(tableSpecs, tableSpec{h, cnt})
|
|
}
|
|
}
|
|
for _, src := range ts.upstream {
|
|
cnt, err := src.count()
|
|
if err != nil {
|
|
return nil, err
|
|
} else if cnt <= 0 {
|
|
return nil, errors.New("no upstream chunks")
|
|
}
|
|
h := src.hash()
|
|
tableSpecs = append(tableSpecs, tableSpec{h, cnt})
|
|
}
|
|
sort.Slice(tableSpecs, func(i, j int) bool {
|
|
return bytes.Compare(tableSpecs[i].name[:], tableSpecs[j].name[:]) < 0
|
|
})
|
|
return tableSpecs, nil
|
|
}
|
|
|
|
func tableSetCalcReads(ts tableSet, reqs []getRecord, blockSize uint64) (reads int, split, remaining bool, err error) {
|
|
all := copyChunkSourceSet(ts.upstream)
|
|
for a, cs := range ts.novel {
|
|
all[a] = cs
|
|
}
|
|
for _, tbl := range all {
|
|
rdr, ok := tbl.(*fileTableReader)
|
|
if !ok {
|
|
err = fmt.Errorf("chunkSource %s is not a fileTableReader", tbl.hash().String())
|
|
return
|
|
}
|
|
|
|
var n int
|
|
var more bool
|
|
n, more, err = rdr.calcReads(reqs, blockSize)
|
|
if err != nil {
|
|
return 0, false, false, err
|
|
}
|
|
|
|
reads += n
|
|
if !more {
|
|
break
|
|
}
|
|
split = true
|
|
}
|
|
return
|
|
}
|