Files
dolt/go/libraries/doltcore/env/environment.go
T

1276 lines
36 KiB
Go

// Copyright 2019-2020 Dolthub, Inc.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
package env
import (
"context"
"errors"
"fmt"
"os"
"path/filepath"
"strings"
"sync"
"time"
goerrors "gopkg.in/src-d/go-errors.v1"
"github.com/dolthub/dolt/go/cmd/dolt/doltversion"
"github.com/dolthub/dolt/go/cmd/dolt/errhand"
"github.com/dolthub/dolt/go/libraries/doltcore/creds"
"github.com/dolthub/dolt/go/libraries/doltcore/dbfactory"
"github.com/dolthub/dolt/go/libraries/doltcore/doltdb"
"github.com/dolthub/dolt/go/libraries/doltcore/grpcendpoint"
"github.com/dolthub/dolt/go/libraries/doltcore/ref"
"github.com/dolthub/dolt/go/libraries/doltcore/table/editor"
"github.com/dolthub/dolt/go/libraries/utils/concurrentmap"
"github.com/dolthub/dolt/go/libraries/utils/config"
"github.com/dolthub/dolt/go/libraries/utils/filesys"
"github.com/dolthub/dolt/go/store/chunks"
"github.com/dolthub/dolt/go/store/datas"
"github.com/dolthub/dolt/go/store/hash"
"github.com/dolthub/dolt/go/store/types"
)
const (
DefaultInitBranch = "main"
DefaultLoginUrl = "https://dolthub.com/settings/credentials"
DefaultRemotesApiHost = "doltremoteapi.dolthub.com"
DefaultRemotesApiPort = "443"
tempTablesDir = "temptf"
TmpDirName = "tmp"
)
var zeroHashStr = (hash.Hash{}).String()
var ErrStateUpdate = errors.New("error updating local data repo state")
var ErrMarshallingSchema = errors.New("error marshalling schema")
var ErrInvalidCredsFile = errors.New("invalid creds file")
var ErrRemoteAlreadyExists = errors.New("remote already exists")
var ErrInvalidRemoteURL = errors.New("remote URL invalid")
var ErrRemoteNotFound = errors.New("remote not found")
var ErrInvalidRemoteName = errors.New("remote name invalid")
var ErrBackupAlreadyExists = errors.New("backup already exists")
var ErrInvalidBackupURL = errors.New("backup URL invalid")
var ErrBackupNotFound = errors.New("backup not found")
var ErrInvalidBackupName = errors.New("backup name invalid")
var ErrFailedToDeleteBackup = errors.New("failed to delete backup")
var ErrFailedToReadFromDb = errors.New("failed to read from db")
var ErrFailedToDeleteRemote = errors.New("failed to delete remote")
var ErrFailedToWriteRepoState = errors.New("failed to write repo state")
var ErrRemoteAddressConflict = errors.New("address conflict with a remote")
var ErrDoltRepositoryNotFound = errors.New("can no longer find .dolt dir on disk")
var ErrFailedToAccessDB = goerrors.NewKind("failed to access '%s' database: can no longer find .dolt dir on disk")
var ErrDatabaseIsLocked = errors.New("the database is locked by another dolt process")
// DoltEnv holds the state of the current environment used by the cli.
type DoltEnv struct {
Version string
Config *DoltCliConfig
CfgLoadErr error
RepoState *RepoState
RSLoadErr error
loadDBOnce sync.Once
doltDB *doltdb.DoltDB
DBLoadError error
FS filesys.Filesys
urlStr string
hdp HomeDirProvider
UserPassConfig *creds.DoltCredsForPass
}
func NewDoltEnv(version string, config *DoltCliConfig, repoState *RepoState, doltDB *doltdb.DoltDB, fs filesys.Filesys) *DoltEnv {
return &DoltEnv{
Version: version,
Config: config,
RepoState: repoState,
doltDB: doltDB,
FS: fs,
}
}
func (dEnv *DoltEnv) DoltDB(ctx context.Context) *doltdb.DoltDB {
if dEnv.doltDB == nil {
LoadDoltDB(ctx, dEnv.FS, dEnv.urlStr, dEnv)
}
return dEnv.doltDB
}
func (dEnv *DoltEnv) LoadDoltDBWithParams(ctx context.Context, nbf *types.NomsBinFormat, urlStr string, fs filesys.Filesys, params map[string]interface{}) error {
if dEnv.doltDB == nil {
ddb, err := doltdb.LoadDoltDBWithParams(ctx, types.Format_Default, urlStr, fs, params)
if err != nil {
return err
}
dEnv.doltDB = ddb
}
return nil
}
// IncompleteEnv returns a DoltEnv that is incomplete. There are cases where we want to know that the structure
// of and database is correct on disk, but don't want to actually load the database.
//
// Callers of this method should not expect the returned DoltEnv to be useful as a database for running queries.
func IncompleteEnv(FS filesys.Filesys) *DoltEnv {
return &DoltEnv{
Version: doltversion.Version,
Config: nil,
RepoState: nil,
RSLoadErr: nil,
doltDB: nil,
DBLoadError: nil,
FS: FS,
urlStr: "",
}
}
func (dEnv *DoltEnv) GetRemoteDB(ctx context.Context, format *types.NomsBinFormat, r Remote, withCaching bool) (*doltdb.DoltDB, error) {
if withCaching {
return r.GetRemoteDB(ctx, format, dEnv)
} else {
return r.GetRemoteDBWithoutCaching(ctx, format, dEnv)
}
}
func (dEnv *DoltEnv) GetConfig() config.ReadableConfig {
return dEnv.Config
}
func (dEnv *DoltEnv) UrlStr() string {
return dEnv.urlStr
}
func createRepoState(fs filesys.Filesys) (*RepoState, error) {
repoState, rsErr := LoadRepoState(fs)
// deep copy remotes and backups ¯\_(ツ)_/¯ (see commit c59cbead)
if repoState != nil {
repoState.Remotes = repoState.Remotes.DeepCopy()
repoState.Backups = repoState.Backups.DeepCopy()
}
return repoState, rsErr
}
func (dEnv *DoltEnv) ReloadRepoState() error {
rs, err := createRepoState(dEnv.FS)
if err != nil {
return err
}
dEnv.RepoState = rs
return nil
}
// LoadWithoutDB creates a DoltEnv without eagerly loading the DB.
// If urlStr is non-empty, then the DB can be loaded later by calling LoadDoltDB.
func LoadWithoutDB(_ context.Context, hdp HomeDirProvider, fs filesys.Filesys, urlStr string, version string) *DoltEnv {
cfg, cfgErr := LoadDoltCliConfig(hdp, fs)
repoState, rsErr := createRepoState(fs)
return &DoltEnv{
Version: version,
Config: cfg,
CfgLoadErr: cfgErr,
RepoState: repoState,
RSLoadErr: rsErr,
FS: fs,
hdp: hdp,
urlStr: urlStr,
}
}
// Load loads the DoltEnv for the .dolt directory determined by resolving the specified urlStr with the specified Filesys.
func Load(ctx context.Context, hdp HomeDirProvider, fs filesys.Filesys, urlStr string, version string) *DoltEnv {
dEnv := LoadWithoutDB(ctx, hdp, fs, urlStr, version)
LoadDoltDB(ctx, dEnv.FS, dEnv.urlStr, dEnv)
return dEnv
}
func LoadDoltDB(ctx context.Context, fs filesys.Filesys, urlStr string, dEnv *DoltEnv) {
dEnv.loadDBOnce.Do(func() {
ddb, dbLoadErr := doltdb.LoadDoltDB(ctx, types.Format_Default, urlStr, fs)
dEnv.doltDB = ddb
dEnv.DBLoadError = dbLoadErr
dEnv.urlStr = urlStr
if dbLoadErr == nil && dEnv.HasDoltDir() {
if !dEnv.HasDoltTempTableDir() {
tmpDir, err := dEnv.TempTableFilesDir()
if err != nil {
dEnv.DBLoadError = err
}
err = dEnv.FS.MkDirs(tmpDir)
dEnv.DBLoadError = err
} else {
// fire and forget cleanup routine. Will delete as many old temp files as it can during the main commands execution.
// The process will not wait for this to finish so this may not always complete.
go func() {
// TODO dEnv.HasDoltTempTableDir() true but dEnv.TempTableFileDir() panics
tmpTableDir, err := dEnv.FS.Abs(filepath.Join(dEnv.urlStr, dbfactory.DoltDir, tempTablesDir))
if err != nil {
return
}
_ = fs.Iter(tmpTableDir, true, func(path string, size int64, isDir bool) (stop bool) {
if !isDir {
lm, exists := fs.LastModified(path)
if exists && time.Now().Sub(lm) > (time.Hour*24) {
_ = fs.DeleteFile(path)
}
}
return false
})
}()
}
}
if dEnv.RSLoadErr == nil && dbLoadErr == nil {
// If the working set isn't present in the DB, create it from the repo state. This step can be removed post 1.0.
_, err := dEnv.WorkingSet(ctx)
if errors.Is(err, doltdb.ErrWorkingSetNotFound) {
_ = dEnv.initWorkingSetFromRepoState(ctx)
} else if err != nil {
dEnv.RSLoadErr = err
}
}
})
}
func GetDefaultInitBranch(cfg config.ReadableConfig) string {
return GetStringOrDefault(cfg, config.InitBranchName, DefaultInitBranch)
}
// Valid returns whether this environment has been properly initialized. This is useful because although every command
// gets a DoltEnv, not all of them require it, and we allow invalid dolt envs to be passed around for this reason.
func (dEnv *DoltEnv) Valid() bool {
return dEnv != nil && dEnv.CfgLoadErr == nil && dEnv.DBLoadError == nil && dEnv.HasDoltDir() && dEnv.HasDoltDataDir()
}
// initWorkingSetFromRepoState sets the working set for the env's head to mirror the contents of the repo state file.
// This is only necessary to migrate repos written before this method was introduced, and can be removed after 1.0
func (dEnv *DoltEnv) initWorkingSetFromRepoState(ctx context.Context) error {
headRef, err := dEnv.RepoStateReader().CWBHeadRef(ctx)
if err != nil {
return err
}
wsRef, err := ref.WorkingSetRefForHead(headRef)
if err != nil {
return err
}
headRoot, err := dEnv.HeadRoot(ctx)
if err != nil {
return err
}
stagedRoot := headRoot
if len(dEnv.RepoState.staged) != 0 && dEnv.RepoState.staged != zeroHashStr {
stagedHash, ok := hash.MaybeParse(dEnv.RepoState.staged)
if !ok {
return fmt.Errorf("Corrupt repo, invalid staged hash %s", stagedHash)
}
stagedRoot, err = dEnv.DoltDB(ctx).ReadRootValue(ctx, stagedHash)
if err != nil {
return err
}
}
workingRoot := stagedRoot
if len(dEnv.RepoState.working) != 0 && dEnv.RepoState.working != zeroHashStr {
workingHash, ok := hash.MaybeParse(dEnv.RepoState.working)
if !ok {
return fmt.Errorf("Corrupt repo, invalid working hash %s", workingHash)
}
workingRoot, err = dEnv.DoltDB(ctx).ReadRootValue(ctx, workingHash)
if err != nil {
return err
}
}
mergeState, err := mergeStateToMergeState(ctx, dEnv.RepoState.merge, dEnv.DoltDB(ctx))
if err != nil {
return err
}
ws := doltdb.EmptyWorkingSet(wsRef).WithWorkingRoot(workingRoot).WithStagedRoot(stagedRoot).WithMergeState(mergeState)
return dEnv.UpdateWorkingSet(ctx, ws)
}
func mergeStateToMergeState(ctx context.Context, mergeState *mergeState, db *doltdb.DoltDB) (*doltdb.MergeState, error) {
if mergeState == nil {
return nil, nil
}
cs, err := doltdb.NewCommitSpec(mergeState.Commit)
if err != nil {
panic("Corrupted repostate. Active merge state is not valid.")
}
optCmt, err := db.Resolve(ctx, cs, nil)
if err != nil {
return nil, err
}
commit, ok := optCmt.ToCommit()
if !ok {
return nil, doltdb.ErrGhostCommitEncountered
}
pmwh := hash.Parse(mergeState.PreMergeWorking)
pmwr, err := db.ReadRootValue(ctx, pmwh)
if err != nil {
return nil, err
}
return doltdb.MergeStateFromCommitAndWorking(commit, pmwr), nil
}
// HasDoltDir returns true if the .dolt directory exists and is a valid directory
func (dEnv *DoltEnv) HasDoltDir() bool {
return dEnv.hasDoltDir("./")
}
func (dEnv *DoltEnv) HasDoltDataDir() bool {
exists, isDir := dEnv.FS.Exists(dbfactory.DoltDataDir)
return exists && isDir
}
// HasDoltSqlServerInfo returns true if this Dolt environment has a sql-server.info file, indicating
// that a sql-server is running from this Dolt environment.
func (dEnv *DoltEnv) HasDoltSqlServerInfo() bool {
exists, _ := dEnv.FS.Exists(filepath.Join(dbfactory.DoltDir, "sql-server.info"))
return exists
}
func (dEnv *DoltEnv) HasDoltTempTableDir() bool {
tmpDir, err := dEnv.TempTableFilesDir()
if err != nil {
return false
}
ex, _ := dEnv.FS.Exists(tmpDir)
return ex
}
func mustAbs(dEnv *DoltEnv, path ...string) string {
absPath, err := dEnv.FS.Abs(filepath.Join(path...))
if err != nil {
panic(err)
}
return absPath
}
// GetDoltDir returns the path to the .dolt directory
func (dEnv *DoltEnv) GetDoltDir() string {
if !dEnv.HasDoltDataDir() {
return ""
}
return mustAbs(dEnv, dbfactory.DoltDir)
}
func (dEnv *DoltEnv) hasDoltDir(path string) bool {
exists, isDir := dEnv.FS.Exists(mustAbs(dEnv, dbfactory.DoltDir))
return exists && isDir
}
// HasLocalConfig returns true if a repository local config file
func (dEnv *DoltEnv) HasLocalConfig() bool {
_, ok := dEnv.Config.GetConfig(LocalConfig)
return ok
}
func (dEnv *DoltEnv) bestEffortDeleteAll(dir string) {
fileToIsDir := make(map[string]bool)
dEnv.FS.Iter(dir, false, func(path string, size int64, isDir bool) (stop bool) {
fileToIsDir[path] = isDir
return false
})
for path, isDir := range fileToIsDir {
if isDir {
dEnv.FS.Delete(path, true)
} else {
dEnv.FS.DeleteFile(path)
}
}
}
// InitRepo takes an empty directory and initializes it with a .dolt directory containing repo state, uncommitted license and readme, and creates a noms
// database with dolt structure.
func (dEnv *DoltEnv) InitRepo(ctx context.Context, nbf *types.NomsBinFormat, name, email, branchName string) error { // should remove name and email args
return dEnv.InitRepoWithTime(ctx, nbf, name, email, branchName, datas.CommitterDate())
}
func (dEnv *DoltEnv) InitRepoWithTime(ctx context.Context, nbf *types.NomsBinFormat, name, email, branchName string, t time.Time) error { // should remove name and email args
return dEnv.InitRepoWithCommitMetaGenerator(ctx, nbf, branchName, datas.MakeCommitMetaGenerator(name, email, t))
}
func (dEnv *DoltEnv) InitRepoWithCommitMetaGenerator(ctx context.Context, nbf *types.NomsBinFormat, branchName string, commitMeta datas.CommitMetaGenerator) error {
doltDir, err := dEnv.createDirectories(".")
if err != nil {
return err
}
err = dEnv.configureRepo(doltDir)
if err == nil {
err = dEnv.InitDBAndRepoStateWithCommitMetaGenerator(ctx, nbf, branchName, commitMeta)
}
if err != nil {
dEnv.bestEffortDeleteAll(dbfactory.DoltDir)
}
return err
}
func (dEnv *DoltEnv) InitRepoWithNoData(ctx context.Context, nbf *types.NomsBinFormat) error {
doltDir, err := dEnv.createDirectories(".")
if err != nil {
return err
}
err = dEnv.configureRepo(doltDir)
if err != nil {
dEnv.bestEffortDeleteAll(dbfactory.DoltDir)
return err
}
dEnv.doltDB, err = doltdb.LoadDoltDB(ctx, nbf, dEnv.urlStr, dEnv.FS)
return err
}
func (dEnv *DoltEnv) createDirectories(dir string) (string, error) {
absPath, err := dEnv.FS.Abs(dir)
if err != nil {
return "", err
}
exists, isDir := dEnv.FS.Exists(absPath)
if !exists {
return "", fmt.Errorf("'%s' does not exist so could not create '%s", absPath, dbfactory.DoltDataDir)
} else if !isDir {
return "", fmt.Errorf("'%s' exists but it's a file not a directory", absPath)
}
if dEnv.hasDoltDir(dir) {
// Special case a completely empty directory, or one which has a "tmp" directory but nothing else.
// The `.dolt/tmp` directory is created while verifying that we can rename table files which is early
// in the startup process. It will only exist if we need it because the TMPDIR environment variable is set to
// a path which is on a different partition than the .dolt directory.
dotDolt := mustAbs(dEnv, dbfactory.DoltDir)
entries, err := os.ReadDir(dotDolt)
if err != nil {
return "", err
}
if len(entries) == 1 {
entry := entries[0]
if (entry.IsDir() && entry.Name() != TmpDirName) || (!entry.IsDir() && entry.Name() != configFile) {
return "", fmt.Errorf(".dolt directory already exists at '%s'", dir)
}
} else if len(entries) != 0 {
return "", fmt.Errorf(".dolt directory already exists at '%s'", dir)
}
}
absDataDir := filepath.Join(absPath, dbfactory.DoltDataDir)
err = dEnv.FS.MkDirs(absDataDir)
if err != nil {
return "", fmt.Errorf("unable to make directory '%s', cause: %s", absDataDir, err.Error())
}
tmpDir, err := dEnv.TempTableFilesDir()
if err != nil {
return "", err
}
err = dEnv.FS.MkDirs(tmpDir)
if err != nil {
return "", fmt.Errorf("unable to make directory '%s', cause: %s", tmpDir, err.Error())
}
return filepath.Join(absPath, dbfactory.DoltDir), nil
}
func (dEnv *DoltEnv) configureRepo(doltDir string) error {
configDir, err := dEnv.FS.Abs(".")
if err != nil {
return fmt.Errorf("unable to resolve current path to create repo local config file: %s", err.Error())
}
err = dEnv.Config.CreateLocalConfig(configDir, map[string]string{})
if err != nil {
return fmt.Errorf("failed creating file %s", getLocalConfigPath())
}
return nil
}
// Inits the dolt DB of this environment with an empty commit at the time given and writes default docs to disk.
// Writes new repo state with a main branch and current root hash.
func (dEnv *DoltEnv) InitDBAndRepoState(ctx context.Context, nbf *types.NomsBinFormat, name, email, branchName string, t time.Time) error {
return dEnv.InitDBAndRepoStateWithCommitMetaGenerator(ctx, nbf, branchName, datas.MakeCommitMetaGenerator(name, email, t))
}
func (dEnv *DoltEnv) InitDBAndRepoStateWithCommitMetaGenerator(ctx context.Context, nbf *types.NomsBinFormat, branchName string, commitMeta datas.CommitMetaGenerator) error {
err := dEnv.InitDBWithCommitMetaGenerator(ctx, nbf, branchName, commitMeta)
if err != nil {
return err
}
return dEnv.InitializeRepoState(ctx, branchName)
}
// Inits the dolt DB of this environment with an empty commit at the time given and writes default docs to disk.
// Does not update repo state.
func (dEnv *DoltEnv) InitDBWithTime(ctx context.Context, nbf *types.NomsBinFormat, name, email, branchName string, t time.Time) error {
return dEnv.InitDBWithCommitMetaGenerator(ctx, nbf, branchName, datas.MakeCommitMetaGenerator(name, email, t))
}
func (dEnv *DoltEnv) InitDBWithCommitMetaGenerator(ctx context.Context, nbf *types.NomsBinFormat, branchName string, commitMeta datas.CommitMetaGenerator) error {
var err error
dEnv.doltDB, err = doltdb.LoadDoltDB(ctx, nbf, dEnv.urlStr, dEnv.FS)
if err != nil {
return err
}
err = dEnv.DoltDB(ctx).WriteEmptyRepoWithCommitMetaGenerator(ctx, branchName, commitMeta)
if err != nil {
return fmt.Errorf("%w: %v", doltdb.ErrNomsIO, err)
}
return nil
}
// InitializeRepoState writes a default repo state to disk, consisting of a main branch and current root hash value.
func (dEnv *DoltEnv) InitializeRepoState(ctx context.Context, branchName string) error {
commit, err := dEnv.DoltDB(ctx).ResolveCommitRef(ctx, ref.NewBranchRef(branchName))
if err != nil {
return err
}
root, err := commit.GetRootValue(ctx)
if err != nil {
return err
}
dEnv.RepoState, err = CreateRepoState(dEnv.FS, branchName)
if err != nil {
return ErrStateUpdate
}
// TODO: combine into one update
err = dEnv.UpdateWorkingRoot(ctx, root)
if err != nil {
return err
}
err = dEnv.UpdateStagedRoot(ctx, root)
if err != nil {
return err
}
dEnv.RSLoadErr = nil
return nil
}
type RootsProvider[C doltdb.Context] interface {
GetRoots(ctx C) (doltdb.Roots, error)
}
// Roots returns the roots for this environment
func (dEnv *DoltEnv) Roots(ctx context.Context) (doltdb.Roots, error) {
ws, err := dEnv.WorkingSet(ctx)
if err != nil {
return doltdb.Roots{}, err
}
headRoot, err := dEnv.HeadRoot(ctx)
if err != nil {
return doltdb.Roots{}, err
}
return doltdb.Roots{
Head: headRoot,
Working: ws.WorkingRoot(),
Staged: ws.StagedRoot(),
}, nil
}
// RecoveryRoots returns the roots for this environment in the case that the
// currently checked out branch has been deleted or HEAD has been updated in a
// non-principled way to point to a branch that does not exist. This is used by
// `dolt checkout`, in particular, to go forward with a `dolt checkout` of an
// existing branch in the degraded state where the current branch was deleted.
func (dEnv *DoltEnv) RecoveryRoots(ctx context.Context) (doltdb.Roots, error) {
ws, err := dEnv.WorkingSet(ctx)
if err != nil {
return doltdb.Roots{}, err
}
headRoot, err := dEnv.HeadRoot(ctx)
if err == doltdb.ErrBranchNotFound {
headRoot = ws.StagedRoot()
err = nil
}
if err != nil {
return doltdb.Roots{}, err
}
return doltdb.Roots{
Head: headRoot,
Working: ws.WorkingRoot(),
Staged: ws.StagedRoot(),
}, nil
}
// UpdateRoots updates the working and staged roots for this environment
func (dEnv *DoltEnv) UpdateRoots(ctx context.Context, roots doltdb.Roots) error {
ws, err := dEnv.WorkingSet(ctx)
if err == doltdb.ErrWorkingSetNotFound {
// first time updating roots
wsRef, err := ref.WorkingSetRefForHead(dEnv.RepoState.CWBHeadRef())
if err != nil {
return err
}
ws = doltdb.EmptyWorkingSet(wsRef)
} else if err != nil {
return err
}
return dEnv.UpdateWorkingSet(ctx, ws.WithWorkingRoot(roots.Working).WithStagedRoot(roots.Staged))
}
// WorkingRoot returns the working root for the current working branch
func (dEnv *DoltEnv) WorkingRoot(ctx context.Context) (doltdb.RootValue, error) {
workingSet, err := dEnv.WorkingSet(ctx)
if err != nil {
return nil, err
}
return workingSet.WorkingRoot(), nil
}
func (dEnv *DoltEnv) WorkingSet(ctx context.Context) (*doltdb.WorkingSet, error) {
return WorkingSet(ctx, dEnv.DoltDB(ctx), dEnv.RepoStateReader())
}
func WorkingSet[C doltdb.Context](ctx C, ddb *doltdb.DoltDB, rsr RepoStateReader[C]) (*doltdb.WorkingSet, error) {
headRef, err := rsr.CWBHeadRef(ctx)
if err != nil {
return nil, err
}
workingSetRef, err := ref.WorkingSetRefForHead(headRef)
if err != nil {
return nil, err
}
workingSet, err := ddb.ResolveWorkingSet(ctx, workingSetRef)
if err != nil {
return nil, err
}
return workingSet, nil
}
// UpdateWorkingRoot updates the working root for the current working branch to the root value given.
// This method can fail if another client updates the working root at the same time.
func (dEnv *DoltEnv) UpdateWorkingRoot(ctx context.Context, newRoot doltdb.RootValue) error {
var h hash.Hash
var wsRef ref.WorkingSetRef
ws, err := dEnv.WorkingSet(ctx)
if err == doltdb.ErrWorkingSetNotFound {
// first time updating root
wsRef, err = ref.WorkingSetRefForHead(dEnv.RepoState.CWBHeadRef())
if err != nil {
return err
}
ws = doltdb.EmptyWorkingSet(wsRef).WithWorkingRoot(newRoot).WithStagedRoot(newRoot)
} else if err != nil {
return err
} else {
h, err = ws.HashOf()
if err != nil {
return err
}
wsRef = ws.Ref()
}
return dEnv.DoltDB(ctx).UpdateWorkingSet(ctx, wsRef, ws.WithWorkingRoot(newRoot), h, dEnv.workingSetMeta(), nil)
}
// UpdateWorkingSet updates the working set for the current working branch to the value given.
// This method can fail if another client updates the working set at the same time.
func (dEnv *DoltEnv) UpdateWorkingSet(ctx context.Context, ws *doltdb.WorkingSet) error {
currentWs, err := dEnv.WorkingSet(ctx)
if err != doltdb.ErrWorkingSetNotFound && err != nil {
return err
}
var h hash.Hash
if currentWs != nil {
h, err = currentWs.HashOf()
if err != nil {
return err
}
}
return dEnv.DoltDB(ctx).UpdateWorkingSet(ctx, ws.Ref(), ws, h, dEnv.workingSetMeta(), nil)
}
type repoStateReader struct {
*DoltEnv
}
func (r *repoStateReader) CWBHeadRef(context.Context) (ref.DoltRef, error) {
if r.RepoState == nil && r.RSLoadErr != nil {
return nil, r.RSLoadErr
}
return r.RepoState.CWBHeadRef(), nil
}
func (r *repoStateReader) CWBHeadSpec(context.Context) (*doltdb.CommitSpec, error) {
if r.RepoState == nil && r.RSLoadErr != nil {
return nil, r.RSLoadErr
}
return r.RepoState.CWBHeadSpec(), nil
}
func (dEnv *DoltEnv) RepoStateReader() RepoStateReader[context.Context] {
return &repoStateReader{dEnv}
}
type repoStateWriter struct {
*DoltEnv
}
func (r *repoStateWriter) SetCWBHeadRef(ctx context.Context, marshalableRef ref.MarshalableRef) error {
if r.RepoState == nil && r.RSLoadErr != nil {
return r.RSLoadErr
}
r.RepoState.Head = marshalableRef
err := r.RepoState.Save(r.FS)
if err != nil {
return ErrStateUpdate
}
return nil
}
func (r *repoStateWriter) AddRemote(remote Remote) error {
return r.DoltEnv.AddRemote(remote)
}
func (r *repoStateWriter) AddBackup(remote Remote) error {
return r.DoltEnv.AddBackup(remote)
}
func (r *repoStateWriter) RemoveRemote(ctx context.Context, name string) error {
return r.DoltEnv.RemoveRemote(ctx, name)
}
func (r *repoStateWriter) RemoveBackup(ctx context.Context, name string) error {
return r.DoltEnv.RemoveBackup(ctx, name)
}
func (dEnv *DoltEnv) RepoStateWriter() RepoStateWriter {
return &repoStateWriter{dEnv}
}
func (dEnv *DoltEnv) HeadRoot(ctx context.Context) (doltdb.RootValue, error) {
commit, err := dEnv.HeadCommit(ctx)
if err != nil {
return nil, err
}
return commit.GetRootValue(ctx)
}
func (dEnv *DoltEnv) HeadCommit(ctx context.Context) (*doltdb.Commit, error) {
return dEnv.DoltDB(ctx).ResolveCommitRef(ctx, dEnv.RepoState.CWBHeadRef())
}
func (dEnv *DoltEnv) DbData(ctx context.Context) DbData[context.Context] {
return DbData[context.Context]{
Ddb: dEnv.DoltDB(ctx),
Rsw: dEnv.RepoStateWriter(),
Rsr: dEnv.RepoStateReader(),
}
}
// StagedRoot returns the staged root value in the current working set
func (dEnv *DoltEnv) StagedRoot(ctx context.Context) (doltdb.RootValue, error) {
workingSet, err := dEnv.WorkingSet(ctx)
if err != nil {
return nil, err
}
return workingSet.StagedRoot(), nil
}
// UpdateStagedRoot updates the staged root for the current working branch. This can fail if multiple clients attempt
// to update at the same time.
func (dEnv *DoltEnv) UpdateStagedRoot(ctx context.Context, newRoot doltdb.RootValue) error {
var h hash.Hash
var wsRef ref.WorkingSetRef
ws, err := dEnv.WorkingSet(ctx)
if err == doltdb.ErrWorkingSetNotFound {
// first time updating root
wsRef, err = ref.WorkingSetRefForHead(dEnv.RepoState.CWBHeadRef())
if err != nil {
return err
}
ws = doltdb.EmptyWorkingSet(wsRef).WithWorkingRoot(newRoot).WithStagedRoot(newRoot)
} else if err != nil {
return err
} else {
h, err = ws.HashOf()
if err != nil {
return err
}
wsRef = ws.Ref()
}
return dEnv.DoltDB(ctx).UpdateWorkingSet(ctx, wsRef, ws.WithStagedRoot(newRoot), h, dEnv.workingSetMeta(), nil)
}
func (dEnv *DoltEnv) AbortMerge(ctx context.Context) error {
ws, err := dEnv.WorkingSet(ctx)
if err != nil {
return err
}
h, err := ws.HashOf()
if err != nil {
return err
}
return dEnv.DoltDB(ctx).UpdateWorkingSet(ctx, ws.Ref(), ws.AbortMerge(), h, dEnv.workingSetMeta(), nil)
}
func (dEnv *DoltEnv) workingSetMeta() *datas.WorkingSetMeta {
return dEnv.NewWorkingSetMeta("updated from dolt environment")
}
func (dEnv *DoltEnv) NewWorkingSetMeta(message string) *datas.WorkingSetMeta {
return &datas.WorkingSetMeta{
Name: dEnv.Config.GetStringOrDefault(config.UserNameKey, ""),
Email: dEnv.Config.GetStringOrDefault(config.UserEmailKey, ""),
Timestamp: uint64(time.Now().Unix()),
Description: message,
}
}
func (dEnv *DoltEnv) CredsDir() (string, error) {
return getCredsDir(dEnv.hdp)
}
func (dEnv *DoltEnv) UserDoltCreds() (creds.DoltCreds, bool, error) {
kid, err := dEnv.Config.GetString(config.UserCreds)
if err == nil && kid != "" {
dir, err := dEnv.CredsDir()
if err != nil {
// not sure why you wouldn't be able to get the creds dir.
panic(err)
}
c, err := creds.JWKCredsReadFromFile(dEnv.FS, filepath.Join(dir, kid+".jwk"))
return c, c.IsPrivKeyValid() && c.IsPubKeyValid(), err
}
return creds.DoltCreds{}, false, nil
}
// GetGRPCDialParams implements dbfactory.GRPCDialProvider
func (dEnv *DoltEnv) GetGRPCDialParams(config grpcendpoint.Config) (dbfactory.GRPCRemoteConfig, error) {
return NewGRPCDialProviderFromDoltEnv(dEnv).GetGRPCDialParams(config)
}
func (dEnv *DoltEnv) GetRemotes() (*concurrentmap.Map[string, Remote], error) {
if dEnv.RSLoadErr != nil {
return nil, dEnv.RSLoadErr
}
return dEnv.RepoState.Remotes, nil
}
// CheckRemoteAddressConflict checks whether any backups or remotes share the given URL. Returns the first remote if multiple match.
// Returns NoRemote and false if none match.
func CheckRemoteAddressConflict(absUrl string, remotes *concurrentmap.Map[string, Remote], backups *concurrentmap.Map[string, Remote]) (Remote, bool) {
if remotes != nil {
var rm *Remote
remotes.Iter(func(key string, value Remote) bool {
if value.Url == absUrl {
rm = &value
return false
}
return true
})
if rm != nil {
return *rm, true
}
}
if backups != nil {
var rm *Remote
backups.Iter(func(key string, value Remote) bool {
if value.Url == absUrl {
rm = &value
return false
}
return true
})
if rm != nil {
return *rm, true
}
}
return NoRemote, false
}
func (dEnv *DoltEnv) AddRemote(r Remote) error {
if _, ok := dEnv.RepoState.Remotes.Get(r.Name); ok {
return ErrRemoteAlreadyExists
}
if strings.IndexAny(r.Name, " \t\n\r./\\!@#$%^&*(){}[],.<>'\"?=+|") != -1 {
return ErrInvalidRemoteName
}
_, absRemoteUrl, err := GetAbsRemoteUrl(dEnv.FS, dEnv.Config, r.Url)
if err != nil {
return fmt.Errorf("%w; %s", ErrInvalidRemoteURL, err.Error())
}
// can have multiple remotes with the same address, but no conflicting backups
if rem, found := CheckRemoteAddressConflict(absRemoteUrl, nil, dEnv.RepoState.Backups); found {
return fmt.Errorf("%w: '%s' -> %s", ErrRemoteAddressConflict, rem.Name, rem.Url)
}
r.Url = absRemoteUrl
dEnv.RepoState.AddRemote(r)
return dEnv.RepoState.Save(dEnv.FS)
}
func (dEnv *DoltEnv) GetBackups() (*concurrentmap.Map[string, Remote], error) {
if dEnv.RSLoadErr != nil {
return nil, dEnv.RSLoadErr
}
return dEnv.RepoState.Backups, nil
}
func (dEnv *DoltEnv) AddBackup(r Remote) error {
if _, ok := dEnv.RepoState.Backups.Get(r.Name); ok {
return ErrBackupAlreadyExists
}
if strings.IndexAny(r.Name, " \t\n\r./\\!@#$%^&*(){}[],.<>'\"?=+|") != -1 {
return ErrInvalidBackupName
}
_, absRemoteUrl, err := GetAbsRemoteUrl(dEnv.FS, dEnv.Config, r.Url)
if err != nil {
return fmt.Errorf("%w; %s", ErrInvalidBackupURL, err.Error())
}
// no conflicting remote or backup addresses
if rem, found := CheckRemoteAddressConflict(absRemoteUrl, dEnv.RepoState.Remotes, dEnv.RepoState.Backups); found {
return fmt.Errorf("%w: '%s' -> %s", ErrRemoteAddressConflict, rem.Name, rem.Url)
}
r.Url = absRemoteUrl
dEnv.RepoState.AddBackup(r)
return dEnv.RepoState.Save(dEnv.FS)
}
func (dEnv *DoltEnv) RemoveRemote(ctx context.Context, name string) error {
remote, ok := dEnv.RepoState.Remotes.Get(name)
if !ok {
return ErrRemoteNotFound
}
ddb := dEnv.DoltDB(ctx)
refs, err := ddb.GetRemoteRefs(ctx)
if err != nil {
return fmt.Errorf("%w: %s", ErrFailedToReadFromDb, err.Error())
}
for _, r := range refs {
rr := r.(ref.RemoteRef)
if rr.GetRemote() == remote.Name {
err = ddb.DeleteBranch(ctx, rr, nil)
if err != nil {
return fmt.Errorf("%w; failed to delete remote tracking ref '%s'; %s", ErrFailedToDeleteRemote, rr.String(), err.Error())
}
}
}
dEnv.RepoState.RemoveRemote(remote)
err = dEnv.RepoState.Save(dEnv.FS)
if err != nil {
return ErrFailedToWriteRepoState
}
return nil
}
func (dEnv *DoltEnv) RemoveBackup(ctx context.Context, name string) error {
backup, ok := dEnv.RepoState.Backups.Get(name)
if !ok {
return ErrBackupNotFound
}
dEnv.RepoState.RemoveBackup(backup)
err := dEnv.RepoState.Save(dEnv.FS)
if err != nil {
return ErrFailedToWriteRepoState
}
return nil
}
func (dEnv *DoltEnv) GetBranches() (*concurrentmap.Map[string, BranchConfig], error) {
if dEnv.RSLoadErr != nil {
return nil, dEnv.RSLoadErr
}
return dEnv.RepoState.Branches, nil
}
func (dEnv *DoltEnv) UpdateBranch(name string, new BranchConfig) error {
if dEnv.RSLoadErr != nil {
return dEnv.RSLoadErr
}
dEnv.RepoState.Branches.Set(name, new)
err := dEnv.RepoState.Save(dEnv.FS)
if err != nil {
return ErrFailedToWriteRepoState
}
return nil
}
var ErrNotACred = errors.New("not a valid credential key id or public key")
func (dEnv *DoltEnv) FindCreds(credsDir, pubKeyOrId string) (string, error) {
if !creds.B32CredsByteSet.ContainsAll([]byte(pubKeyOrId)) {
return "", creds.ErrBadB32CredsEncoding
}
if len(pubKeyOrId) == creds.B32EncodedPubKeyLen {
pubKeyOrId, _ = creds.PubKeyStrToKIDStr(pubKeyOrId)
}
if len(pubKeyOrId) != creds.B32EncodedKeyIdLen {
return "", ErrNotACred
}
path := mustAbs(dEnv, credsDir, pubKeyOrId+creds.JWKFileExtension)
exists, isDir := dEnv.FS.Exists(path)
if isDir {
return path, filesys.ErrIsDir
} else if !exists {
return "", creds.ErrCredsNotFound
} else {
return path, nil
}
}
func (dEnv *DoltEnv) FindRef(ctx context.Context, refStr string) (ref.DoltRef, error) {
localRef := ref.NewBranchRef(refStr)
if hasRef, err := dEnv.DoltDB(ctx).HasRef(ctx, localRef); err != nil {
return nil, err
} else if hasRef {
return localRef, nil
} else {
if strings.HasPrefix(refStr, "remotes/") {
refStr = refStr[len("remotes/"):]
}
slashIdx := strings.IndexRune(refStr, '/')
if slashIdx > 0 {
remoteName := refStr[:slashIdx]
if _, ok := dEnv.RepoState.Remotes.Get(remoteName); ok {
remoteRef, err := ref.NewRemoteRefFromPathStr(refStr)
if err != nil {
return nil, err
}
if hasRef, err = dEnv.DoltDB(ctx).HasRef(ctx, remoteRef); err != nil {
return nil, err
} else if hasRef {
return remoteRef, nil
}
}
}
}
return nil, doltdb.ErrBranchNotFound
}
// GetRefSpecs takes an optional remoteName and returns all refspecs associated with that remote. If "" is passed as
// the remoteName then the default remote is used.
func GetRefSpecs[C doltdb.Context](rsr RepoStateReader[C], remoteName string) ([]ref.RemoteRefSpec, error) {
var remote Remote
var err error
remotes, err := rsr.GetRemotes()
if err != nil {
return nil, err
}
if remoteName == "" {
remote, err = GetDefaultRemote(rsr)
} else if r, ok := remotes.Get(remoteName); ok {
remote = r
} else {
err = ErrInvalidRepository.New(remoteName)
}
if err != nil {
return nil, err
}
var refSpecs []ref.RemoteRefSpec
for _, fs := range remote.FetchSpecs {
rs, err := ref.ParseRefSpecForRemote(remote.Name, fs)
if err != nil {
return nil, errhand.BuildDError("error: for '%s', '%s' is not a valid refspec.", remote.Name, fs).Build()
}
if rrs, ok := rs.(ref.RemoteRefSpec); !ok {
return nil, fmt.Errorf("%w; '%s' is not a valid refspec referring to a remote tracking branch", ref.ErrInvalidRefSpec, remote.Name)
} else if rrs.GetRemote() != remote.Name {
return nil, ErrInvalidRefSpecRemote
} else {
refSpecs = append(refSpecs, rrs)
}
}
return refSpecs, nil
}
var ErrInvalidRefSpecRemote = errors.New("refspec refers to different remote")
var ErrNoRemote = errors.New("no remote")
var ErrUnknownRemote = errors.New("unknown remote")
var ErrCantDetermineDefault = errors.New("unable to determine the default remote")
// GetDefaultRemote gets the default remote for the environment. Not fully implemented yet. Needs to support multiple
// repos and a configurable default.
func GetDefaultRemote[C doltdb.Context](rsr RepoStateReader[C]) (Remote, error) {
remotes, err := rsr.GetRemotes()
if err != nil {
return NoRemote, err
}
remotesLen := remotes.Len()
if remotesLen == 0 {
return NoRemote, ErrNoRemote
} else if remotesLen == 1 {
var remote *Remote
remotes.Iter(func(key string, value Remote) bool {
remote = &value
return false
})
if remote != nil {
return *remote, nil
}
}
if remote, ok := remotes.Get("origin"); ok {
return remote, nil
}
return NoRemote, ErrCantDetermineDefault
}
// GetUserHomeDir returns the user's home dir
// based on current filesys
func (dEnv *DoltEnv) GetUserHomeDir() (string, error) {
return getHomeDir(dEnv.hdp)
}
func (dEnv *DoltEnv) TempTableFilesDir() (string, error) {
doltDir := dEnv.GetDoltDir()
if doltDir == "" {
return "", ErrDoltRepositoryNotFound
}
absPath, err := dEnv.FS.Abs(filepath.Join(doltDir, tempTablesDir))
if err != nil {
return "", err
}
return absPath, nil
}
func (dEnv *DoltEnv) DbEaFactory(ctx context.Context) editor.DbEaFactory {
tmpDir, err := dEnv.TempTableFilesDir()
if err != nil {
return nil
}
return editor.NewDbEaFactory(tmpDir, dEnv.DoltDB(ctx).ValueReadWriter())
}
func (dEnv *DoltEnv) BulkDbEaFactory(ctx context.Context) editor.DbEaFactory {
tmpDir, err := dEnv.TempTableFilesDir()
if err != nil {
return nil
}
return editor.NewBulkImportTEAFactory(dEnv.DoltDB(ctx).ValueReadWriter(), tmpDir)
}
func (dEnv *DoltEnv) IsAccessModeReadOnly(ctx context.Context) bool {
return dEnv.DoltDB(ctx).AccessMode() == chunks.ExclusiveAccessMode_ReadOnly
}