1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377
|
package objectpool
import (
"bufio"
"bytes"
"context"
"fmt"
"io"
"strings"
"github.com/grpc-ecosystem/go-grpc-middleware/logging/logrus/ctxlogrus"
"github.com/sirupsen/logrus"
"gitlab.com/gitlab-org/gitaly/v16/internal/git"
"gitlab.com/gitlab-org/gitaly/v16/internal/git/housekeeping"
"gitlab.com/gitlab-org/gitaly/v16/internal/git/localrepo"
"gitlab.com/gitlab-org/gitaly/v16/internal/git/stats"
"gitlab.com/gitlab-org/gitaly/v16/internal/git/updateref"
"gitlab.com/gitlab-org/gitaly/v16/internal/gitaly/transaction"
"gitlab.com/gitlab-org/gitaly/v16/internal/structerr"
"gitlab.com/gitlab-org/gitaly/v16/internal/transaction/voting"
)
var objectPoolRefspec = fmt.Sprintf("+refs/*:%s/*", git.ObjectPoolRefNamespace)
// FetchFromOrigin initializes the pool and fetches the objects from its origin repository
func (o *ObjectPool) FetchFromOrigin(ctx context.Context, origin *localrepo.Repo) error {
logger := ctxlogrus.Extract(ctx)
if !o.Exists() {
return structerr.NewInvalidArgument("object pool does not exist")
}
originPath, err := origin.Path()
if err != nil {
return fmt.Errorf("computing origin repo's path: %w", err)
}
if err := o.housekeepingManager.CleanStaleData(ctx, o.Repo, housekeeping.DefaultStaleDataCleanup()); err != nil {
return fmt.Errorf("cleaning stale data: %w", err)
}
if err := o.logStats(ctx, logger.WithField("when", "before fetch")); err != nil {
return fmt.Errorf("computing stats before fetch: %w", err)
}
// Ideally we wouldn't want to prune old references at all so that we can keep alive all
// objects without having to create loads of dangling references. But unfortunately keeping
// around old refs can lead to D/F conflicts between old references that have since
// been deleted in the pool and new references that have been added in the pool member we're
// fetching from. E.g. if we have the old reference `refs/heads/branch` and the pool member
// has replaced that since with a new reference `refs/heads/branch/conflict` then
// the fetch would now always fail because of that conflict.
//
// Due to the lack of an alternative to resolve that conflict we are thus forced to enable
// pruning. This isn't too bad given that we know to keep alive the old objects via dangling
// refs anyway, but I'd sleep easier if we didn't have to do this.
//
// Note that we need to perform the pruning separately from the fetch: if the fetch is using
// `--atomic` and `--prune` together then it still wouldn't be able to recover from the D/F
// conflict. So we first to a preliminary prune that only prunes refs without fetching
// objects yet to avoid that scenario.
if err := o.pruneReferences(ctx, origin); err != nil {
return fmt.Errorf("pruning references: %w", err)
}
var stderr bytes.Buffer
if err := o.Repo.ExecAndWait(ctx,
git.Command{
Name: "fetch",
Flags: []git.Option{
git.Flag{Name: "--quiet"},
git.Flag{Name: "--atomic"},
// We already fetch tags via our refspec, so we don't
// want to fetch them a second time via Git's default
// tag refspec.
git.Flag{Name: "--no-tags"},
// We don't need FETCH_HEAD, and it can potentially be hundreds of
// megabytes when doing a mirror-sync of repos with huge numbers of
// references.
git.Flag{Name: "--no-write-fetch-head"},
// Disable showing forced updates, which may take a considerable
// amount of time to compute. We don't display any output anyway,
// which makes this computation kind of moot.
git.Flag{Name: "--no-show-forced-updates"},
},
Args: []string{originPath, objectPoolRefspec},
},
git.WithRefTxHook(o.Repo),
git.WithStderr(&stderr),
git.WithConfig(git.ConfigPair{
// Git is so kind to point out that we asked it to not show forced updates
// by default, so we need to ask it not to do that.
Key: "advice.fetchShowForcedUpdates", Value: "false",
}),
); err != nil {
return fmt.Errorf("fetch into object pool: %w, stderr: %q", err,
stderr.String())
}
if err := o.rescueDanglingObjects(ctx); err != nil {
return fmt.Errorf("rescuing dangling objects: %w", err)
}
if err := o.logStats(ctx, logger.WithField("when", "after fetch")); err != nil {
return fmt.Errorf("computing stats after fetch: %w", err)
}
if err := o.housekeepingManager.OptimizeRepository(ctx, o.Repo); err != nil {
return fmt.Errorf("optimizing pool repo: %w", err)
}
return nil
}
// pruneReferences prunes any references that have been deleted in the origin repository.
func (o *ObjectPool) pruneReferences(ctx context.Context, origin *localrepo.Repo) (returnedErr error) {
originPath, err := origin.Path()
if err != nil {
return fmt.Errorf("computing origin repo's path: %w", err)
}
// Ideally, we'd just use `git remote prune` directly. But unfortunately, this command does
// not support atomic updates, but will instead use a separate reference transaction for
// updating the packed-refs file and for updating each of the loose references. This can be
// really expensive in case we are about to prune a lot of references given that every time,
// the reference-transaction hook needs to vote on the deletion and reach quorum.
//
// Instead we ask for a dry-run, parse the output and queue up every reference into a
// git-update-ref(1) process. While ugly, it works around the performance issues.
prune, err := o.Repo.Exec(ctx,
git.Command{
Name: "remote",
Action: "prune",
Args: []string{"origin"},
Flags: []git.Option{
git.Flag{Name: "--dry-run"},
},
},
git.WithConfig(git.ConfigPair{Key: "remote.origin.url", Value: originPath}),
git.WithConfig(git.ConfigPair{Key: "remote.origin.fetch", Value: objectPoolRefspec}),
// This is a dry-run, only, so we don't have to enable hooks.
git.WithDisabledHooks(),
)
if err != nil {
return fmt.Errorf("spawning prune: %w", err)
}
updater, err := updateref.New(ctx, o.Repo)
if err != nil {
return fmt.Errorf("spawning updater: %w", err)
}
defer func() {
if err := updater.Close(); err != nil && returnedErr == nil {
returnedErr = fmt.Errorf("cancel updater: %w", err)
}
}()
if err := updater.Start(); err != nil {
return fmt.Errorf("start reference transaction: %w", err)
}
objectHash, err := o.ObjectHash(ctx)
if err != nil {
return fmt.Errorf("detecting object hash: %w", err)
}
// We need to manually compute a vote because all deletions we queue up here are
// force-deletions. We are forced to filter out force-deletions because these may also
// happen when evicting references from the packed-refs file.
voteHash := voting.NewVoteHash()
scanner := bufio.NewScanner(prune)
for scanner.Scan() {
line := scanner.Bytes()
// We need to skip the first two lines that represent the header of git-remote(1)'s
// output. While we should ideally just use a state machine here, it doesn't feel
// worth it given that the output is comparatively simple and given that the pruned
// branches are distinguished by a special prefix.
switch {
case bytes.Equal(line, []byte("Pruning origin")):
continue
case bytes.HasPrefix(line, []byte("URL: ")):
continue
case bytes.HasPrefix(line, []byte(" * [would prune] ")):
// The references announced by git-remote(1) only have the remote's name as
// prefix, which is "origin". We thus have to reassemble the complete name
// of every reference here.
deletedRef := "refs/remotes/" + string(bytes.TrimPrefix(line, []byte(" * [would prune] ")))
if _, err := io.Copy(voteHash, strings.NewReader(fmt.Sprintf("%[1]s %[1]s %s\n", objectHash.ZeroOID, deletedRef))); err != nil {
return fmt.Errorf("hashing reference deletion: %w", err)
}
if err := updater.Delete(git.ReferenceName(deletedRef)); err != nil {
return fmt.Errorf("queueing ref for deletion: %w", err)
}
default:
return fmt.Errorf("unexpected line: %q", line)
}
}
if err := scanner.Err(); err != nil {
return fmt.Errorf("scanning deleted refs: %w", err)
}
if err := prune.Wait(); err != nil {
return fmt.Errorf("waiting for prune: %w", err)
}
vote, err := voteHash.Vote()
if err != nil {
return fmt.Errorf("computing vote: %w", err)
}
// Prepare references so that they're locked and cannot be written by any concurrent
// processes. This also verifies that we can indeed delete the references.
if err := updater.Prepare(); err != nil {
return fmt.Errorf("preparing deletion of references: %w", err)
}
// Vote on the references we're about to delete.
if err := transaction.VoteOnContext(ctx, o.txManager, vote, voting.Prepared); err != nil {
return fmt.Errorf("preparational vote on pruned references: %w", err)
}
// Commit the pruned references to disk so that the change gets applied.
if err := updater.Commit(); err != nil {
return fmt.Errorf("deleting references: %w", err)
}
// And then confirm that we actually deleted the references.
if err := transaction.VoteOnContext(ctx, o.txManager, vote, voting.Committed); err != nil {
return fmt.Errorf("preparational vote on pruned references: %w", err)
}
return nil
}
const danglingObjectNamespace = "refs/dangling/"
// rescueDanglingObjects creates refs for all dangling objects if finds
// with `git fsck`, which converts those objects from "dangling" to
// "not-dangling". This guards against any object ever being deleted from
// a pool repository. This is a defense in depth against accidental use
// of `git prune`, which could remove Git objects that a pool member
// relies on. There is currently no way for us to reliably determine if
// an object is still used anywhere, so the only safe thing to do is to
// assume that every object _is_ used.
func (o *ObjectPool) rescueDanglingObjects(ctx context.Context) (returnedErr error) {
stderr := &bytes.Buffer{}
fsck, err := o.Repo.Exec(ctx, git.Command{
Name: "fsck",
Flags: []git.Option{git.Flag{Name: "--connectivity-only"}, git.Flag{Name: "--dangling"}},
},
git.WithStderr(stderr),
)
if err != nil {
return err
}
updater, err := updateref.New(ctx, o.Repo, updateref.WithDisabledTransactions())
if err != nil {
return err
}
defer func() {
if err := updater.Close(); err != nil && returnedErr == nil {
returnedErr = fmt.Errorf("cancel updater: %w", err)
}
}()
if err := updater.Start(); err != nil {
return fmt.Errorf("start reference transaction: %w", err)
}
objectHash, err := o.ObjectHash(ctx)
if err != nil {
return fmt.Errorf("detecting object hash: %w", err)
}
scanner := bufio.NewScanner(fsck)
for scanner.Scan() {
split := strings.SplitN(scanner.Text(), " ", 3)
if len(split) != 3 {
continue
}
if split[0] != "dangling" {
continue
}
danglingObjectID, err := objectHash.FromHex(split[2])
if err != nil {
return fmt.Errorf("parsing object ID %q: %w", split[2], err)
}
ref := git.ReferenceName(danglingObjectNamespace + split[2])
if err := updater.Create(ref, danglingObjectID); err != nil {
return err
}
}
if err := scanner.Err(); err != nil {
return err
}
if err := fsck.Wait(); err != nil {
return fmt.Errorf("git fsck: %w, stderr: %q", err, stderr.String())
}
return updater.Commit()
}
type referencedObjectTypes struct {
Blobs uint64 `json:"blobs"`
Commits uint64 `json:"commits"`
Tags uint64 `json:"tags"`
Trees uint64 `json:"trees"`
}
func (o *ObjectPool) logStats(ctx context.Context, logger *logrus.Entry) error {
fields := logrus.Fields{}
repoInfo, err := stats.RepositoryInfoForRepository(o.Repo)
if err != nil {
return fmt.Errorf("deriving repository info: %w", err)
}
fields["repository_info"] = repoInfo
forEachRef, err := o.Repo.Exec(ctx, git.Command{
Name: "for-each-ref",
Flags: []git.Option{git.Flag{Name: "--format=%(objecttype)%00%(refname)"}},
Args: []string{"refs/"},
})
if err != nil {
return fmt.Errorf("spawning for-each-ref: %w", err)
}
var danglingTypes, normalTypes referencedObjectTypes
scanner := bufio.NewScanner(forEachRef)
for scanner.Scan() {
objectType, refname, found := bytes.Cut(scanner.Bytes(), []byte{0})
if !found {
continue
}
types := &normalTypes
if bytes.HasPrefix(refname, []byte(danglingObjectNamespace)) {
types = &danglingTypes
}
switch {
case bytes.Equal(objectType, []byte("blob")):
types.Blobs++
case bytes.Equal(objectType, []byte("commit")):
types.Commits++
case bytes.Equal(objectType, []byte("tag")):
types.Tags++
case bytes.Equal(objectType, []byte("tree")):
types.Trees++
}
}
if err := scanner.Err(); err != nil {
return fmt.Errorf("scanning references: %w", err)
}
if err := forEachRef.Wait(); err != nil {
return fmt.Errorf("waiting for for-each-ref: %w", err)
}
fields["references.dangling"] = danglingTypes
fields["references.normal"] = normalTypes
logger.WithFields(fields).Info("pool dangling ref stats")
return nil
}
|