mirror of
https://codeberg.org/forgejo/forgejo.git
synced 2024-11-15 23:29:50 +01:00
5e6a008fba
This PR adds basic repository LFS management UI including the ability to find all possible pointers within the repository. Locks are not managed at present but would be addable through some simple additions. * Add basic repository lfs management * add auto-associate function * Add functionality to find commits with this lfs file * Add link to find commits on the lfs file view * Adjust commit view to state the likely branch causing the commit * Only read Oid from database
126 lines
4.3 KiB
Go
126 lines
4.3 KiB
Go
// Copyright 2019 The Gitea Authors.
|
|
// All rights reserved.
|
|
// Use of this source code is governed by a MIT-style
|
|
// license that can be found in the LICENSE file.
|
|
|
|
package pull
|
|
|
|
import (
|
|
"bufio"
|
|
"io"
|
|
"strconv"
|
|
"sync"
|
|
|
|
"code.gitea.io/gitea/models"
|
|
"code.gitea.io/gitea/modules/git/pipeline"
|
|
"code.gitea.io/gitea/modules/lfs"
|
|
"code.gitea.io/gitea/modules/log"
|
|
)
|
|
|
|
// LFSPush pushes lfs objects referred to in new commits in the head repository from the base repository
|
|
func LFSPush(tmpBasePath, mergeHeadSHA, mergeBaseSHA string, pr *models.PullRequest) error {
|
|
// Now we have to implement git lfs push
|
|
// git rev-list --objects --filter=blob:limit=1k HEAD --not base
|
|
// pass blob shas in to git cat-file --batch-check (possibly unnecessary)
|
|
// ensure only blobs and <=1k size then pass in to git cat-file --batch
|
|
// to read each sha and check each as a pointer
|
|
// Then if they are lfs -> add them to the baseRepo
|
|
revListReader, revListWriter := io.Pipe()
|
|
shasToCheckReader, shasToCheckWriter := io.Pipe()
|
|
catFileCheckReader, catFileCheckWriter := io.Pipe()
|
|
shasToBatchReader, shasToBatchWriter := io.Pipe()
|
|
catFileBatchReader, catFileBatchWriter := io.Pipe()
|
|
errChan := make(chan error, 1)
|
|
wg := sync.WaitGroup{}
|
|
wg.Add(6)
|
|
// Create the go-routines in reverse order.
|
|
|
|
// 6. Take the output of cat-file --batch and check if each file in turn
|
|
// to see if they're pointers to files in the LFS store associated with
|
|
// the head repo and add them to the base repo if so
|
|
go createLFSMetaObjectsFromCatFileBatch(catFileBatchReader, &wg, pr)
|
|
|
|
// 5. Take the shas of the blobs and batch read them
|
|
go pipeline.CatFileBatch(shasToBatchReader, catFileBatchWriter, &wg, tmpBasePath)
|
|
|
|
// 4. From the provided objects restrict to blobs <=1k
|
|
go pipeline.BlobsLessThan1024FromCatFileBatchCheck(catFileCheckReader, shasToBatchWriter, &wg)
|
|
|
|
// 3. Run batch-check on the objects retrieved from rev-list
|
|
go pipeline.CatFileBatchCheck(shasToCheckReader, catFileCheckWriter, &wg, tmpBasePath)
|
|
|
|
// 2. Check each object retrieved rejecting those without names as they will be commits or trees
|
|
go pipeline.BlobsFromRevListObjects(revListReader, shasToCheckWriter, &wg)
|
|
|
|
// 1. Run rev-list objects from mergeHead to mergeBase
|
|
go pipeline.RevListObjects(revListWriter, &wg, tmpBasePath, mergeHeadSHA, mergeBaseSHA, errChan)
|
|
|
|
wg.Wait()
|
|
select {
|
|
case err, has := <-errChan:
|
|
if has {
|
|
return err
|
|
}
|
|
default:
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func createLFSMetaObjectsFromCatFileBatch(catFileBatchReader *io.PipeReader, wg *sync.WaitGroup, pr *models.PullRequest) {
|
|
defer wg.Done()
|
|
defer catFileBatchReader.Close()
|
|
|
|
bufferedReader := bufio.NewReader(catFileBatchReader)
|
|
buf := make([]byte, 1025)
|
|
for {
|
|
// File descriptor line: sha
|
|
_, err := bufferedReader.ReadString(' ')
|
|
if err != nil {
|
|
_ = catFileBatchReader.CloseWithError(err)
|
|
break
|
|
}
|
|
// Throw away the blob
|
|
if _, err := bufferedReader.ReadString(' '); err != nil {
|
|
_ = catFileBatchReader.CloseWithError(err)
|
|
break
|
|
}
|
|
sizeStr, err := bufferedReader.ReadString('\n')
|
|
if err != nil {
|
|
_ = catFileBatchReader.CloseWithError(err)
|
|
break
|
|
}
|
|
size, err := strconv.Atoi(sizeStr[:len(sizeStr)-1])
|
|
if err != nil {
|
|
_ = catFileBatchReader.CloseWithError(err)
|
|
break
|
|
}
|
|
pointerBuf := buf[:size+1]
|
|
if _, err := io.ReadFull(bufferedReader, pointerBuf); err != nil {
|
|
_ = catFileBatchReader.CloseWithError(err)
|
|
break
|
|
}
|
|
pointerBuf = pointerBuf[:size]
|
|
// Now we need to check if the pointerBuf is an LFS pointer
|
|
pointer := lfs.IsPointerFile(&pointerBuf)
|
|
if pointer == nil {
|
|
continue
|
|
}
|
|
// Then we need to check that this pointer is in the db
|
|
if _, err := pr.HeadRepo.GetLFSMetaObjectByOid(pointer.Oid); err != nil {
|
|
if err == models.ErrLFSObjectNotExist {
|
|
log.Warn("During merge of: %d in %-v, there is a pointer to LFS Oid: %s which although present in the LFS store is not associated with the head repo %-v", pr.Index, pr.BaseRepo, pointer.Oid, pr.HeadRepo)
|
|
continue
|
|
}
|
|
_ = catFileBatchReader.CloseWithError(err)
|
|
break
|
|
}
|
|
// OK we have a pointer that is associated with the head repo
|
|
// and is actually a file in the LFS
|
|
// Therefore it should be associated with the base repo
|
|
pointer.RepositoryID = pr.BaseRepoID
|
|
if _, err := models.NewLFSMetaObject(pointer); err != nil {
|
|
_ = catFileBatchReader.CloseWithError(err)
|
|
break
|
|
}
|
|
}
|
|
}
|