mirror of
https://codeberg.org/forgejo/forgejo.git
synced 2024-11-22 08:42:32 -05:00
test: LFS gc should not delete all metadata objects
and ComputeBlobHash should depend on the blob content (not only the length)
This commit is contained in:
parent
67d6c674df
commit
3dfa5ba43a
3 changed files with 16 additions and 3 deletions
|
@ -9,7 +9,7 @@
|
||||||
|
|
||||||
-
|
-
|
||||||
|
|
||||||
id: 2
|
id: 2 # this is an LFS orphan object
|
||||||
oid: 2eccdb43825d2a49d99d542daa20075cff1d97d9d2349a8977efe9c03661737c
|
oid: 2eccdb43825d2a49d99d542daa20075cff1d97d9d2349a8977efe9c03661737c
|
||||||
size: 107
|
size: 107
|
||||||
repository_id: 54
|
repository_id: 54
|
||||||
|
|
|
@ -18,4 +18,8 @@ func TestIsValidSHAPattern(t *testing.T) {
|
||||||
assert.False(t, h.IsValid("abc"))
|
assert.False(t, h.IsValid("abc"))
|
||||||
assert.False(t, h.IsValid("123g"))
|
assert.False(t, h.IsValid("123g"))
|
||||||
assert.False(t, h.IsValid("some random text"))
|
assert.False(t, h.IsValid("some random text"))
|
||||||
|
|
||||||
|
assert.Equal(t, "79ee38a6416c1ede423ec7ee0a8639ceea4aad22", ComputeBlobHash(Sha1ObjectFormat, []byte("some random blob")).String())
|
||||||
|
assert.Equal(t, "d5c6407415d85df49592672aa421aed39b9db5e3", ComputeBlobHash(Sha1ObjectFormat, []byte("same length blob")).String())
|
||||||
|
assert.Equal(t, "df0b5174ed06ae65aea40d43316bcbc21d82c9e3158ce2661df2ad28d7931dd6", ComputeBlobHash(Sha256ObjectFormat, []byte("some random blob")).String())
|
||||||
}
|
}
|
||||||
|
|
|
@ -28,9 +28,13 @@ func TestGarbageCollectLFSMetaObjects(t *testing.T) {
|
||||||
err := storage.Init()
|
err := storage.Init()
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
|
|
||||||
repo, err := repo_model.GetRepositoryByOwnerAndName(db.DefaultContext, "user2", "repo1")
|
repo, err := repo_model.GetRepositoryByOwnerAndName(db.DefaultContext, "user2", "lfs")
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
|
|
||||||
|
validLFSObjects, err := db.GetEngine(db.DefaultContext).Count(git_model.LFSMetaObject{RepositoryID: repo.ID})
|
||||||
|
assert.NoError(t, err)
|
||||||
|
assert.Greater(t, validLFSObjects, int64(1))
|
||||||
|
|
||||||
// add lfs object
|
// add lfs object
|
||||||
lfsContent := []byte("gitea1")
|
lfsContent := []byte("gitea1")
|
||||||
lfsOid := storeObjectInRepo(t, repo.ID, &lfsContent)
|
lfsOid := storeObjectInRepo(t, repo.ID, &lfsContent)
|
||||||
|
@ -39,13 +43,18 @@ func TestGarbageCollectLFSMetaObjects(t *testing.T) {
|
||||||
err = repo_service.GarbageCollectLFSMetaObjects(context.Background(), repo_service.GarbageCollectLFSMetaObjectsOptions{
|
err = repo_service.GarbageCollectLFSMetaObjects(context.Background(), repo_service.GarbageCollectLFSMetaObjectsOptions{
|
||||||
AutoFix: true,
|
AutoFix: true,
|
||||||
OlderThan: time.Now().Add(7 * 24 * time.Hour).Add(5 * 24 * time.Hour),
|
OlderThan: time.Now().Add(7 * 24 * time.Hour).Add(5 * 24 * time.Hour),
|
||||||
UpdatedLessRecentlyThan: time.Now().Add(7 * 24 * time.Hour).Add(3 * 24 * time.Hour),
|
UpdatedLessRecentlyThan: time.Time{}, // ensure that the models/fixtures/lfs_meta_object.yml objects are considered as well
|
||||||
|
LogDetail: t.Logf,
|
||||||
})
|
})
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
|
|
||||||
// lfs meta has been deleted
|
// lfs meta has been deleted
|
||||||
_, err = git_model.GetLFSMetaObjectByOid(db.DefaultContext, repo.ID, lfsOid)
|
_, err = git_model.GetLFSMetaObjectByOid(db.DefaultContext, repo.ID, lfsOid)
|
||||||
assert.ErrorIs(t, err, git_model.ErrLFSObjectNotExist)
|
assert.ErrorIs(t, err, git_model.ErrLFSObjectNotExist)
|
||||||
|
|
||||||
|
remainingLFSObjects, err := db.GetEngine(db.DefaultContext).Count(git_model.LFSMetaObject{RepositoryID: repo.ID})
|
||||||
|
assert.NoError(t, err)
|
||||||
|
assert.Equal(t, validLFSObjects-1, remainingLFSObjects)
|
||||||
}
|
}
|
||||||
|
|
||||||
func storeObjectInRepo(t *testing.T, repositoryID int64, content *[]byte) string {
|
func storeObjectInRepo(t *testing.T, repositoryID int64, content *[]byte) string {
|
||||||
|
|
Loading…
Reference in a new issue