mirror of
https://codeberg.org/forgejo/forgejo.git
synced 2024-11-25 08:59:31 -05:00
Refactor indexer (#25174)
Refactor `modules/indexer` to make it more maintainable. And it can be easier to support more features. I'm trying to solve some of issue searching, this is a precursor to making functional changes. Current supported engines and the index versions: | engines | issues | code | | - | - | - | | db | Just a wrapper for database queries, doesn't need version | - | | bleve | The version of index is **2** | The version of index is **6** | | elasticsearch | The old index has no version, will be treated as version **0** in this PR | The version of index is **1** | | meilisearch | The old index has no version, will be treated as version **0** in this PR | - | ## Changes ### Split Splited it into mutiple packages ```text indexer ├── internal │ ├── bleve │ ├── db │ ├── elasticsearch │ └── meilisearch ├── code │ ├── bleve │ ├── elasticsearch │ └── internal └── issues ├── bleve ├── db ├── elasticsearch ├── internal └── meilisearch ``` - `indexer/interanal`: Internal shared package for indexer. - `indexer/interanal/[engine]`: Internal shared package for each engine (bleve/db/elasticsearch/meilisearch). - `indexer/code`: Implementations for code indexer. - `indexer/code/internal`: Internal shared package for code indexer. - `indexer/code/[engine]`: Implementation via each engine for code indexer. - `indexer/issues`: Implementations for issues indexer. ### Deduplication - Combine `Init/Ping/Close` for code indexer and issues indexer. - ~Combine `issues.indexerHolder` and `code.wrappedIndexer` to `internal.IndexHolder`.~ Remove it, use dummy indexer instead when the indexer is not ready. - Duplicate two copies of creating ES clients. - Duplicate two copies of `indexerID()`. ### Enhancement - [x] Support index version for elasticsearch issues indexer, the old index without version will be treated as version 0. - [x] Fix spell of `elastic_search/ElasticSearch`, it should be `Elasticsearch`. - [x] Improve versioning of ES index. We don't need `Aliases`: - Gitea does't need aliases for "Zero Downtime" because it never delete old indexes. - The old code of issues indexer uses the orignal name to create issue index, so it's tricky to convert it to an alias. - [x] Support index version for meilisearch issues indexer, the old index without version will be treated as version 0. - [x] Do "ping" only when `Ping` has been called, don't ping periodically and cache the status. - [x] Support the context parameter whenever possible. - [x] Fix outdated example config. - [x] Give up the requeue logic of issues indexer: When indexing fails, call Ping to check if it was caused by the engine being unavailable, and only requeue the task if the engine is unavailable. - It is fragile and tricky, could cause data losing (It did happen when I was doing some tests for this PR). And it works for ES only. - Just always requeue the failed task, if it caused by bad data, it's a bug of Gitea which should be fixed. --------- Co-authored-by: Giteabot <teabot@gitea.io>
This commit is contained in:
parent
b0215c40cd
commit
375fd15fbf
43 changed files with 1374 additions and 1426 deletions
|
@ -1334,10 +1334,10 @@ LEVEL = Info
|
||||||
;; Issue indexer storage path, available when ISSUE_INDEXER_TYPE is bleve
|
;; Issue indexer storage path, available when ISSUE_INDEXER_TYPE is bleve
|
||||||
;ISSUE_INDEXER_PATH = indexers/issues.bleve ; Relative paths will be made absolute against _`AppWorkPath`_.
|
;ISSUE_INDEXER_PATH = indexers/issues.bleve ; Relative paths will be made absolute against _`AppWorkPath`_.
|
||||||
;;
|
;;
|
||||||
;; Issue indexer connection string, available when ISSUE_INDEXER_TYPE is elasticsearch or meilisearch
|
;; Issue indexer connection string, available when ISSUE_INDEXER_TYPE is elasticsearch (e.g. http://elastic:password@localhost:9200) or meilisearch (e.g. http://:apikey@localhost:7700)
|
||||||
;ISSUE_INDEXER_CONN_STR = http://elastic:changeme@localhost:9200
|
;ISSUE_INDEXER_CONN_STR =
|
||||||
;;
|
;;
|
||||||
;; Issue indexer name, available when ISSUE_INDEXER_TYPE is elasticsearch
|
;; Issue indexer name, available when ISSUE_INDEXER_TYPE is elasticsearch or meilisearch.
|
||||||
;ISSUE_INDEXER_NAME = gitea_issues
|
;ISSUE_INDEXER_NAME = gitea_issues
|
||||||
;;
|
;;
|
||||||
;; Timeout the indexer if it takes longer than this to start.
|
;; Timeout the indexer if it takes longer than this to start.
|
||||||
|
|
|
@ -458,15 +458,15 @@ relation to port exhaustion.
|
||||||
## Indexer (`indexer`)
|
## Indexer (`indexer`)
|
||||||
|
|
||||||
- `ISSUE_INDEXER_TYPE`: **bleve**: Issue indexer type, currently supported: `bleve`, `db`, `elasticsearch` or `meilisearch`.
|
- `ISSUE_INDEXER_TYPE`: **bleve**: Issue indexer type, currently supported: `bleve`, `db`, `elasticsearch` or `meilisearch`.
|
||||||
- `ISSUE_INDEXER_CONN_STR`: ****: Issue indexer connection string, available when ISSUE_INDEXER_TYPE is elasticsearch, or meilisearch. i.e. http://elastic:changeme@localhost:9200
|
- `ISSUE_INDEXER_CONN_STR`: ****: Issue indexer connection string, available when ISSUE_INDEXER_TYPE is elasticsearch (e.g. http://elastic:password@localhost:9200) or meilisearch (e.g. http://:apikey@localhost:7700)
|
||||||
- `ISSUE_INDEXER_NAME`: **gitea_issues**: Issue indexer name, available when ISSUE_INDEXER_TYPE is elasticsearch
|
- `ISSUE_INDEXER_NAME`: **gitea_issues**: Issue indexer name, available when ISSUE_INDEXER_TYPE is elasticsearch or meilisearch.
|
||||||
- `ISSUE_INDEXER_PATH`: **indexers/issues.bleve**: Index file used for issue search; available when ISSUE_INDEXER_TYPE is bleve and elasticsearch. Relative paths will be made absolute against _`AppWorkPath`_.
|
- `ISSUE_INDEXER_PATH`: **indexers/issues.bleve**: Index file used for issue search; available when ISSUE_INDEXER_TYPE is bleve and elasticsearch. Relative paths will be made absolute against _`AppWorkPath`_.
|
||||||
|
|
||||||
- `REPO_INDEXER_ENABLED`: **false**: Enables code search (uses a lot of disk space, about 6 times more than the repository size).
|
- `REPO_INDEXER_ENABLED`: **false**: Enables code search (uses a lot of disk space, about 6 times more than the repository size).
|
||||||
- `REPO_INDEXER_REPO_TYPES`: **sources,forks,mirrors,templates**: Repo indexer units. The items to index could be `sources`, `forks`, `mirrors`, `templates` or any combination of them separated by a comma. If empty then it defaults to `sources` only, as if you'd like to disable fully please see `REPO_INDEXER_ENABLED`.
|
- `REPO_INDEXER_REPO_TYPES`: **sources,forks,mirrors,templates**: Repo indexer units. The items to index could be `sources`, `forks`, `mirrors`, `templates` or any combination of them separated by a comma. If empty then it defaults to `sources` only, as if you'd like to disable fully please see `REPO_INDEXER_ENABLED`.
|
||||||
- `REPO_INDEXER_TYPE`: **bleve**: Code search engine type, could be `bleve` or `elasticsearch`.
|
- `REPO_INDEXER_TYPE`: **bleve**: Code search engine type, could be `bleve` or `elasticsearch`.
|
||||||
- `REPO_INDEXER_PATH`: **indexers/repos.bleve**: Index file used for code search.
|
- `REPO_INDEXER_PATH`: **indexers/repos.bleve**: Index file used for code search.
|
||||||
- `REPO_INDEXER_CONN_STR`: ****: Code indexer connection string, available when `REPO_INDEXER_TYPE` is elasticsearch. i.e. http://elastic:changeme@localhost:9200
|
- `REPO_INDEXER_CONN_STR`: ****: Code indexer connection string, available when `REPO_INDEXER_TYPE` is elasticsearch. i.e. http://elastic:password@localhost:9200
|
||||||
- `REPO_INDEXER_NAME`: **gitea_codes**: Code indexer name, available when `REPO_INDEXER_TYPE` is elasticsearch
|
- `REPO_INDEXER_NAME`: **gitea_codes**: Code indexer name, available when `REPO_INDEXER_TYPE` is elasticsearch
|
||||||
|
|
||||||
- `REPO_INDEXER_INCLUDE`: **empty**: A comma separated list of glob patterns (see https://github.com/gobwas/glob) to **include** in the index. Use `**.txt` to match any files with .txt extension. An empty list means include all files.
|
- `REPO_INDEXER_INCLUDE`: **empty**: A comma separated list of glob patterns (see https://github.com/gobwas/glob) to **include** in the index. Use `**.txt` to match any files with .txt extension. An empty list means include all files.
|
||||||
|
|
|
@ -593,7 +593,7 @@ func RepoAssignment(ctx *Context) (cancel context.CancelFunc) {
|
||||||
|
|
||||||
ctx.Data["RepoSearchEnabled"] = setting.Indexer.RepoIndexerEnabled
|
ctx.Data["RepoSearchEnabled"] = setting.Indexer.RepoIndexerEnabled
|
||||||
if setting.Indexer.RepoIndexerEnabled {
|
if setting.Indexer.RepoIndexerEnabled {
|
||||||
ctx.Data["CodeIndexerUnavailable"] = !code_indexer.IsAvailable()
|
ctx.Data["CodeIndexerUnavailable"] = !code_indexer.IsAvailable(ctx)
|
||||||
}
|
}
|
||||||
|
|
||||||
if ctx.IsSigned {
|
if ctx.IsSigned {
|
||||||
|
|
|
@ -1,14 +1,13 @@
|
||||||
// Copyright 2019 The Gitea Authors. All rights reserved.
|
// Copyright 2019 The Gitea Authors. All rights reserved.
|
||||||
// SPDX-License-Identifier: MIT
|
// SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
package code
|
package bleve
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"bufio"
|
"bufio"
|
||||||
"context"
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
"io"
|
"io"
|
||||||
"os"
|
|
||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
"time"
|
"time"
|
||||||
|
@ -17,12 +16,13 @@ import (
|
||||||
"code.gitea.io/gitea/modules/analyze"
|
"code.gitea.io/gitea/modules/analyze"
|
||||||
"code.gitea.io/gitea/modules/charset"
|
"code.gitea.io/gitea/modules/charset"
|
||||||
"code.gitea.io/gitea/modules/git"
|
"code.gitea.io/gitea/modules/git"
|
||||||
gitea_bleve "code.gitea.io/gitea/modules/indexer/bleve"
|
"code.gitea.io/gitea/modules/indexer/code/internal"
|
||||||
|
indexer_internal "code.gitea.io/gitea/modules/indexer/internal"
|
||||||
|
inner_bleve "code.gitea.io/gitea/modules/indexer/internal/bleve"
|
||||||
"code.gitea.io/gitea/modules/log"
|
"code.gitea.io/gitea/modules/log"
|
||||||
"code.gitea.io/gitea/modules/setting"
|
"code.gitea.io/gitea/modules/setting"
|
||||||
"code.gitea.io/gitea/modules/timeutil"
|
"code.gitea.io/gitea/modules/timeutil"
|
||||||
"code.gitea.io/gitea/modules/typesniffer"
|
"code.gitea.io/gitea/modules/typesniffer"
|
||||||
"code.gitea.io/gitea/modules/util"
|
|
||||||
|
|
||||||
"github.com/blevesearch/bleve/v2"
|
"github.com/blevesearch/bleve/v2"
|
||||||
analyzer_custom "github.com/blevesearch/bleve/v2/analysis/analyzer/custom"
|
analyzer_custom "github.com/blevesearch/bleve/v2/analysis/analyzer/custom"
|
||||||
|
@ -31,10 +31,8 @@ import (
|
||||||
"github.com/blevesearch/bleve/v2/analysis/token/lowercase"
|
"github.com/blevesearch/bleve/v2/analysis/token/lowercase"
|
||||||
"github.com/blevesearch/bleve/v2/analysis/token/unicodenorm"
|
"github.com/blevesearch/bleve/v2/analysis/token/unicodenorm"
|
||||||
"github.com/blevesearch/bleve/v2/analysis/tokenizer/unicode"
|
"github.com/blevesearch/bleve/v2/analysis/tokenizer/unicode"
|
||||||
"github.com/blevesearch/bleve/v2/index/upsidedown"
|
|
||||||
"github.com/blevesearch/bleve/v2/mapping"
|
"github.com/blevesearch/bleve/v2/mapping"
|
||||||
"github.com/blevesearch/bleve/v2/search/query"
|
"github.com/blevesearch/bleve/v2/search/query"
|
||||||
"github.com/ethantkoenig/rupture"
|
|
||||||
"github.com/go-enry/go-enry/v2"
|
"github.com/go-enry/go-enry/v2"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -59,38 +57,6 @@ func addUnicodeNormalizeTokenFilter(m *mapping.IndexMappingImpl) error {
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
// openBleveIndexer open the index at the specified path, checking for metadata
|
|
||||||
// updates and bleve version updates. If index needs to be created (or
|
|
||||||
// re-created), returns (nil, nil)
|
|
||||||
func openBleveIndexer(path string, latestVersion int) (bleve.Index, error) {
|
|
||||||
_, err := os.Stat(path)
|
|
||||||
if err != nil && os.IsNotExist(err) {
|
|
||||||
return nil, nil
|
|
||||||
} else if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
|
|
||||||
metadata, err := rupture.ReadIndexMetadata(path)
|
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
if metadata.Version < latestVersion {
|
|
||||||
// the indexer is using a previous version, so we should delete it and
|
|
||||||
// re-populate
|
|
||||||
return nil, util.RemoveAll(path)
|
|
||||||
}
|
|
||||||
|
|
||||||
index, err := bleve.Open(path)
|
|
||||||
if err != nil && err == upsidedown.IncompatibleVersion {
|
|
||||||
// the indexer was built with a previous version of bleve, so we should
|
|
||||||
// delete it and re-populate
|
|
||||||
return nil, util.RemoveAll(path)
|
|
||||||
} else if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
return index, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// RepoIndexerData data stored in the repo indexer
|
// RepoIndexerData data stored in the repo indexer
|
||||||
type RepoIndexerData struct {
|
type RepoIndexerData struct {
|
||||||
RepoID int64
|
RepoID int64
|
||||||
|
@ -111,8 +77,8 @@ const (
|
||||||
repoIndexerLatestVersion = 6
|
repoIndexerLatestVersion = 6
|
||||||
)
|
)
|
||||||
|
|
||||||
// createBleveIndexer create a bleve repo indexer if one does not already exist
|
// generateBleveIndexMapping generates a bleve index mapping for the repo indexer
|
||||||
func createBleveIndexer(path string, latestVersion int) (bleve.Index, error) {
|
func generateBleveIndexMapping() (mapping.IndexMapping, error) {
|
||||||
docMapping := bleve.NewDocumentMapping()
|
docMapping := bleve.NewDocumentMapping()
|
||||||
numericFieldMapping := bleve.NewNumericFieldMapping()
|
numericFieldMapping := bleve.NewNumericFieldMapping()
|
||||||
numericFieldMapping.IncludeInAll = false
|
numericFieldMapping.IncludeInAll = false
|
||||||
|
@ -147,42 +113,28 @@ func createBleveIndexer(path string, latestVersion int) (bleve.Index, error) {
|
||||||
mapping.AddDocumentMapping(repoIndexerDocType, docMapping)
|
mapping.AddDocumentMapping(repoIndexerDocType, docMapping)
|
||||||
mapping.AddDocumentMapping("_all", bleve.NewDocumentDisabledMapping())
|
mapping.AddDocumentMapping("_all", bleve.NewDocumentDisabledMapping())
|
||||||
|
|
||||||
indexer, err := bleve.New(path, mapping)
|
return mapping, nil
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
|
|
||||||
if err = rupture.WriteIndexMetadata(path, &rupture.IndexMetadata{
|
|
||||||
Version: latestVersion,
|
|
||||||
}); err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
return indexer, nil
|
|
||||||
}
|
}
|
||||||
|
|
||||||
var _ Indexer = &BleveIndexer{}
|
var _ internal.Indexer = &Indexer{}
|
||||||
|
|
||||||
// BleveIndexer represents a bleve indexer implementation
|
// Indexer represents a bleve indexer implementation
|
||||||
type BleveIndexer struct {
|
type Indexer struct {
|
||||||
indexDir string
|
inner *inner_bleve.Indexer
|
||||||
indexer bleve.Index
|
indexer_internal.Indexer // do not composite inner_bleve.Indexer directly to avoid exposing too much
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewBleveIndexer creates a new bleve local indexer
|
// NewIndexer creates a new bleve local indexer
|
||||||
func NewBleveIndexer(indexDir string) (*BleveIndexer, bool, error) {
|
func NewIndexer(indexDir string) *Indexer {
|
||||||
indexer := &BleveIndexer{
|
inner := inner_bleve.NewIndexer(indexDir, repoIndexerLatestVersion, generateBleveIndexMapping)
|
||||||
indexDir: indexDir,
|
return &Indexer{
|
||||||
|
Indexer: inner,
|
||||||
|
inner: inner,
|
||||||
}
|
}
|
||||||
created, err := indexer.init()
|
|
||||||
if err != nil {
|
|
||||||
indexer.Close()
|
|
||||||
return nil, false, err
|
|
||||||
}
|
|
||||||
return indexer, created, err
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (b *BleveIndexer) addUpdate(ctx context.Context, batchWriter git.WriteCloserError, batchReader *bufio.Reader, commitSha string,
|
func (b *Indexer) addUpdate(ctx context.Context, batchWriter git.WriteCloserError, batchReader *bufio.Reader, commitSha string,
|
||||||
update fileUpdate, repo *repo_model.Repository, batch *gitea_bleve.FlushingBatch,
|
update internal.FileUpdate, repo *repo_model.Repository, batch *inner_bleve.FlushingBatch,
|
||||||
) error {
|
) error {
|
||||||
// Ignore vendored files in code search
|
// Ignore vendored files in code search
|
||||||
if setting.Indexer.ExcludeVendored && analyze.IsVendor(update.Filename) {
|
if setting.Indexer.ExcludeVendored && analyze.IsVendor(update.Filename) {
|
||||||
|
@ -227,7 +179,7 @@ func (b *BleveIndexer) addUpdate(ctx context.Context, batchWriter git.WriteClose
|
||||||
if _, err = batchReader.Discard(1); err != nil {
|
if _, err = batchReader.Discard(1); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
id := filenameIndexerID(repo.ID, update.Filename)
|
id := internal.FilenameIndexerID(repo.ID, update.Filename)
|
||||||
return batch.Index(id, &RepoIndexerData{
|
return batch.Index(id, &RepoIndexerData{
|
||||||
RepoID: repo.ID,
|
RepoID: repo.ID,
|
||||||
CommitID: commitSha,
|
CommitID: commitSha,
|
||||||
|
@ -237,50 +189,14 @@ func (b *BleveIndexer) addUpdate(ctx context.Context, batchWriter git.WriteClose
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
func (b *BleveIndexer) addDelete(filename string, repo *repo_model.Repository, batch *gitea_bleve.FlushingBatch) error {
|
func (b *Indexer) addDelete(filename string, repo *repo_model.Repository, batch *inner_bleve.FlushingBatch) error {
|
||||||
id := filenameIndexerID(repo.ID, filename)
|
id := internal.FilenameIndexerID(repo.ID, filename)
|
||||||
return batch.Delete(id)
|
return batch.Delete(id)
|
||||||
}
|
}
|
||||||
|
|
||||||
// init init the indexer
|
|
||||||
func (b *BleveIndexer) init() (bool, error) {
|
|
||||||
var err error
|
|
||||||
b.indexer, err = openBleveIndexer(b.indexDir, repoIndexerLatestVersion)
|
|
||||||
if err != nil {
|
|
||||||
return false, err
|
|
||||||
}
|
|
||||||
if b.indexer != nil {
|
|
||||||
return false, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
b.indexer, err = createBleveIndexer(b.indexDir, repoIndexerLatestVersion)
|
|
||||||
if err != nil {
|
|
||||||
return false, err
|
|
||||||
}
|
|
||||||
|
|
||||||
return true, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Close close the indexer
|
|
||||||
func (b *BleveIndexer) Close() {
|
|
||||||
log.Debug("Closing repo indexer")
|
|
||||||
if b.indexer != nil {
|
|
||||||
err := b.indexer.Close()
|
|
||||||
if err != nil {
|
|
||||||
log.Error("Error whilst closing the repository indexer: %v", err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
log.Info("PID: %d Repository Indexer closed", os.Getpid())
|
|
||||||
}
|
|
||||||
|
|
||||||
// Ping does nothing
|
|
||||||
func (b *BleveIndexer) Ping() bool {
|
|
||||||
return true
|
|
||||||
}
|
|
||||||
|
|
||||||
// Index indexes the data
|
// Index indexes the data
|
||||||
func (b *BleveIndexer) Index(ctx context.Context, repo *repo_model.Repository, sha string, changes *repoChanges) error {
|
func (b *Indexer) Index(ctx context.Context, repo *repo_model.Repository, sha string, changes *internal.RepoChanges) error {
|
||||||
batch := gitea_bleve.NewFlushingBatch(b.indexer, maxBatchSize)
|
batch := inner_bleve.NewFlushingBatch(b.inner.Indexer, maxBatchSize)
|
||||||
if len(changes.Updates) > 0 {
|
if len(changes.Updates) > 0 {
|
||||||
|
|
||||||
// Now because of some insanity with git cat-file not immediately failing if not run in a valid git directory we need to run git rev-parse first!
|
// Now because of some insanity with git cat-file not immediately failing if not run in a valid git directory we need to run git rev-parse first!
|
||||||
|
@ -308,14 +224,14 @@ func (b *BleveIndexer) Index(ctx context.Context, repo *repo_model.Repository, s
|
||||||
}
|
}
|
||||||
|
|
||||||
// Delete deletes indexes by ids
|
// Delete deletes indexes by ids
|
||||||
func (b *BleveIndexer) Delete(repoID int64) error {
|
func (b *Indexer) Delete(_ context.Context, repoID int64) error {
|
||||||
query := numericEqualityQuery(repoID, "RepoID")
|
query := numericEqualityQuery(repoID, "RepoID")
|
||||||
searchRequest := bleve.NewSearchRequestOptions(query, 2147483647, 0, false)
|
searchRequest := bleve.NewSearchRequestOptions(query, 2147483647, 0, false)
|
||||||
result, err := b.indexer.Search(searchRequest)
|
result, err := b.inner.Indexer.Search(searchRequest)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
batch := gitea_bleve.NewFlushingBatch(b.indexer, maxBatchSize)
|
batch := inner_bleve.NewFlushingBatch(b.inner.Indexer, maxBatchSize)
|
||||||
for _, hit := range result.Hits {
|
for _, hit := range result.Hits {
|
||||||
if err = batch.Delete(hit.ID); err != nil {
|
if err = batch.Delete(hit.ID); err != nil {
|
||||||
return err
|
return err
|
||||||
|
@ -326,7 +242,7 @@ func (b *BleveIndexer) Delete(repoID int64) error {
|
||||||
|
|
||||||
// Search searches for files in the specified repo.
|
// Search searches for files in the specified repo.
|
||||||
// Returns the matching file-paths
|
// Returns the matching file-paths
|
||||||
func (b *BleveIndexer) Search(ctx context.Context, repoIDs []int64, language, keyword string, page, pageSize int, isMatch bool) (int64, []*SearchResult, []*SearchResultLanguages, error) {
|
func (b *Indexer) Search(ctx context.Context, repoIDs []int64, language, keyword string, page, pageSize int, isMatch bool) (int64, []*internal.SearchResult, []*internal.SearchResultLanguages, error) {
|
||||||
var (
|
var (
|
||||||
indexerQuery query.Query
|
indexerQuery query.Query
|
||||||
keywordQuery query.Query
|
keywordQuery query.Query
|
||||||
|
@ -379,14 +295,14 @@ func (b *BleveIndexer) Search(ctx context.Context, repoIDs []int64, language, ke
|
||||||
searchRequest.AddFacet("languages", bleve.NewFacetRequest("Language", 10))
|
searchRequest.AddFacet("languages", bleve.NewFacetRequest("Language", 10))
|
||||||
}
|
}
|
||||||
|
|
||||||
result, err := b.indexer.SearchInContext(ctx, searchRequest)
|
result, err := b.inner.Indexer.SearchInContext(ctx, searchRequest)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return 0, nil, nil, err
|
return 0, nil, nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
total := int64(result.Total)
|
total := int64(result.Total)
|
||||||
|
|
||||||
searchResults := make([]*SearchResult, len(result.Hits))
|
searchResults := make([]*internal.SearchResult, len(result.Hits))
|
||||||
for i, hit := range result.Hits {
|
for i, hit := range result.Hits {
|
||||||
startIndex, endIndex := -1, -1
|
startIndex, endIndex := -1, -1
|
||||||
for _, locations := range hit.Locations["Content"] {
|
for _, locations := range hit.Locations["Content"] {
|
||||||
|
@ -405,11 +321,11 @@ func (b *BleveIndexer) Search(ctx context.Context, repoIDs []int64, language, ke
|
||||||
if t, err := time.Parse(time.RFC3339, hit.Fields["UpdatedAt"].(string)); err == nil {
|
if t, err := time.Parse(time.RFC3339, hit.Fields["UpdatedAt"].(string)); err == nil {
|
||||||
updatedUnix = timeutil.TimeStamp(t.Unix())
|
updatedUnix = timeutil.TimeStamp(t.Unix())
|
||||||
}
|
}
|
||||||
searchResults[i] = &SearchResult{
|
searchResults[i] = &internal.SearchResult{
|
||||||
RepoID: int64(hit.Fields["RepoID"].(float64)),
|
RepoID: int64(hit.Fields["RepoID"].(float64)),
|
||||||
StartIndex: startIndex,
|
StartIndex: startIndex,
|
||||||
EndIndex: endIndex,
|
EndIndex: endIndex,
|
||||||
Filename: filenameOfIndexerID(hit.ID),
|
Filename: internal.FilenameOfIndexerID(hit.ID),
|
||||||
Content: hit.Fields["Content"].(string),
|
Content: hit.Fields["Content"].(string),
|
||||||
CommitID: hit.Fields["CommitID"].(string),
|
CommitID: hit.Fields["CommitID"].(string),
|
||||||
UpdatedUnix: updatedUnix,
|
UpdatedUnix: updatedUnix,
|
||||||
|
@ -418,7 +334,7 @@ func (b *BleveIndexer) Search(ctx context.Context, repoIDs []int64, language, ke
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
searchResultLanguages := make([]*SearchResultLanguages, 0, 10)
|
searchResultLanguages := make([]*internal.SearchResultLanguages, 0, 10)
|
||||||
if len(language) > 0 {
|
if len(language) > 0 {
|
||||||
// Use separate query to go get all language counts
|
// Use separate query to go get all language counts
|
||||||
facetRequest := bleve.NewSearchRequestOptions(facetQuery, 1, 0, false)
|
facetRequest := bleve.NewSearchRequestOptions(facetQuery, 1, 0, false)
|
||||||
|
@ -426,7 +342,7 @@ func (b *BleveIndexer) Search(ctx context.Context, repoIDs []int64, language, ke
|
||||||
facetRequest.IncludeLocations = true
|
facetRequest.IncludeLocations = true
|
||||||
facetRequest.AddFacet("languages", bleve.NewFacetRequest("Language", 10))
|
facetRequest.AddFacet("languages", bleve.NewFacetRequest("Language", 10))
|
||||||
|
|
||||||
if result, err = b.indexer.Search(facetRequest); err != nil {
|
if result, err = b.inner.Indexer.Search(facetRequest); err != nil {
|
||||||
return 0, nil, nil, err
|
return 0, nil, nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -436,7 +352,7 @@ func (b *BleveIndexer) Search(ctx context.Context, repoIDs []int64, language, ke
|
||||||
if len(term.Term) == 0 {
|
if len(term.Term) == 0 {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
searchResultLanguages = append(searchResultLanguages, &SearchResultLanguages{
|
searchResultLanguages = append(searchResultLanguages, &internal.SearchResultLanguages{
|
||||||
Language: term.Term,
|
Language: term.Term,
|
||||||
Color: enry.GetColor(term.Term),
|
Color: enry.GetColor(term.Term),
|
||||||
Count: term.Count,
|
Count: term.Count,
|
|
@ -1,30 +0,0 @@
|
||||||
// Copyright 2019 The Gitea Authors. All rights reserved.
|
|
||||||
// SPDX-License-Identifier: MIT
|
|
||||||
|
|
||||||
package code
|
|
||||||
|
|
||||||
import (
|
|
||||||
"testing"
|
|
||||||
|
|
||||||
"code.gitea.io/gitea/models/unittest"
|
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
|
||||||
)
|
|
||||||
|
|
||||||
func TestBleveIndexAndSearch(t *testing.T) {
|
|
||||||
unittest.PrepareTestEnv(t)
|
|
||||||
|
|
||||||
dir := t.TempDir()
|
|
||||||
|
|
||||||
idx, _, err := NewBleveIndexer(dir)
|
|
||||||
if err != nil {
|
|
||||||
assert.Fail(t, "Unable to create bleve indexer Error: %v", err)
|
|
||||||
if idx != nil {
|
|
||||||
idx.Close()
|
|
||||||
}
|
|
||||||
return
|
|
||||||
}
|
|
||||||
defer idx.Close()
|
|
||||||
|
|
||||||
testIndexer("beleve", t, idx)
|
|
||||||
}
|
|
|
@ -1,41 +0,0 @@
|
||||||
// Copyright 2020 The Gitea Authors. All rights reserved.
|
|
||||||
// SPDX-License-Identifier: MIT
|
|
||||||
|
|
||||||
package code
|
|
||||||
|
|
||||||
import (
|
|
||||||
"os"
|
|
||||||
"testing"
|
|
||||||
|
|
||||||
"code.gitea.io/gitea/models/unittest"
|
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
|
||||||
)
|
|
||||||
|
|
||||||
func TestESIndexAndSearch(t *testing.T) {
|
|
||||||
unittest.PrepareTestEnv(t)
|
|
||||||
|
|
||||||
u := os.Getenv("TEST_INDEXER_CODE_ES_URL")
|
|
||||||
if u == "" {
|
|
||||||
t.SkipNow()
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
indexer, _, err := NewElasticSearchIndexer(u, "gitea_codes")
|
|
||||||
if err != nil {
|
|
||||||
assert.Fail(t, "Unable to create ES indexer Error: %v", err)
|
|
||||||
if indexer != nil {
|
|
||||||
indexer.Close()
|
|
||||||
}
|
|
||||||
return
|
|
||||||
}
|
|
||||||
defer indexer.Close()
|
|
||||||
|
|
||||||
testIndexer("elastic_search", t, indexer)
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestIndexPos(t *testing.T) {
|
|
||||||
startIdx, endIdx := indexPos("test index start and end", "start", "end")
|
|
||||||
assert.EqualValues(t, 11, startIdx)
|
|
||||||
assert.EqualValues(t, 24, endIdx)
|
|
||||||
}
|
|
|
@ -1,25 +1,23 @@
|
||||||
// Copyright 2020 The Gitea Authors. All rights reserved.
|
// Copyright 2020 The Gitea Authors. All rights reserved.
|
||||||
// SPDX-License-Identifier: MIT
|
// SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
package code
|
package elasticsearch
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"bufio"
|
"bufio"
|
||||||
"context"
|
"context"
|
||||||
"errors"
|
|
||||||
"fmt"
|
"fmt"
|
||||||
"io"
|
"io"
|
||||||
"net"
|
|
||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
"sync"
|
|
||||||
"time"
|
|
||||||
|
|
||||||
repo_model "code.gitea.io/gitea/models/repo"
|
repo_model "code.gitea.io/gitea/models/repo"
|
||||||
"code.gitea.io/gitea/modules/analyze"
|
"code.gitea.io/gitea/modules/analyze"
|
||||||
"code.gitea.io/gitea/modules/charset"
|
"code.gitea.io/gitea/modules/charset"
|
||||||
"code.gitea.io/gitea/modules/git"
|
"code.gitea.io/gitea/modules/git"
|
||||||
"code.gitea.io/gitea/modules/graceful"
|
"code.gitea.io/gitea/modules/indexer/code/internal"
|
||||||
|
indexer_internal "code.gitea.io/gitea/modules/indexer/internal"
|
||||||
|
inner_elasticsearch "code.gitea.io/gitea/modules/indexer/internal/elasticsearch"
|
||||||
"code.gitea.io/gitea/modules/json"
|
"code.gitea.io/gitea/modules/json"
|
||||||
"code.gitea.io/gitea/modules/log"
|
"code.gitea.io/gitea/modules/log"
|
||||||
"code.gitea.io/gitea/modules/setting"
|
"code.gitea.io/gitea/modules/setting"
|
||||||
|
@ -38,63 +36,22 @@ const (
|
||||||
esMultiMatchTypePhrasePrefix = "phrase_prefix"
|
esMultiMatchTypePhrasePrefix = "phrase_prefix"
|
||||||
)
|
)
|
||||||
|
|
||||||
var _ Indexer = &ElasticSearchIndexer{}
|
var _ internal.Indexer = &Indexer{}
|
||||||
|
|
||||||
// ElasticSearchIndexer implements Indexer interface
|
// Indexer implements Indexer interface
|
||||||
type ElasticSearchIndexer struct {
|
type Indexer struct {
|
||||||
client *elastic.Client
|
inner *inner_elasticsearch.Indexer
|
||||||
indexerAliasName string
|
indexer_internal.Indexer // do not composite inner_elasticsearch.Indexer directly to avoid exposing too much
|
||||||
available bool
|
|
||||||
stopTimer chan struct{}
|
|
||||||
lock sync.RWMutex
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewElasticSearchIndexer creates a new elasticsearch indexer
|
// NewIndexer creates a new elasticsearch indexer
|
||||||
func NewElasticSearchIndexer(url, indexerName string) (*ElasticSearchIndexer, bool, error) {
|
func NewIndexer(url, indexerName string) *Indexer {
|
||||||
opts := []elastic.ClientOptionFunc{
|
inner := inner_elasticsearch.NewIndexer(url, indexerName, esRepoIndexerLatestVersion, defaultMapping)
|
||||||
elastic.SetURL(url),
|
indexer := &Indexer{
|
||||||
elastic.SetSniff(false),
|
inner: inner,
|
||||||
elastic.SetHealthcheckInterval(10 * time.Second),
|
Indexer: inner,
|
||||||
elastic.SetGzip(false),
|
|
||||||
}
|
}
|
||||||
|
return indexer
|
||||||
logger := log.GetLogger(log.DEFAULT)
|
|
||||||
|
|
||||||
opts = append(opts, elastic.SetTraceLog(&log.PrintfLogger{Logf: logger.Trace}))
|
|
||||||
opts = append(opts, elastic.SetInfoLog(&log.PrintfLogger{Logf: logger.Info}))
|
|
||||||
opts = append(opts, elastic.SetErrorLog(&log.PrintfLogger{Logf: logger.Error}))
|
|
||||||
|
|
||||||
client, err := elastic.NewClient(opts...)
|
|
||||||
if err != nil {
|
|
||||||
return nil, false, err
|
|
||||||
}
|
|
||||||
|
|
||||||
indexer := &ElasticSearchIndexer{
|
|
||||||
client: client,
|
|
||||||
indexerAliasName: indexerName,
|
|
||||||
available: true,
|
|
||||||
stopTimer: make(chan struct{}),
|
|
||||||
}
|
|
||||||
|
|
||||||
ticker := time.NewTicker(10 * time.Second)
|
|
||||||
go func() {
|
|
||||||
for {
|
|
||||||
select {
|
|
||||||
case <-ticker.C:
|
|
||||||
indexer.checkAvailability()
|
|
||||||
case <-indexer.stopTimer:
|
|
||||||
ticker.Stop()
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}()
|
|
||||||
|
|
||||||
exists, err := indexer.init()
|
|
||||||
if err != nil {
|
|
||||||
indexer.Close()
|
|
||||||
return nil, false, err
|
|
||||||
}
|
|
||||||
return indexer, !exists, err
|
|
||||||
}
|
}
|
||||||
|
|
||||||
const (
|
const (
|
||||||
|
@ -127,72 +84,7 @@ const (
|
||||||
}`
|
}`
|
||||||
)
|
)
|
||||||
|
|
||||||
func (b *ElasticSearchIndexer) realIndexerName() string {
|
func (b *Indexer) addUpdate(ctx context.Context, batchWriter git.WriteCloserError, batchReader *bufio.Reader, sha string, update internal.FileUpdate, repo *repo_model.Repository) ([]elastic.BulkableRequest, error) {
|
||||||
return fmt.Sprintf("%s.v%d", b.indexerAliasName, esRepoIndexerLatestVersion)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Init will initialize the indexer
|
|
||||||
func (b *ElasticSearchIndexer) init() (bool, error) {
|
|
||||||
ctx := graceful.GetManager().HammerContext()
|
|
||||||
exists, err := b.client.IndexExists(b.realIndexerName()).Do(ctx)
|
|
||||||
if err != nil {
|
|
||||||
return false, b.checkError(err)
|
|
||||||
}
|
|
||||||
if !exists {
|
|
||||||
mapping := defaultMapping
|
|
||||||
|
|
||||||
createIndex, err := b.client.CreateIndex(b.realIndexerName()).BodyString(mapping).Do(ctx)
|
|
||||||
if err != nil {
|
|
||||||
return false, b.checkError(err)
|
|
||||||
}
|
|
||||||
if !createIndex.Acknowledged {
|
|
||||||
return false, fmt.Errorf("create index %s with %s failed", b.realIndexerName(), mapping)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// check version
|
|
||||||
r, err := b.client.Aliases().Do(ctx)
|
|
||||||
if err != nil {
|
|
||||||
return false, b.checkError(err)
|
|
||||||
}
|
|
||||||
|
|
||||||
realIndexerNames := r.IndicesByAlias(b.indexerAliasName)
|
|
||||||
if len(realIndexerNames) < 1 {
|
|
||||||
res, err := b.client.Alias().
|
|
||||||
Add(b.realIndexerName(), b.indexerAliasName).
|
|
||||||
Do(ctx)
|
|
||||||
if err != nil {
|
|
||||||
return false, b.checkError(err)
|
|
||||||
}
|
|
||||||
if !res.Acknowledged {
|
|
||||||
return false, fmt.Errorf("create alias %s to index %s failed", b.indexerAliasName, b.realIndexerName())
|
|
||||||
}
|
|
||||||
} else if len(realIndexerNames) >= 1 && realIndexerNames[0] < b.realIndexerName() {
|
|
||||||
log.Warn("Found older gitea indexer named %s, but we will create a new one %s and keep the old NOT DELETED. You can delete the old version after the upgrade succeed.",
|
|
||||||
realIndexerNames[0], b.realIndexerName())
|
|
||||||
res, err := b.client.Alias().
|
|
||||||
Remove(realIndexerNames[0], b.indexerAliasName).
|
|
||||||
Add(b.realIndexerName(), b.indexerAliasName).
|
|
||||||
Do(ctx)
|
|
||||||
if err != nil {
|
|
||||||
return false, b.checkError(err)
|
|
||||||
}
|
|
||||||
if !res.Acknowledged {
|
|
||||||
return false, fmt.Errorf("change alias %s to index %s failed", b.indexerAliasName, b.realIndexerName())
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return exists, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Ping checks if elastic is available
|
|
||||||
func (b *ElasticSearchIndexer) Ping() bool {
|
|
||||||
b.lock.RLock()
|
|
||||||
defer b.lock.RUnlock()
|
|
||||||
return b.available
|
|
||||||
}
|
|
||||||
|
|
||||||
func (b *ElasticSearchIndexer) addUpdate(ctx context.Context, batchWriter git.WriteCloserError, batchReader *bufio.Reader, sha string, update fileUpdate, repo *repo_model.Repository) ([]elastic.BulkableRequest, error) {
|
|
||||||
// Ignore vendored files in code search
|
// Ignore vendored files in code search
|
||||||
if setting.Indexer.ExcludeVendored && analyze.IsVendor(update.Filename) {
|
if setting.Indexer.ExcludeVendored && analyze.IsVendor(update.Filename) {
|
||||||
return nil, nil
|
return nil, nil
|
||||||
|
@ -235,11 +127,11 @@ func (b *ElasticSearchIndexer) addUpdate(ctx context.Context, batchWriter git.Wr
|
||||||
if _, err = batchReader.Discard(1); err != nil {
|
if _, err = batchReader.Discard(1); err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
id := filenameIndexerID(repo.ID, update.Filename)
|
id := internal.FilenameIndexerID(repo.ID, update.Filename)
|
||||||
|
|
||||||
return []elastic.BulkableRequest{
|
return []elastic.BulkableRequest{
|
||||||
elastic.NewBulkIndexRequest().
|
elastic.NewBulkIndexRequest().
|
||||||
Index(b.indexerAliasName).
|
Index(b.inner.VersionedIndexName()).
|
||||||
Id(id).
|
Id(id).
|
||||||
Doc(map[string]interface{}{
|
Doc(map[string]interface{}{
|
||||||
"repo_id": repo.ID,
|
"repo_id": repo.ID,
|
||||||
|
@ -251,15 +143,15 @@ func (b *ElasticSearchIndexer) addUpdate(ctx context.Context, batchWriter git.Wr
|
||||||
}, nil
|
}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (b *ElasticSearchIndexer) addDelete(filename string, repo *repo_model.Repository) elastic.BulkableRequest {
|
func (b *Indexer) addDelete(filename string, repo *repo_model.Repository) elastic.BulkableRequest {
|
||||||
id := filenameIndexerID(repo.ID, filename)
|
id := internal.FilenameIndexerID(repo.ID, filename)
|
||||||
return elastic.NewBulkDeleteRequest().
|
return elastic.NewBulkDeleteRequest().
|
||||||
Index(b.indexerAliasName).
|
Index(b.inner.VersionedIndexName()).
|
||||||
Id(id)
|
Id(id)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Index will save the index data
|
// Index will save the index data
|
||||||
func (b *ElasticSearchIndexer) Index(ctx context.Context, repo *repo_model.Repository, sha string, changes *repoChanges) error {
|
func (b *Indexer) Index(ctx context.Context, repo *repo_model.Repository, sha string, changes *internal.RepoChanges) error {
|
||||||
reqs := make([]elastic.BulkableRequest, 0)
|
reqs := make([]elastic.BulkableRequest, 0)
|
||||||
if len(changes.Updates) > 0 {
|
if len(changes.Updates) > 0 {
|
||||||
// Now because of some insanity with git cat-file not immediately failing if not run in a valid git directory we need to run git rev-parse first!
|
// Now because of some insanity with git cat-file not immediately failing if not run in a valid git directory we need to run git rev-parse first!
|
||||||
|
@ -288,21 +180,21 @@ func (b *ElasticSearchIndexer) Index(ctx context.Context, repo *repo_model.Repos
|
||||||
}
|
}
|
||||||
|
|
||||||
if len(reqs) > 0 {
|
if len(reqs) > 0 {
|
||||||
_, err := b.client.Bulk().
|
_, err := b.inner.Client.Bulk().
|
||||||
Index(b.indexerAliasName).
|
Index(b.inner.VersionedIndexName()).
|
||||||
Add(reqs...).
|
Add(reqs...).
|
||||||
Do(ctx)
|
Do(ctx)
|
||||||
return b.checkError(err)
|
return err
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Delete deletes indexes by ids
|
// Delete deletes indexes by ids
|
||||||
func (b *ElasticSearchIndexer) Delete(repoID int64) error {
|
func (b *Indexer) Delete(ctx context.Context, repoID int64) error {
|
||||||
_, err := b.client.DeleteByQuery(b.indexerAliasName).
|
_, err := b.inner.Client.DeleteByQuery(b.inner.VersionedIndexName()).
|
||||||
Query(elastic.NewTermsQuery("repo_id", repoID)).
|
Query(elastic.NewTermsQuery("repo_id", repoID)).
|
||||||
Do(graceful.GetManager().HammerContext())
|
Do(ctx)
|
||||||
return b.checkError(err)
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
// indexPos find words positions for start and the following end on content. It will
|
// indexPos find words positions for start and the following end on content. It will
|
||||||
|
@ -321,8 +213,8 @@ func indexPos(content, start, end string) (int, int) {
|
||||||
return startIdx, startIdx + len(start) + endIdx + len(end)
|
return startIdx, startIdx + len(start) + endIdx + len(end)
|
||||||
}
|
}
|
||||||
|
|
||||||
func convertResult(searchResult *elastic.SearchResult, kw string, pageSize int) (int64, []*SearchResult, []*SearchResultLanguages, error) {
|
func convertResult(searchResult *elastic.SearchResult, kw string, pageSize int) (int64, []*internal.SearchResult, []*internal.SearchResultLanguages, error) {
|
||||||
hits := make([]*SearchResult, 0, pageSize)
|
hits := make([]*internal.SearchResult, 0, pageSize)
|
||||||
for _, hit := range searchResult.Hits.Hits {
|
for _, hit := range searchResult.Hits.Hits {
|
||||||
// FIXME: There is no way to get the position the keyword on the content currently on the same request.
|
// FIXME: There is no way to get the position the keyword on the content currently on the same request.
|
||||||
// So we get it from content, this may made the query slower. See
|
// So we get it from content, this may made the query slower. See
|
||||||
|
@ -341,7 +233,7 @@ func convertResult(searchResult *elastic.SearchResult, kw string, pageSize int)
|
||||||
panic(fmt.Sprintf("2===%#v", hit.Highlight))
|
panic(fmt.Sprintf("2===%#v", hit.Highlight))
|
||||||
}
|
}
|
||||||
|
|
||||||
repoID, fileName := parseIndexerID(hit.Id)
|
repoID, fileName := internal.ParseIndexerID(hit.Id)
|
||||||
res := make(map[string]interface{})
|
res := make(map[string]interface{})
|
||||||
if err := json.Unmarshal(hit.Source, &res); err != nil {
|
if err := json.Unmarshal(hit.Source, &res); err != nil {
|
||||||
return 0, nil, nil, err
|
return 0, nil, nil, err
|
||||||
|
@ -349,7 +241,7 @@ func convertResult(searchResult *elastic.SearchResult, kw string, pageSize int)
|
||||||
|
|
||||||
language := res["language"].(string)
|
language := res["language"].(string)
|
||||||
|
|
||||||
hits = append(hits, &SearchResult{
|
hits = append(hits, &internal.SearchResult{
|
||||||
RepoID: repoID,
|
RepoID: repoID,
|
||||||
Filename: fileName,
|
Filename: fileName,
|
||||||
CommitID: res["commit_id"].(string),
|
CommitID: res["commit_id"].(string),
|
||||||
|
@ -365,14 +257,14 @@ func convertResult(searchResult *elastic.SearchResult, kw string, pageSize int)
|
||||||
return searchResult.TotalHits(), hits, extractAggs(searchResult), nil
|
return searchResult.TotalHits(), hits, extractAggs(searchResult), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func extractAggs(searchResult *elastic.SearchResult) []*SearchResultLanguages {
|
func extractAggs(searchResult *elastic.SearchResult) []*internal.SearchResultLanguages {
|
||||||
var searchResultLanguages []*SearchResultLanguages
|
var searchResultLanguages []*internal.SearchResultLanguages
|
||||||
agg, found := searchResult.Aggregations.Terms("language")
|
agg, found := searchResult.Aggregations.Terms("language")
|
||||||
if found {
|
if found {
|
||||||
searchResultLanguages = make([]*SearchResultLanguages, 0, 10)
|
searchResultLanguages = make([]*internal.SearchResultLanguages, 0, 10)
|
||||||
|
|
||||||
for _, bucket := range agg.Buckets {
|
for _, bucket := range agg.Buckets {
|
||||||
searchResultLanguages = append(searchResultLanguages, &SearchResultLanguages{
|
searchResultLanguages = append(searchResultLanguages, &internal.SearchResultLanguages{
|
||||||
Language: bucket.Key.(string),
|
Language: bucket.Key.(string),
|
||||||
Color: enry.GetColor(bucket.Key.(string)),
|
Color: enry.GetColor(bucket.Key.(string)),
|
||||||
Count: int(bucket.DocCount),
|
Count: int(bucket.DocCount),
|
||||||
|
@ -383,7 +275,7 @@ func extractAggs(searchResult *elastic.SearchResult) []*SearchResultLanguages {
|
||||||
}
|
}
|
||||||
|
|
||||||
// Search searches for codes and language stats by given conditions.
|
// Search searches for codes and language stats by given conditions.
|
||||||
func (b *ElasticSearchIndexer) Search(ctx context.Context, repoIDs []int64, language, keyword string, page, pageSize int, isMatch bool) (int64, []*SearchResult, []*SearchResultLanguages, error) {
|
func (b *Indexer) Search(ctx context.Context, repoIDs []int64, language, keyword string, page, pageSize int, isMatch bool) (int64, []*internal.SearchResult, []*internal.SearchResultLanguages, error) {
|
||||||
searchType := esMultiMatchTypeBestFields
|
searchType := esMultiMatchTypeBestFields
|
||||||
if isMatch {
|
if isMatch {
|
||||||
searchType = esMultiMatchTypePhrasePrefix
|
searchType = esMultiMatchTypePhrasePrefix
|
||||||
|
@ -412,8 +304,8 @@ func (b *ElasticSearchIndexer) Search(ctx context.Context, repoIDs []int64, lang
|
||||||
}
|
}
|
||||||
|
|
||||||
if len(language) == 0 {
|
if len(language) == 0 {
|
||||||
searchResult, err := b.client.Search().
|
searchResult, err := b.inner.Client.Search().
|
||||||
Index(b.indexerAliasName).
|
Index(b.inner.VersionedIndexName()).
|
||||||
Aggregation("language", aggregation).
|
Aggregation("language", aggregation).
|
||||||
Query(query).
|
Query(query).
|
||||||
Highlight(
|
Highlight(
|
||||||
|
@ -426,26 +318,26 @@ func (b *ElasticSearchIndexer) Search(ctx context.Context, repoIDs []int64, lang
|
||||||
From(start).Size(pageSize).
|
From(start).Size(pageSize).
|
||||||
Do(ctx)
|
Do(ctx)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return 0, nil, nil, b.checkError(err)
|
return 0, nil, nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
return convertResult(searchResult, kw, pageSize)
|
return convertResult(searchResult, kw, pageSize)
|
||||||
}
|
}
|
||||||
|
|
||||||
langQuery := elastic.NewMatchQuery("language", language)
|
langQuery := elastic.NewMatchQuery("language", language)
|
||||||
countResult, err := b.client.Search().
|
countResult, err := b.inner.Client.Search().
|
||||||
Index(b.indexerAliasName).
|
Index(b.inner.VersionedIndexName()).
|
||||||
Aggregation("language", aggregation).
|
Aggregation("language", aggregation).
|
||||||
Query(query).
|
Query(query).
|
||||||
Size(0). // We only needs stats information
|
Size(0). // We only need stats information
|
||||||
Do(ctx)
|
Do(ctx)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return 0, nil, nil, b.checkError(err)
|
return 0, nil, nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
query = query.Must(langQuery)
|
query = query.Must(langQuery)
|
||||||
searchResult, err := b.client.Search().
|
searchResult, err := b.inner.Client.Search().
|
||||||
Index(b.indexerAliasName).
|
Index(b.inner.VersionedIndexName()).
|
||||||
Query(query).
|
Query(query).
|
||||||
Highlight(
|
Highlight(
|
||||||
elastic.NewHighlight().
|
elastic.NewHighlight().
|
||||||
|
@ -457,56 +349,10 @@ func (b *ElasticSearchIndexer) Search(ctx context.Context, repoIDs []int64, lang
|
||||||
From(start).Size(pageSize).
|
From(start).Size(pageSize).
|
||||||
Do(ctx)
|
Do(ctx)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return 0, nil, nil, b.checkError(err)
|
return 0, nil, nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
total, hits, _, err := convertResult(searchResult, kw, pageSize)
|
total, hits, _, err := convertResult(searchResult, kw, pageSize)
|
||||||
|
|
||||||
return total, hits, extractAggs(countResult), err
|
return total, hits, extractAggs(countResult), err
|
||||||
}
|
}
|
||||||
|
|
||||||
// Close implements indexer
|
|
||||||
func (b *ElasticSearchIndexer) Close() {
|
|
||||||
select {
|
|
||||||
case <-b.stopTimer:
|
|
||||||
default:
|
|
||||||
close(b.stopTimer)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (b *ElasticSearchIndexer) checkError(err error) error {
|
|
||||||
var opErr *net.OpError
|
|
||||||
if !(elastic.IsConnErr(err) || (errors.As(err, &opErr) && (opErr.Op == "dial" || opErr.Op == "read"))) {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
b.setAvailability(false)
|
|
||||||
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
func (b *ElasticSearchIndexer) checkAvailability() {
|
|
||||||
if b.Ping() {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
// Request cluster state to check if elastic is available again
|
|
||||||
_, err := b.client.ClusterState().Do(graceful.GetManager().ShutdownContext())
|
|
||||||
if err != nil {
|
|
||||||
b.setAvailability(false)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
b.setAvailability(true)
|
|
||||||
}
|
|
||||||
|
|
||||||
func (b *ElasticSearchIndexer) setAvailability(available bool) {
|
|
||||||
b.lock.Lock()
|
|
||||||
defer b.lock.Unlock()
|
|
||||||
|
|
||||||
if b.available == available {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
b.available = available
|
|
||||||
}
|
|
16
modules/indexer/code/elasticsearch/elasticsearch_test.go
Normal file
16
modules/indexer/code/elasticsearch/elasticsearch_test.go
Normal file
|
@ -0,0 +1,16 @@
|
||||||
|
// Copyright 2020 The Gitea Authors. All rights reserved.
|
||||||
|
// SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
|
package elasticsearch
|
||||||
|
|
||||||
|
import (
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
"github.com/stretchr/testify/assert"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestIndexPos(t *testing.T) {
|
||||||
|
startIdx, endIdx := indexPos("test index start and end", "start", "end")
|
||||||
|
assert.EqualValues(t, 11, startIdx)
|
||||||
|
assert.EqualValues(t, 24, endIdx)
|
||||||
|
}
|
|
@ -10,23 +10,11 @@ import (
|
||||||
|
|
||||||
repo_model "code.gitea.io/gitea/models/repo"
|
repo_model "code.gitea.io/gitea/models/repo"
|
||||||
"code.gitea.io/gitea/modules/git"
|
"code.gitea.io/gitea/modules/git"
|
||||||
|
"code.gitea.io/gitea/modules/indexer/code/internal"
|
||||||
"code.gitea.io/gitea/modules/log"
|
"code.gitea.io/gitea/modules/log"
|
||||||
"code.gitea.io/gitea/modules/setting"
|
"code.gitea.io/gitea/modules/setting"
|
||||||
)
|
)
|
||||||
|
|
||||||
type fileUpdate struct {
|
|
||||||
Filename string
|
|
||||||
BlobSha string
|
|
||||||
Size int64
|
|
||||||
Sized bool
|
|
||||||
}
|
|
||||||
|
|
||||||
// repoChanges changes (file additions/updates/removals) to a repo
|
|
||||||
type repoChanges struct {
|
|
||||||
Updates []fileUpdate
|
|
||||||
RemovedFilenames []string
|
|
||||||
}
|
|
||||||
|
|
||||||
func getDefaultBranchSha(ctx context.Context, repo *repo_model.Repository) (string, error) {
|
func getDefaultBranchSha(ctx context.Context, repo *repo_model.Repository) (string, error) {
|
||||||
stdout, _, err := git.NewCommand(ctx, "show-ref", "-s").AddDynamicArguments(git.BranchPrefix + repo.DefaultBranch).RunStdString(&git.RunOpts{Dir: repo.RepoPath()})
|
stdout, _, err := git.NewCommand(ctx, "show-ref", "-s").AddDynamicArguments(git.BranchPrefix + repo.DefaultBranch).RunStdString(&git.RunOpts{Dir: repo.RepoPath()})
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
@ -36,7 +24,7 @@ func getDefaultBranchSha(ctx context.Context, repo *repo_model.Repository) (stri
|
||||||
}
|
}
|
||||||
|
|
||||||
// getRepoChanges returns changes to repo since last indexer update
|
// getRepoChanges returns changes to repo since last indexer update
|
||||||
func getRepoChanges(ctx context.Context, repo *repo_model.Repository, revision string) (*repoChanges, error) {
|
func getRepoChanges(ctx context.Context, repo *repo_model.Repository, revision string) (*internal.RepoChanges, error) {
|
||||||
status, err := repo_model.GetIndexerStatus(ctx, repo, repo_model.RepoIndexerTypeCode)
|
status, err := repo_model.GetIndexerStatus(ctx, repo, repo_model.RepoIndexerTypeCode)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
|
@ -67,16 +55,16 @@ func isIndexable(entry *git.TreeEntry) bool {
|
||||||
}
|
}
|
||||||
|
|
||||||
// parseGitLsTreeOutput parses the output of a `git ls-tree -r --full-name` command
|
// parseGitLsTreeOutput parses the output of a `git ls-tree -r --full-name` command
|
||||||
func parseGitLsTreeOutput(stdout []byte) ([]fileUpdate, error) {
|
func parseGitLsTreeOutput(stdout []byte) ([]internal.FileUpdate, error) {
|
||||||
entries, err := git.ParseTreeEntries(stdout)
|
entries, err := git.ParseTreeEntries(stdout)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
idxCount := 0
|
idxCount := 0
|
||||||
updates := make([]fileUpdate, len(entries))
|
updates := make([]internal.FileUpdate, len(entries))
|
||||||
for _, entry := range entries {
|
for _, entry := range entries {
|
||||||
if isIndexable(entry) {
|
if isIndexable(entry) {
|
||||||
updates[idxCount] = fileUpdate{
|
updates[idxCount] = internal.FileUpdate{
|
||||||
Filename: entry.Name(),
|
Filename: entry.Name(),
|
||||||
BlobSha: entry.ID.String(),
|
BlobSha: entry.ID.String(),
|
||||||
Size: entry.Size(),
|
Size: entry.Size(),
|
||||||
|
@ -89,8 +77,8 @@ func parseGitLsTreeOutput(stdout []byte) ([]fileUpdate, error) {
|
||||||
}
|
}
|
||||||
|
|
||||||
// genesisChanges get changes to add repo to the indexer for the first time
|
// genesisChanges get changes to add repo to the indexer for the first time
|
||||||
func genesisChanges(ctx context.Context, repo *repo_model.Repository, revision string) (*repoChanges, error) {
|
func genesisChanges(ctx context.Context, repo *repo_model.Repository, revision string) (*internal.RepoChanges, error) {
|
||||||
var changes repoChanges
|
var changes internal.RepoChanges
|
||||||
stdout, _, runErr := git.NewCommand(ctx, "ls-tree", "--full-tree", "-l", "-r").AddDynamicArguments(revision).RunStdBytes(&git.RunOpts{Dir: repo.RepoPath()})
|
stdout, _, runErr := git.NewCommand(ctx, "ls-tree", "--full-tree", "-l", "-r").AddDynamicArguments(revision).RunStdBytes(&git.RunOpts{Dir: repo.RepoPath()})
|
||||||
if runErr != nil {
|
if runErr != nil {
|
||||||
return nil, runErr
|
return nil, runErr
|
||||||
|
@ -102,20 +90,20 @@ func genesisChanges(ctx context.Context, repo *repo_model.Repository, revision s
|
||||||
}
|
}
|
||||||
|
|
||||||
// nonGenesisChanges get changes since the previous indexer update
|
// nonGenesisChanges get changes since the previous indexer update
|
||||||
func nonGenesisChanges(ctx context.Context, repo *repo_model.Repository, revision string) (*repoChanges, error) {
|
func nonGenesisChanges(ctx context.Context, repo *repo_model.Repository, revision string) (*internal.RepoChanges, error) {
|
||||||
diffCmd := git.NewCommand(ctx, "diff", "--name-status").AddDynamicArguments(repo.CodeIndexerStatus.CommitSha, revision)
|
diffCmd := git.NewCommand(ctx, "diff", "--name-status").AddDynamicArguments(repo.CodeIndexerStatus.CommitSha, revision)
|
||||||
stdout, _, runErr := diffCmd.RunStdString(&git.RunOpts{Dir: repo.RepoPath()})
|
stdout, _, runErr := diffCmd.RunStdString(&git.RunOpts{Dir: repo.RepoPath()})
|
||||||
if runErr != nil {
|
if runErr != nil {
|
||||||
// previous commit sha may have been removed by a force push, so
|
// previous commit sha may have been removed by a force push, so
|
||||||
// try rebuilding from scratch
|
// try rebuilding from scratch
|
||||||
log.Warn("git diff: %v", runErr)
|
log.Warn("git diff: %v", runErr)
|
||||||
if err := indexer.Delete(repo.ID); err != nil {
|
if err := (*globalIndexer.Load()).Delete(ctx, repo.ID); err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
return genesisChanges(ctx, repo, revision)
|
return genesisChanges(ctx, repo, revision)
|
||||||
}
|
}
|
||||||
|
|
||||||
var changes repoChanges
|
var changes internal.RepoChanges
|
||||||
var err error
|
var err error
|
||||||
updatedFilenames := make([]string, 0, 10)
|
updatedFilenames := make([]string, 0, 10)
|
||||||
for _, line := range strings.Split(stdout, "\n") {
|
for _, line := range strings.Split(stdout, "\n") {
|
||||||
|
|
|
@ -7,86 +7,41 @@ import (
|
||||||
"context"
|
"context"
|
||||||
"os"
|
"os"
|
||||||
"runtime/pprof"
|
"runtime/pprof"
|
||||||
"strconv"
|
"sync/atomic"
|
||||||
"strings"
|
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"code.gitea.io/gitea/models/db"
|
"code.gitea.io/gitea/models/db"
|
||||||
repo_model "code.gitea.io/gitea/models/repo"
|
repo_model "code.gitea.io/gitea/models/repo"
|
||||||
"code.gitea.io/gitea/modules/graceful"
|
"code.gitea.io/gitea/modules/graceful"
|
||||||
|
"code.gitea.io/gitea/modules/indexer/code/bleve"
|
||||||
|
"code.gitea.io/gitea/modules/indexer/code/elasticsearch"
|
||||||
|
"code.gitea.io/gitea/modules/indexer/code/internal"
|
||||||
"code.gitea.io/gitea/modules/log"
|
"code.gitea.io/gitea/modules/log"
|
||||||
"code.gitea.io/gitea/modules/process"
|
"code.gitea.io/gitea/modules/process"
|
||||||
"code.gitea.io/gitea/modules/queue"
|
"code.gitea.io/gitea/modules/queue"
|
||||||
"code.gitea.io/gitea/modules/setting"
|
"code.gitea.io/gitea/modules/setting"
|
||||||
"code.gitea.io/gitea/modules/timeutil"
|
|
||||||
"code.gitea.io/gitea/modules/util"
|
"code.gitea.io/gitea/modules/util"
|
||||||
)
|
)
|
||||||
|
|
||||||
// SearchResult result of performing a search in a repo
|
var (
|
||||||
type SearchResult struct {
|
indexerQueue *queue.WorkerPoolQueue[*internal.IndexerData]
|
||||||
RepoID int64
|
// globalIndexer is the global indexer, it cannot be nil.
|
||||||
StartIndex int
|
// When the real indexer is not ready, it will be a dummy indexer which will return error to explain it's not ready.
|
||||||
EndIndex int
|
// So it's always safe use it as *globalIndexer.Load() and call its methods.
|
||||||
Filename string
|
globalIndexer atomic.Pointer[internal.Indexer]
|
||||||
Content string
|
dummyIndexer *internal.Indexer
|
||||||
CommitID string
|
)
|
||||||
UpdatedUnix timeutil.TimeStamp
|
|
||||||
Language string
|
func init() {
|
||||||
Color string
|
i := internal.NewDummyIndexer()
|
||||||
|
dummyIndexer = &i
|
||||||
|
globalIndexer.Store(dummyIndexer)
|
||||||
}
|
}
|
||||||
|
|
||||||
// SearchResultLanguages result of top languages count in search results
|
func index(ctx context.Context, indexer internal.Indexer, repoID int64) error {
|
||||||
type SearchResultLanguages struct {
|
|
||||||
Language string
|
|
||||||
Color string
|
|
||||||
Count int
|
|
||||||
}
|
|
||||||
|
|
||||||
// Indexer defines an interface to index and search code contents
|
|
||||||
type Indexer interface {
|
|
||||||
Ping() bool
|
|
||||||
Index(ctx context.Context, repo *repo_model.Repository, sha string, changes *repoChanges) error
|
|
||||||
Delete(repoID int64) error
|
|
||||||
Search(ctx context.Context, repoIDs []int64, language, keyword string, page, pageSize int, isMatch bool) (int64, []*SearchResult, []*SearchResultLanguages, error)
|
|
||||||
Close()
|
|
||||||
}
|
|
||||||
|
|
||||||
func filenameIndexerID(repoID int64, filename string) string {
|
|
||||||
return indexerID(repoID) + "_" + filename
|
|
||||||
}
|
|
||||||
|
|
||||||
func indexerID(id int64) string {
|
|
||||||
return strconv.FormatInt(id, 36)
|
|
||||||
}
|
|
||||||
|
|
||||||
func parseIndexerID(indexerID string) (int64, string) {
|
|
||||||
index := strings.IndexByte(indexerID, '_')
|
|
||||||
if index == -1 {
|
|
||||||
log.Error("Unexpected ID in repo indexer: %s", indexerID)
|
|
||||||
}
|
|
||||||
repoID, _ := strconv.ParseInt(indexerID[:index], 36, 64)
|
|
||||||
return repoID, indexerID[index+1:]
|
|
||||||
}
|
|
||||||
|
|
||||||
func filenameOfIndexerID(indexerID string) string {
|
|
||||||
index := strings.IndexByte(indexerID, '_')
|
|
||||||
if index == -1 {
|
|
||||||
log.Error("Unexpected ID in repo indexer: %s", indexerID)
|
|
||||||
}
|
|
||||||
return indexerID[index+1:]
|
|
||||||
}
|
|
||||||
|
|
||||||
// IndexerData represents data stored in the code indexer
|
|
||||||
type IndexerData struct {
|
|
||||||
RepoID int64
|
|
||||||
}
|
|
||||||
|
|
||||||
var indexerQueue *queue.WorkerPoolQueue[*IndexerData]
|
|
||||||
|
|
||||||
func index(ctx context.Context, indexer Indexer, repoID int64) error {
|
|
||||||
repo, err := repo_model.GetRepositoryByID(ctx, repoID)
|
repo, err := repo_model.GetRepositoryByID(ctx, repoID)
|
||||||
if repo_model.IsErrRepoNotExist(err) {
|
if repo_model.IsErrRepoNotExist(err) {
|
||||||
return indexer.Delete(repoID)
|
return indexer.Delete(ctx, repoID)
|
||||||
}
|
}
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
|
@ -139,7 +94,7 @@ func index(ctx context.Context, indexer Indexer, repoID int64) error {
|
||||||
// Init initialize the repo indexer
|
// Init initialize the repo indexer
|
||||||
func Init() {
|
func Init() {
|
||||||
if !setting.Indexer.RepoIndexerEnabled {
|
if !setting.Indexer.RepoIndexerEnabled {
|
||||||
indexer.Close()
|
(*globalIndexer.Load()).Close()
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -153,7 +108,7 @@ func Init() {
|
||||||
}
|
}
|
||||||
cancel()
|
cancel()
|
||||||
log.Debug("Closing repository indexer")
|
log.Debug("Closing repository indexer")
|
||||||
indexer.Close()
|
(*globalIndexer.Load()).Close()
|
||||||
log.Info("PID: %d Repository Indexer closed", os.Getpid())
|
log.Info("PID: %d Repository Indexer closed", os.Getpid())
|
||||||
finished()
|
finished()
|
||||||
})
|
})
|
||||||
|
@ -163,13 +118,8 @@ func Init() {
|
||||||
// Create the Queue
|
// Create the Queue
|
||||||
switch setting.Indexer.RepoType {
|
switch setting.Indexer.RepoType {
|
||||||
case "bleve", "elasticsearch":
|
case "bleve", "elasticsearch":
|
||||||
handler := func(items ...*IndexerData) (unhandled []*IndexerData) {
|
handler := func(items ...*internal.IndexerData) (unhandled []*internal.IndexerData) {
|
||||||
idx, err := indexer.get()
|
indexer := *globalIndexer.Load()
|
||||||
if idx == nil || err != nil {
|
|
||||||
log.Warn("Codes indexer handler: indexer is not ready, retry later.")
|
|
||||||
return items
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, indexerData := range items {
|
for _, indexerData := range items {
|
||||||
log.Trace("IndexerData Process Repo: %d", indexerData.RepoID)
|
log.Trace("IndexerData Process Repo: %d", indexerData.RepoID)
|
||||||
|
|
||||||
|
@ -188,11 +138,7 @@ func Init() {
|
||||||
code.gitea.io/gitea/modules/indexer/code.index(indexer.go:105)
|
code.gitea.io/gitea/modules/indexer/code.index(indexer.go:105)
|
||||||
*/
|
*/
|
||||||
if err := index(ctx, indexer, indexerData.RepoID); err != nil {
|
if err := index(ctx, indexer, indexerData.RepoID); err != nil {
|
||||||
if !idx.Ping() {
|
unhandled = append(unhandled, indexerData)
|
||||||
log.Error("Code indexer handler: indexer is unavailable.")
|
|
||||||
unhandled = append(unhandled, indexerData)
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
if !setting.IsInTesting {
|
if !setting.IsInTesting {
|
||||||
log.Error("Codes indexer handler: index error for repo %v: %v", indexerData.RepoID, err)
|
log.Error("Codes indexer handler: index error for repo %v: %v", indexerData.RepoID, err)
|
||||||
}
|
}
|
||||||
|
@ -213,8 +159,8 @@ func Init() {
|
||||||
pprof.SetGoroutineLabels(ctx)
|
pprof.SetGoroutineLabels(ctx)
|
||||||
start := time.Now()
|
start := time.Now()
|
||||||
var (
|
var (
|
||||||
rIndexer Indexer
|
rIndexer internal.Indexer
|
||||||
populate bool
|
existed bool
|
||||||
err error
|
err error
|
||||||
)
|
)
|
||||||
switch setting.Indexer.RepoType {
|
switch setting.Indexer.RepoType {
|
||||||
|
@ -228,10 +174,11 @@ func Init() {
|
||||||
}
|
}
|
||||||
}()
|
}()
|
||||||
|
|
||||||
rIndexer, populate, err = NewBleveIndexer(setting.Indexer.RepoPath)
|
rIndexer = bleve.NewIndexer(setting.Indexer.RepoPath)
|
||||||
|
existed, err = rIndexer.Init(ctx)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
cancel()
|
cancel()
|
||||||
indexer.Close()
|
(*globalIndexer.Load()).Close()
|
||||||
close(waitChannel)
|
close(waitChannel)
|
||||||
log.Fatal("PID: %d Unable to initialize the bleve Repository Indexer at path: %s Error: %v", os.Getpid(), setting.Indexer.RepoPath, err)
|
log.Fatal("PID: %d Unable to initialize the bleve Repository Indexer at path: %s Error: %v", os.Getpid(), setting.Indexer.RepoPath, err)
|
||||||
}
|
}
|
||||||
|
@ -245,23 +192,31 @@ func Init() {
|
||||||
}
|
}
|
||||||
}()
|
}()
|
||||||
|
|
||||||
rIndexer, populate, err = NewElasticSearchIndexer(setting.Indexer.RepoConnStr, setting.Indexer.RepoIndexerName)
|
rIndexer = elasticsearch.NewIndexer(setting.Indexer.RepoConnStr, setting.Indexer.RepoIndexerName)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
cancel()
|
cancel()
|
||||||
indexer.Close()
|
(*globalIndexer.Load()).Close()
|
||||||
|
close(waitChannel)
|
||||||
|
log.Fatal("PID: %d Unable to create the elasticsearch Repository Indexer connstr: %s Error: %v", os.Getpid(), setting.Indexer.RepoConnStr, err)
|
||||||
|
}
|
||||||
|
existed, err = rIndexer.Init(ctx)
|
||||||
|
if err != nil {
|
||||||
|
cancel()
|
||||||
|
(*globalIndexer.Load()).Close()
|
||||||
close(waitChannel)
|
close(waitChannel)
|
||||||
log.Fatal("PID: %d Unable to initialize the elasticsearch Repository Indexer connstr: %s Error: %v", os.Getpid(), setting.Indexer.RepoConnStr, err)
|
log.Fatal("PID: %d Unable to initialize the elasticsearch Repository Indexer connstr: %s Error: %v", os.Getpid(), setting.Indexer.RepoConnStr, err)
|
||||||
}
|
}
|
||||||
|
|
||||||
default:
|
default:
|
||||||
log.Fatal("PID: %d Unknown Indexer type: %s", os.Getpid(), setting.Indexer.RepoType)
|
log.Fatal("PID: %d Unknown Indexer type: %s", os.Getpid(), setting.Indexer.RepoType)
|
||||||
}
|
}
|
||||||
|
|
||||||
indexer.set(rIndexer)
|
globalIndexer.Store(&rIndexer)
|
||||||
|
|
||||||
// Start processing the queue
|
// Start processing the queue
|
||||||
go graceful.GetManager().RunWithCancel(indexerQueue)
|
go graceful.GetManager().RunWithCancel(indexerQueue)
|
||||||
|
|
||||||
if populate {
|
if !existed { // populate the index because it's created for the first time
|
||||||
go graceful.GetManager().RunWithShutdownContext(populateRepoIndexer)
|
go graceful.GetManager().RunWithShutdownContext(populateRepoIndexer)
|
||||||
}
|
}
|
||||||
select {
|
select {
|
||||||
|
@ -283,18 +238,18 @@ func Init() {
|
||||||
case <-graceful.GetManager().IsShutdown():
|
case <-graceful.GetManager().IsShutdown():
|
||||||
log.Warn("Shutdown before Repository Indexer completed initialization")
|
log.Warn("Shutdown before Repository Indexer completed initialization")
|
||||||
cancel()
|
cancel()
|
||||||
indexer.Close()
|
(*globalIndexer.Load()).Close()
|
||||||
case duration, ok := <-waitChannel:
|
case duration, ok := <-waitChannel:
|
||||||
if !ok {
|
if !ok {
|
||||||
log.Warn("Repository Indexer Initialization failed")
|
log.Warn("Repository Indexer Initialization failed")
|
||||||
cancel()
|
cancel()
|
||||||
indexer.Close()
|
(*globalIndexer.Load()).Close()
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
log.Info("Repository Indexer Initialization took %v", duration)
|
log.Info("Repository Indexer Initialization took %v", duration)
|
||||||
case <-time.After(timeout):
|
case <-time.After(timeout):
|
||||||
cancel()
|
cancel()
|
||||||
indexer.Close()
|
(*globalIndexer.Load()).Close()
|
||||||
log.Fatal("Repository Indexer Initialization Timed-Out after: %v", timeout)
|
log.Fatal("Repository Indexer Initialization Timed-Out after: %v", timeout)
|
||||||
}
|
}
|
||||||
}()
|
}()
|
||||||
|
@ -303,21 +258,15 @@ func Init() {
|
||||||
|
|
||||||
// UpdateRepoIndexer update a repository's entries in the indexer
|
// UpdateRepoIndexer update a repository's entries in the indexer
|
||||||
func UpdateRepoIndexer(repo *repo_model.Repository) {
|
func UpdateRepoIndexer(repo *repo_model.Repository) {
|
||||||
indexData := &IndexerData{RepoID: repo.ID}
|
indexData := &internal.IndexerData{RepoID: repo.ID}
|
||||||
if err := indexerQueue.Push(indexData); err != nil {
|
if err := indexerQueue.Push(indexData); err != nil {
|
||||||
log.Error("Update repo index data %v failed: %v", indexData, err)
|
log.Error("Update repo index data %v failed: %v", indexData, err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// IsAvailable checks if issue indexer is available
|
// IsAvailable checks if issue indexer is available
|
||||||
func IsAvailable() bool {
|
func IsAvailable(ctx context.Context) bool {
|
||||||
idx, err := indexer.get()
|
return (*globalIndexer.Load()).Ping(ctx) == nil
|
||||||
if err != nil {
|
|
||||||
log.Error("IsAvailable(): unable to get indexer: %v", err)
|
|
||||||
return false
|
|
||||||
}
|
|
||||||
|
|
||||||
return idx.Ping()
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// populateRepoIndexer populate the repo indexer with pre-existing data. This
|
// populateRepoIndexer populate the repo indexer with pre-existing data. This
|
||||||
|
@ -368,7 +317,7 @@ func populateRepoIndexer(ctx context.Context) {
|
||||||
return
|
return
|
||||||
default:
|
default:
|
||||||
}
|
}
|
||||||
if err := indexerQueue.Push(&IndexerData{RepoID: id}); err != nil {
|
if err := indexerQueue.Push(&internal.IndexerData{RepoID: id}); err != nil {
|
||||||
log.Error("indexerQueue.Push: %v", err)
|
log.Error("indexerQueue.Push: %v", err)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
|
@ -5,11 +5,15 @@ package code
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
|
"os"
|
||||||
"path/filepath"
|
"path/filepath"
|
||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"code.gitea.io/gitea/models/unittest"
|
"code.gitea.io/gitea/models/unittest"
|
||||||
"code.gitea.io/gitea/modules/git"
|
"code.gitea.io/gitea/modules/git"
|
||||||
|
"code.gitea.io/gitea/modules/indexer/code/bleve"
|
||||||
|
"code.gitea.io/gitea/modules/indexer/code/elasticsearch"
|
||||||
|
"code.gitea.io/gitea/modules/indexer/code/internal"
|
||||||
|
|
||||||
_ "code.gitea.io/gitea/models"
|
_ "code.gitea.io/gitea/models"
|
||||||
|
|
||||||
|
@ -22,7 +26,7 @@ func TestMain(m *testing.M) {
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
func testIndexer(name string, t *testing.T, indexer Indexer) {
|
func testIndexer(name string, t *testing.T, indexer internal.Indexer) {
|
||||||
t.Run(name, func(t *testing.T) {
|
t.Run(name, func(t *testing.T) {
|
||||||
var repoID int64 = 1
|
var repoID int64 = 1
|
||||||
err := index(git.DefaultContext, indexer, repoID)
|
err := index(git.DefaultContext, indexer, repoID)
|
||||||
|
@ -81,6 +85,48 @@ func testIndexer(name string, t *testing.T, indexer Indexer) {
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
assert.NoError(t, indexer.Delete(repoID))
|
assert.NoError(t, indexer.Delete(context.Background(), repoID))
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func TestBleveIndexAndSearch(t *testing.T) {
|
||||||
|
unittest.PrepareTestEnv(t)
|
||||||
|
|
||||||
|
dir := t.TempDir()
|
||||||
|
|
||||||
|
idx := bleve.NewIndexer(dir)
|
||||||
|
_, err := idx.Init(context.Background())
|
||||||
|
if err != nil {
|
||||||
|
assert.Fail(t, "Unable to create bleve indexer Error: %v", err)
|
||||||
|
if idx != nil {
|
||||||
|
idx.Close()
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
defer idx.Close()
|
||||||
|
|
||||||
|
testIndexer("beleve", t, idx)
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestESIndexAndSearch(t *testing.T) {
|
||||||
|
unittest.PrepareTestEnv(t)
|
||||||
|
|
||||||
|
u := os.Getenv("TEST_INDEXER_CODE_ES_URL")
|
||||||
|
if u == "" {
|
||||||
|
t.SkipNow()
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
indexer := elasticsearch.NewIndexer(u, "gitea_codes")
|
||||||
|
if _, err := indexer.Init(context.Background()); err != nil {
|
||||||
|
assert.Fail(t, "Unable to init ES indexer Error: %v", err)
|
||||||
|
if indexer != nil {
|
||||||
|
indexer.Close()
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
defer indexer.Close()
|
||||||
|
|
||||||
|
testIndexer("elastic_search", t, indexer)
|
||||||
|
}
|
||||||
|
|
43
modules/indexer/code/internal/indexer.go
Normal file
43
modules/indexer/code/internal/indexer.go
Normal file
|
@ -0,0 +1,43 @@
|
||||||
|
// Copyright 2023 The Gitea Authors. All rights reserved.
|
||||||
|
// SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
|
package internal
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
repo_model "code.gitea.io/gitea/models/repo"
|
||||||
|
"code.gitea.io/gitea/modules/indexer/internal"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Indexer defines an interface to index and search code contents
|
||||||
|
type Indexer interface {
|
||||||
|
internal.Indexer
|
||||||
|
Index(ctx context.Context, repo *repo_model.Repository, sha string, changes *RepoChanges) error
|
||||||
|
Delete(ctx context.Context, repoID int64) error
|
||||||
|
Search(ctx context.Context, repoIDs []int64, language, keyword string, page, pageSize int, isMatch bool) (int64, []*SearchResult, []*SearchResultLanguages, error)
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewDummyIndexer returns a dummy indexer
|
||||||
|
func NewDummyIndexer() Indexer {
|
||||||
|
return &dummyIndexer{
|
||||||
|
Indexer: internal.NewDummyIndexer(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
type dummyIndexer struct {
|
||||||
|
internal.Indexer
|
||||||
|
}
|
||||||
|
|
||||||
|
func (d *dummyIndexer) Index(ctx context.Context, repo *repo_model.Repository, sha string, changes *RepoChanges) error {
|
||||||
|
return fmt.Errorf("indexer is not ready")
|
||||||
|
}
|
||||||
|
|
||||||
|
func (d *dummyIndexer) Delete(ctx context.Context, repoID int64) error {
|
||||||
|
return fmt.Errorf("indexer is not ready")
|
||||||
|
}
|
||||||
|
|
||||||
|
func (d *dummyIndexer) Search(ctx context.Context, repoIDs []int64, language, keyword string, page, pageSize int, isMatch bool) (int64, []*SearchResult, []*SearchResultLanguages, error) {
|
||||||
|
return 0, nil, nil, fmt.Errorf("indexer is not ready")
|
||||||
|
}
|
44
modules/indexer/code/internal/model.go
Normal file
44
modules/indexer/code/internal/model.go
Normal file
|
@ -0,0 +1,44 @@
|
||||||
|
// Copyright 2023 The Gitea Authors. All rights reserved.
|
||||||
|
// SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
|
package internal
|
||||||
|
|
||||||
|
import "code.gitea.io/gitea/modules/timeutil"
|
||||||
|
|
||||||
|
type FileUpdate struct {
|
||||||
|
Filename string
|
||||||
|
BlobSha string
|
||||||
|
Size int64
|
||||||
|
Sized bool
|
||||||
|
}
|
||||||
|
|
||||||
|
// RepoChanges changes (file additions/updates/removals) to a repo
|
||||||
|
type RepoChanges struct {
|
||||||
|
Updates []FileUpdate
|
||||||
|
RemovedFilenames []string
|
||||||
|
}
|
||||||
|
|
||||||
|
// IndexerData represents data stored in the code indexer
|
||||||
|
type IndexerData struct {
|
||||||
|
RepoID int64
|
||||||
|
}
|
||||||
|
|
||||||
|
// SearchResult result of performing a search in a repo
|
||||||
|
type SearchResult struct {
|
||||||
|
RepoID int64
|
||||||
|
StartIndex int
|
||||||
|
EndIndex int
|
||||||
|
Filename string
|
||||||
|
Content string
|
||||||
|
CommitID string
|
||||||
|
UpdatedUnix timeutil.TimeStamp
|
||||||
|
Language string
|
||||||
|
Color string
|
||||||
|
}
|
||||||
|
|
||||||
|
// SearchResultLanguages result of top languages count in search results
|
||||||
|
type SearchResultLanguages struct {
|
||||||
|
Language string
|
||||||
|
Color string
|
||||||
|
Count int
|
||||||
|
}
|
32
modules/indexer/code/internal/util.go
Normal file
32
modules/indexer/code/internal/util.go
Normal file
|
@ -0,0 +1,32 @@
|
||||||
|
// Copyright 2023 The Gitea Authors. All rights reserved.
|
||||||
|
// SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
|
package internal
|
||||||
|
|
||||||
|
import (
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
"code.gitea.io/gitea/modules/indexer/internal"
|
||||||
|
"code.gitea.io/gitea/modules/log"
|
||||||
|
)
|
||||||
|
|
||||||
|
func FilenameIndexerID(repoID int64, filename string) string {
|
||||||
|
return internal.Base36(repoID) + "_" + filename
|
||||||
|
}
|
||||||
|
|
||||||
|
func ParseIndexerID(indexerID string) (int64, string) {
|
||||||
|
index := strings.IndexByte(indexerID, '_')
|
||||||
|
if index == -1 {
|
||||||
|
log.Error("Unexpected ID in repo indexer: %s", indexerID)
|
||||||
|
}
|
||||||
|
repoID, _ := internal.ParseBase36(indexerID[:index])
|
||||||
|
return repoID, indexerID[index+1:]
|
||||||
|
}
|
||||||
|
|
||||||
|
func FilenameOfIndexerID(indexerID string) string {
|
||||||
|
index := strings.IndexByte(indexerID, '_')
|
||||||
|
if index == -1 {
|
||||||
|
log.Error("Unexpected ID in repo indexer: %s", indexerID)
|
||||||
|
}
|
||||||
|
return indexerID[index+1:]
|
||||||
|
}
|
|
@ -9,6 +9,7 @@ import (
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
"code.gitea.io/gitea/modules/highlight"
|
"code.gitea.io/gitea/modules/highlight"
|
||||||
|
"code.gitea.io/gitea/modules/indexer/code/internal"
|
||||||
"code.gitea.io/gitea/modules/timeutil"
|
"code.gitea.io/gitea/modules/timeutil"
|
||||||
"code.gitea.io/gitea/modules/util"
|
"code.gitea.io/gitea/modules/util"
|
||||||
)
|
)
|
||||||
|
@ -25,6 +26,8 @@ type Result struct {
|
||||||
FormattedLines string
|
FormattedLines string
|
||||||
}
|
}
|
||||||
|
|
||||||
|
type SearchResultLanguages = internal.SearchResultLanguages
|
||||||
|
|
||||||
func indices(content string, selectionStartIndex, selectionEndIndex int) (int, int) {
|
func indices(content string, selectionStartIndex, selectionEndIndex int) (int, int) {
|
||||||
startIndex := selectionStartIndex
|
startIndex := selectionStartIndex
|
||||||
numLinesBefore := 0
|
numLinesBefore := 0
|
||||||
|
@ -61,7 +64,7 @@ func writeStrings(buf *bytes.Buffer, strs ...string) error {
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func searchResult(result *SearchResult, startIndex, endIndex int) (*Result, error) {
|
func searchResult(result *internal.SearchResult, startIndex, endIndex int) (*Result, error) {
|
||||||
startLineNum := 1 + strings.Count(result.Content[:startIndex], "\n")
|
startLineNum := 1 + strings.Count(result.Content[:startIndex], "\n")
|
||||||
|
|
||||||
var formattedLinesBuffer bytes.Buffer
|
var formattedLinesBuffer bytes.Buffer
|
||||||
|
@ -109,12 +112,12 @@ func searchResult(result *SearchResult, startIndex, endIndex int) (*Result, erro
|
||||||
}
|
}
|
||||||
|
|
||||||
// PerformSearch perform a search on a repository
|
// PerformSearch perform a search on a repository
|
||||||
func PerformSearch(ctx context.Context, repoIDs []int64, language, keyword string, page, pageSize int, isMatch bool) (int, []*Result, []*SearchResultLanguages, error) {
|
func PerformSearch(ctx context.Context, repoIDs []int64, language, keyword string, page, pageSize int, isMatch bool) (int, []*Result, []*internal.SearchResultLanguages, error) {
|
||||||
if len(keyword) == 0 {
|
if len(keyword) == 0 {
|
||||||
return 0, nil, nil, nil
|
return 0, nil, nil, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
total, results, resultLanguages, err := indexer.Search(ctx, repoIDs, language, keyword, page, pageSize, isMatch)
|
total, results, resultLanguages, err := (*globalIndexer.Load()).Search(ctx, repoIDs, language, keyword, page, pageSize, isMatch)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return 0, nil, nil, err
|
return 0, nil, nil, err
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,104 +0,0 @@
|
||||||
// Copyright 2019 The Gitea Authors. All rights reserved.
|
|
||||||
// SPDX-License-Identifier: MIT
|
|
||||||
|
|
||||||
package code
|
|
||||||
|
|
||||||
import (
|
|
||||||
"context"
|
|
||||||
"fmt"
|
|
||||||
"sync"
|
|
||||||
|
|
||||||
repo_model "code.gitea.io/gitea/models/repo"
|
|
||||||
"code.gitea.io/gitea/modules/log"
|
|
||||||
)
|
|
||||||
|
|
||||||
var indexer = newWrappedIndexer()
|
|
||||||
|
|
||||||
// ErrWrappedIndexerClosed is the error returned if the indexer was closed before it was ready
|
|
||||||
var ErrWrappedIndexerClosed = fmt.Errorf("Indexer closed before ready")
|
|
||||||
|
|
||||||
type wrappedIndexer struct {
|
|
||||||
internal Indexer
|
|
||||||
lock sync.RWMutex
|
|
||||||
cond *sync.Cond
|
|
||||||
closed bool
|
|
||||||
}
|
|
||||||
|
|
||||||
func newWrappedIndexer() *wrappedIndexer {
|
|
||||||
w := &wrappedIndexer{}
|
|
||||||
w.cond = sync.NewCond(w.lock.RLocker())
|
|
||||||
return w
|
|
||||||
}
|
|
||||||
|
|
||||||
func (w *wrappedIndexer) set(indexer Indexer) {
|
|
||||||
w.lock.Lock()
|
|
||||||
defer w.lock.Unlock()
|
|
||||||
if w.closed {
|
|
||||||
// Too late!
|
|
||||||
indexer.Close()
|
|
||||||
}
|
|
||||||
w.internal = indexer
|
|
||||||
w.cond.Broadcast()
|
|
||||||
}
|
|
||||||
|
|
||||||
func (w *wrappedIndexer) get() (Indexer, error) {
|
|
||||||
w.lock.RLock()
|
|
||||||
defer w.lock.RUnlock()
|
|
||||||
if w.internal == nil {
|
|
||||||
if w.closed {
|
|
||||||
return nil, ErrWrappedIndexerClosed
|
|
||||||
}
|
|
||||||
w.cond.Wait()
|
|
||||||
if w.closed {
|
|
||||||
return nil, ErrWrappedIndexerClosed
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return w.internal, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Ping checks if elastic is available
|
|
||||||
func (w *wrappedIndexer) Ping() bool {
|
|
||||||
indexer, err := w.get()
|
|
||||||
if err != nil {
|
|
||||||
log.Warn("Failed to get indexer: %v", err)
|
|
||||||
return false
|
|
||||||
}
|
|
||||||
return indexer.Ping()
|
|
||||||
}
|
|
||||||
|
|
||||||
func (w *wrappedIndexer) Index(ctx context.Context, repo *repo_model.Repository, sha string, changes *repoChanges) error {
|
|
||||||
indexer, err := w.get()
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
return indexer.Index(ctx, repo, sha, changes)
|
|
||||||
}
|
|
||||||
|
|
||||||
func (w *wrappedIndexer) Delete(repoID int64) error {
|
|
||||||
indexer, err := w.get()
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
return indexer.Delete(repoID)
|
|
||||||
}
|
|
||||||
|
|
||||||
func (w *wrappedIndexer) Search(ctx context.Context, repoIDs []int64, language, keyword string, page, pageSize int, isMatch bool) (int64, []*SearchResult, []*SearchResultLanguages, error) {
|
|
||||||
indexer, err := w.get()
|
|
||||||
if err != nil {
|
|
||||||
return 0, nil, nil, err
|
|
||||||
}
|
|
||||||
return indexer.Search(ctx, repoIDs, language, keyword, page, pageSize, isMatch)
|
|
||||||
}
|
|
||||||
|
|
||||||
func (w *wrappedIndexer) Close() {
|
|
||||||
w.lock.Lock()
|
|
||||||
defer w.lock.Unlock()
|
|
||||||
if w.closed {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
w.closed = true
|
|
||||||
w.cond.Broadcast()
|
|
||||||
if w.internal != nil {
|
|
||||||
w.internal.Close()
|
|
||||||
}
|
|
||||||
}
|
|
21
modules/indexer/internal/base32.go
Normal file
21
modules/indexer/internal/base32.go
Normal file
|
@ -0,0 +1,21 @@
|
||||||
|
// Copyright 2023 The Gitea Authors. All rights reserved.
|
||||||
|
// SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
|
package internal
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"strconv"
|
||||||
|
)
|
||||||
|
|
||||||
|
func Base36(i int64) string {
|
||||||
|
return strconv.FormatInt(i, 36)
|
||||||
|
}
|
||||||
|
|
||||||
|
func ParseBase36(s string) (int64, error) {
|
||||||
|
i, err := strconv.ParseInt(s, 36, 64)
|
||||||
|
if err != nil {
|
||||||
|
return 0, fmt.Errorf("invalid base36 integer %q: %w", s, err)
|
||||||
|
}
|
||||||
|
return i, nil
|
||||||
|
}
|
103
modules/indexer/internal/bleve/indexer.go
Normal file
103
modules/indexer/internal/bleve/indexer.go
Normal file
|
@ -0,0 +1,103 @@
|
||||||
|
// Copyright 2023 The Gitea Authors. All rights reserved.
|
||||||
|
// SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
|
package bleve
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
"code.gitea.io/gitea/modules/indexer/internal"
|
||||||
|
"code.gitea.io/gitea/modules/log"
|
||||||
|
|
||||||
|
"github.com/blevesearch/bleve/v2"
|
||||||
|
"github.com/blevesearch/bleve/v2/mapping"
|
||||||
|
"github.com/ethantkoenig/rupture"
|
||||||
|
)
|
||||||
|
|
||||||
|
var _ internal.Indexer = &Indexer{}
|
||||||
|
|
||||||
|
// Indexer represents a basic bleve indexer implementation
|
||||||
|
type Indexer struct {
|
||||||
|
Indexer bleve.Index
|
||||||
|
|
||||||
|
indexDir string
|
||||||
|
version int
|
||||||
|
mappingGetter MappingGetter
|
||||||
|
}
|
||||||
|
|
||||||
|
type MappingGetter func() (mapping.IndexMapping, error)
|
||||||
|
|
||||||
|
func NewIndexer(indexDir string, version int, mappingGetter func() (mapping.IndexMapping, error)) *Indexer {
|
||||||
|
return &Indexer{
|
||||||
|
indexDir: indexDir,
|
||||||
|
version: version,
|
||||||
|
mappingGetter: mappingGetter,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Init initializes the indexer
|
||||||
|
func (i *Indexer) Init(_ context.Context) (bool, error) {
|
||||||
|
if i == nil {
|
||||||
|
return false, fmt.Errorf("cannot init nil indexer")
|
||||||
|
}
|
||||||
|
|
||||||
|
if i.Indexer != nil {
|
||||||
|
return false, fmt.Errorf("indexer is already initialized")
|
||||||
|
}
|
||||||
|
|
||||||
|
indexer, version, err := openIndexer(i.indexDir, i.version)
|
||||||
|
if err != nil {
|
||||||
|
return false, err
|
||||||
|
}
|
||||||
|
if indexer != nil {
|
||||||
|
i.Indexer = indexer
|
||||||
|
return true, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
if version != 0 {
|
||||||
|
log.Warn("Found older bleve index with version %d, Gitea will remove it and rebuild", version)
|
||||||
|
}
|
||||||
|
|
||||||
|
indexMapping, err := i.mappingGetter()
|
||||||
|
if err != nil {
|
||||||
|
return false, err
|
||||||
|
}
|
||||||
|
|
||||||
|
indexer, err = bleve.New(i.indexDir, indexMapping)
|
||||||
|
if err != nil {
|
||||||
|
return false, err
|
||||||
|
}
|
||||||
|
|
||||||
|
if err = rupture.WriteIndexMetadata(i.indexDir, &rupture.IndexMetadata{
|
||||||
|
Version: i.version,
|
||||||
|
}); err != nil {
|
||||||
|
return false, err
|
||||||
|
}
|
||||||
|
|
||||||
|
i.Indexer = indexer
|
||||||
|
|
||||||
|
return false, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Ping checks if the indexer is available
|
||||||
|
func (i *Indexer) Ping(_ context.Context) error {
|
||||||
|
if i == nil {
|
||||||
|
return fmt.Errorf("cannot ping nil indexer")
|
||||||
|
}
|
||||||
|
if i.Indexer == nil {
|
||||||
|
return fmt.Errorf("indexer is not initialized")
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (i *Indexer) Close() {
|
||||||
|
if i == nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
if err := i.Indexer.Close(); err != nil {
|
||||||
|
log.Error("Failed to close bleve indexer in %q: %v", i.indexDir, err)
|
||||||
|
}
|
||||||
|
i.Indexer = nil
|
||||||
|
}
|
49
modules/indexer/internal/bleve/util.go
Normal file
49
modules/indexer/internal/bleve/util.go
Normal file
|
@ -0,0 +1,49 @@
|
||||||
|
// Copyright 2023 The Gitea Authors. All rights reserved.
|
||||||
|
// SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
|
package bleve
|
||||||
|
|
||||||
|
import (
|
||||||
|
"errors"
|
||||||
|
"os"
|
||||||
|
|
||||||
|
"code.gitea.io/gitea/modules/log"
|
||||||
|
"code.gitea.io/gitea/modules/util"
|
||||||
|
|
||||||
|
"github.com/blevesearch/bleve/v2"
|
||||||
|
"github.com/blevesearch/bleve/v2/index/upsidedown"
|
||||||
|
"github.com/ethantkoenig/rupture"
|
||||||
|
)
|
||||||
|
|
||||||
|
// openIndexer open the index at the specified path, checking for metadata
|
||||||
|
// updates and bleve version updates. If index needs to be created (or
|
||||||
|
// re-created), returns (nil, nil)
|
||||||
|
func openIndexer(path string, latestVersion int) (bleve.Index, int, error) {
|
||||||
|
_, err := os.Stat(path)
|
||||||
|
if err != nil && os.IsNotExist(err) {
|
||||||
|
return nil, 0, nil
|
||||||
|
} else if err != nil {
|
||||||
|
return nil, 0, err
|
||||||
|
}
|
||||||
|
|
||||||
|
metadata, err := rupture.ReadIndexMetadata(path)
|
||||||
|
if err != nil {
|
||||||
|
return nil, 0, err
|
||||||
|
}
|
||||||
|
if metadata.Version < latestVersion {
|
||||||
|
// the indexer is using a previous version, so we should delete it and
|
||||||
|
// re-populate
|
||||||
|
return nil, metadata.Version, util.RemoveAll(path)
|
||||||
|
}
|
||||||
|
|
||||||
|
index, err := bleve.Open(path)
|
||||||
|
if err != nil {
|
||||||
|
if errors.Is(err, upsidedown.IncompatibleVersion) {
|
||||||
|
log.Warn("Indexer was built with a previous version of bleve, deleting and rebuilding")
|
||||||
|
return nil, 0, util.RemoveAll(path)
|
||||||
|
}
|
||||||
|
return nil, 0, err
|
||||||
|
}
|
||||||
|
|
||||||
|
return index, 0, nil
|
||||||
|
}
|
33
modules/indexer/internal/db/indexer.go
Normal file
33
modules/indexer/internal/db/indexer.go
Normal file
|
@ -0,0 +1,33 @@
|
||||||
|
// Copyright 2023 The Gitea Authors. All rights reserved.
|
||||||
|
// SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
|
package db
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
|
||||||
|
"code.gitea.io/gitea/modules/indexer/internal"
|
||||||
|
)
|
||||||
|
|
||||||
|
var _ internal.Indexer = &Indexer{}
|
||||||
|
|
||||||
|
// Indexer represents a basic db indexer implementation
|
||||||
|
type Indexer struct{}
|
||||||
|
|
||||||
|
// Init initializes the indexer
|
||||||
|
func (i *Indexer) Init(_ context.Context) (bool, error) {
|
||||||
|
// nothing to do
|
||||||
|
return false, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Ping checks if the indexer is available
|
||||||
|
func (i *Indexer) Ping(_ context.Context) error {
|
||||||
|
// No need to ping database to check if it is available.
|
||||||
|
// If the database goes down, Gitea will go down, so nobody will care if the indexer is available.
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Close closes the indexer
|
||||||
|
func (i *Indexer) Close() {
|
||||||
|
// nothing to do
|
||||||
|
}
|
92
modules/indexer/internal/elasticsearch/indexer.go
Normal file
92
modules/indexer/internal/elasticsearch/indexer.go
Normal file
|
@ -0,0 +1,92 @@
|
||||||
|
// Copyright 2023 The Gitea Authors. All rights reserved.
|
||||||
|
// SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
|
package elasticsearch
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
"code.gitea.io/gitea/modules/indexer/internal"
|
||||||
|
|
||||||
|
"github.com/olivere/elastic/v7"
|
||||||
|
)
|
||||||
|
|
||||||
|
var _ internal.Indexer = &Indexer{}
|
||||||
|
|
||||||
|
// Indexer represents a basic elasticsearch indexer implementation
|
||||||
|
type Indexer struct {
|
||||||
|
Client *elastic.Client
|
||||||
|
|
||||||
|
url string
|
||||||
|
indexName string
|
||||||
|
version int
|
||||||
|
mapping string
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewIndexer(url, indexName string, version int, mapping string) *Indexer {
|
||||||
|
return &Indexer{
|
||||||
|
url: url,
|
||||||
|
indexName: indexName,
|
||||||
|
version: version,
|
||||||
|
mapping: mapping,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Init initializes the indexer
|
||||||
|
func (i *Indexer) Init(ctx context.Context) (bool, error) {
|
||||||
|
if i == nil {
|
||||||
|
return false, fmt.Errorf("cannot init nil indexer")
|
||||||
|
}
|
||||||
|
if i.Client != nil {
|
||||||
|
return false, fmt.Errorf("indexer is already initialized")
|
||||||
|
}
|
||||||
|
|
||||||
|
client, err := i.initClient()
|
||||||
|
if err != nil {
|
||||||
|
return false, err
|
||||||
|
}
|
||||||
|
i.Client = client
|
||||||
|
|
||||||
|
exists, err := i.Client.IndexExists(i.VersionedIndexName()).Do(ctx)
|
||||||
|
if err != nil {
|
||||||
|
return false, err
|
||||||
|
}
|
||||||
|
if exists {
|
||||||
|
return true, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
if err := i.createIndex(ctx); err != nil {
|
||||||
|
return false, err
|
||||||
|
}
|
||||||
|
|
||||||
|
return exists, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Ping checks if the indexer is available
|
||||||
|
func (i *Indexer) Ping(ctx context.Context) error {
|
||||||
|
if i == nil {
|
||||||
|
return fmt.Errorf("cannot ping nil indexer")
|
||||||
|
}
|
||||||
|
if i.Client == nil {
|
||||||
|
return fmt.Errorf("indexer is not initialized")
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := i.Client.ClusterHealth().Do(ctx)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
if resp.Status != "green" {
|
||||||
|
// see https://www.elastic.co/guide/en/elasticsearch/reference/current/cluster-health.html
|
||||||
|
return fmt.Errorf("status of elasticsearch cluster is %s", resp.Status)
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Close closes the indexer
|
||||||
|
func (i *Indexer) Close() {
|
||||||
|
if i == nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
i.Client = nil
|
||||||
|
}
|
68
modules/indexer/internal/elasticsearch/util.go
Normal file
68
modules/indexer/internal/elasticsearch/util.go
Normal file
|
@ -0,0 +1,68 @@
|
||||||
|
// Copyright 2023 The Gitea Authors. All rights reserved.
|
||||||
|
// SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
|
package elasticsearch
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"fmt"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"code.gitea.io/gitea/modules/log"
|
||||||
|
|
||||||
|
"github.com/olivere/elastic/v7"
|
||||||
|
)
|
||||||
|
|
||||||
|
// VersionedIndexName returns the full index name with version
|
||||||
|
func (i *Indexer) VersionedIndexName() string {
|
||||||
|
return versionedIndexName(i.indexName, i.version)
|
||||||
|
}
|
||||||
|
|
||||||
|
func versionedIndexName(indexName string, version int) string {
|
||||||
|
if version == 0 {
|
||||||
|
// Old index name without version
|
||||||
|
return indexName
|
||||||
|
}
|
||||||
|
return fmt.Sprintf("%s.v%d", indexName, version)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (i *Indexer) createIndex(ctx context.Context) error {
|
||||||
|
createIndex, err := i.Client.CreateIndex(i.VersionedIndexName()).BodyString(i.mapping).Do(ctx)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
if !createIndex.Acknowledged {
|
||||||
|
return fmt.Errorf("create index %s with %s failed", i.VersionedIndexName(), i.mapping)
|
||||||
|
}
|
||||||
|
|
||||||
|
i.checkOldIndexes(ctx)
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (i *Indexer) initClient() (*elastic.Client, error) {
|
||||||
|
opts := []elastic.ClientOptionFunc{
|
||||||
|
elastic.SetURL(i.url),
|
||||||
|
elastic.SetSniff(false),
|
||||||
|
elastic.SetHealthcheckInterval(10 * time.Second),
|
||||||
|
elastic.SetGzip(false),
|
||||||
|
}
|
||||||
|
|
||||||
|
logger := log.GetLogger(log.DEFAULT)
|
||||||
|
|
||||||
|
opts = append(opts, elastic.SetTraceLog(&log.PrintfLogger{Logf: logger.Trace}))
|
||||||
|
opts = append(opts, elastic.SetInfoLog(&log.PrintfLogger{Logf: logger.Info}))
|
||||||
|
opts = append(opts, elastic.SetErrorLog(&log.PrintfLogger{Logf: logger.Error}))
|
||||||
|
|
||||||
|
return elastic.NewClient(opts...)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (i *Indexer) checkOldIndexes(ctx context.Context) {
|
||||||
|
for v := 0; v < i.version; v++ {
|
||||||
|
indexName := versionedIndexName(i.indexName, v)
|
||||||
|
exists, err := i.Client.IndexExists(indexName).Do(ctx)
|
||||||
|
if err == nil && exists {
|
||||||
|
log.Warn("Found older elasticsearch index named %q, Gitea will keep the old NOT DELETED. You can delete the old version after the upgrade succeed.", indexName)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
37
modules/indexer/internal/indexer.go
Normal file
37
modules/indexer/internal/indexer.go
Normal file
|
@ -0,0 +1,37 @@
|
||||||
|
// Copyright 2023 The Gitea Authors. All rights reserved.
|
||||||
|
// SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
|
package internal
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"fmt"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Indexer defines an basic indexer interface
|
||||||
|
type Indexer interface {
|
||||||
|
// Init initializes the indexer
|
||||||
|
// returns true if the index was opened/existed (with data populated), false if it was created/not-existed (with no data)
|
||||||
|
Init(ctx context.Context) (bool, error)
|
||||||
|
// Ping checks if the indexer is available
|
||||||
|
Ping(ctx context.Context) error
|
||||||
|
// Close closes the indexer
|
||||||
|
Close()
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewDummyIndexer returns a dummy indexer
|
||||||
|
func NewDummyIndexer() Indexer {
|
||||||
|
return &dummyIndexer{}
|
||||||
|
}
|
||||||
|
|
||||||
|
type dummyIndexer struct{}
|
||||||
|
|
||||||
|
func (d *dummyIndexer) Init(ctx context.Context) (bool, error) {
|
||||||
|
return false, fmt.Errorf("indexer is not ready")
|
||||||
|
}
|
||||||
|
|
||||||
|
func (d *dummyIndexer) Ping(ctx context.Context) error {
|
||||||
|
return fmt.Errorf("indexer is not ready")
|
||||||
|
}
|
||||||
|
|
||||||
|
func (d *dummyIndexer) Close() {}
|
92
modules/indexer/internal/meilisearch/indexer.go
Normal file
92
modules/indexer/internal/meilisearch/indexer.go
Normal file
|
@ -0,0 +1,92 @@
|
||||||
|
// Copyright 2023 The Gitea Authors. All rights reserved.
|
||||||
|
// SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
|
package meilisearch
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
"github.com/meilisearch/meilisearch-go"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Indexer represents a basic meilisearch indexer implementation
|
||||||
|
type Indexer struct {
|
||||||
|
Client *meilisearch.Client
|
||||||
|
|
||||||
|
url, apiKey string
|
||||||
|
indexName string
|
||||||
|
version int
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewIndexer(url, apiKey, indexName string, version int) *Indexer {
|
||||||
|
return &Indexer{
|
||||||
|
url: url,
|
||||||
|
apiKey: apiKey,
|
||||||
|
indexName: indexName,
|
||||||
|
version: version,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Init initializes the indexer
|
||||||
|
func (i *Indexer) Init(_ context.Context) (bool, error) {
|
||||||
|
if i == nil {
|
||||||
|
return false, fmt.Errorf("cannot init nil indexer")
|
||||||
|
}
|
||||||
|
|
||||||
|
if i.Client != nil {
|
||||||
|
return false, fmt.Errorf("indexer is already initialized")
|
||||||
|
}
|
||||||
|
|
||||||
|
i.Client = meilisearch.NewClient(meilisearch.ClientConfig{
|
||||||
|
Host: i.url,
|
||||||
|
APIKey: i.apiKey,
|
||||||
|
})
|
||||||
|
|
||||||
|
_, err := i.Client.GetIndex(i.VersionedIndexName())
|
||||||
|
if err == nil {
|
||||||
|
return true, nil
|
||||||
|
}
|
||||||
|
_, err = i.Client.CreateIndex(&meilisearch.IndexConfig{
|
||||||
|
Uid: i.VersionedIndexName(),
|
||||||
|
PrimaryKey: "id",
|
||||||
|
})
|
||||||
|
if err != nil {
|
||||||
|
return false, err
|
||||||
|
}
|
||||||
|
|
||||||
|
i.checkOldIndexes()
|
||||||
|
|
||||||
|
_, err = i.Client.Index(i.VersionedIndexName()).UpdateFilterableAttributes(&[]string{"repo_id"})
|
||||||
|
return false, err
|
||||||
|
}
|
||||||
|
|
||||||
|
// Ping checks if the indexer is available
|
||||||
|
func (i *Indexer) Ping(ctx context.Context) error {
|
||||||
|
if i == nil {
|
||||||
|
return fmt.Errorf("cannot ping nil indexer")
|
||||||
|
}
|
||||||
|
if i.Client == nil {
|
||||||
|
return fmt.Errorf("indexer is not initialized")
|
||||||
|
}
|
||||||
|
resp, err := i.Client.Health()
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
if resp.Status != "available" {
|
||||||
|
// See https://docs.meilisearch.com/reference/api/health.html#status
|
||||||
|
return fmt.Errorf("status of meilisearch is not available: %s", resp.Status)
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Close closes the indexer
|
||||||
|
func (i *Indexer) Close() {
|
||||||
|
if i == nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if i.Client == nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
i.Client = nil
|
||||||
|
}
|
38
modules/indexer/internal/meilisearch/util.go
Normal file
38
modules/indexer/internal/meilisearch/util.go
Normal file
|
@ -0,0 +1,38 @@
|
||||||
|
// Copyright 2023 The Gitea Authors. All rights reserved.
|
||||||
|
// SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
|
package meilisearch
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
"code.gitea.io/gitea/modules/log"
|
||||||
|
)
|
||||||
|
|
||||||
|
// VersionedIndexName returns the full index name with version
|
||||||
|
func (i *Indexer) VersionedIndexName() string {
|
||||||
|
return versionedIndexName(i.indexName, i.version)
|
||||||
|
}
|
||||||
|
|
||||||
|
func versionedIndexName(indexName string, version int) string {
|
||||||
|
if version == 0 {
|
||||||
|
// Old index name without version
|
||||||
|
return indexName
|
||||||
|
}
|
||||||
|
|
||||||
|
// The format of the index name is <index_name>_v<version>, not <index_name>.v<version> like elasticsearch.
|
||||||
|
// Because meilisearch does not support "." in index name, it should contain only alphanumeric characters, hyphens (-) and underscores (_).
|
||||||
|
// See https://www.meilisearch.com/docs/learn/core_concepts/indexes#index-uid
|
||||||
|
|
||||||
|
return fmt.Sprintf("%s_v%d", indexName, version)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (i *Indexer) checkOldIndexes() {
|
||||||
|
for v := 0; v < i.version; v++ {
|
||||||
|
indexName := versionedIndexName(i.indexName, v)
|
||||||
|
_, err := i.Client.GetIndex(indexName)
|
||||||
|
if err == nil {
|
||||||
|
log.Warn("Found older meilisearch index named %q, Gitea will keep the old NOT DELETED. You can delete the old version after the upgrade succeed.", indexName)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
|
@ -1,17 +1,14 @@
|
||||||
// Copyright 2018 The Gitea Authors. All rights reserved.
|
// Copyright 2018 The Gitea Authors. All rights reserved.
|
||||||
// SPDX-License-Identifier: MIT
|
// SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
package issues
|
package bleve
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"fmt"
|
|
||||||
"os"
|
|
||||||
"strconv"
|
|
||||||
|
|
||||||
gitea_bleve "code.gitea.io/gitea/modules/indexer/bleve"
|
indexer_internal "code.gitea.io/gitea/modules/indexer/internal"
|
||||||
"code.gitea.io/gitea/modules/log"
|
inner_bleve "code.gitea.io/gitea/modules/indexer/internal/bleve"
|
||||||
"code.gitea.io/gitea/modules/util"
|
"code.gitea.io/gitea/modules/indexer/issues/internal"
|
||||||
|
|
||||||
"github.com/blevesearch/bleve/v2"
|
"github.com/blevesearch/bleve/v2"
|
||||||
"github.com/blevesearch/bleve/v2/analysis/analyzer/custom"
|
"github.com/blevesearch/bleve/v2/analysis/analyzer/custom"
|
||||||
|
@ -19,10 +16,8 @@ import (
|
||||||
"github.com/blevesearch/bleve/v2/analysis/token/lowercase"
|
"github.com/blevesearch/bleve/v2/analysis/token/lowercase"
|
||||||
"github.com/blevesearch/bleve/v2/analysis/token/unicodenorm"
|
"github.com/blevesearch/bleve/v2/analysis/token/unicodenorm"
|
||||||
"github.com/blevesearch/bleve/v2/analysis/tokenizer/unicode"
|
"github.com/blevesearch/bleve/v2/analysis/tokenizer/unicode"
|
||||||
"github.com/blevesearch/bleve/v2/index/upsidedown"
|
|
||||||
"github.com/blevesearch/bleve/v2/mapping"
|
"github.com/blevesearch/bleve/v2/mapping"
|
||||||
"github.com/blevesearch/bleve/v2/search/query"
|
"github.com/blevesearch/bleve/v2/search/query"
|
||||||
"github.com/ethantkoenig/rupture"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
|
@ -31,20 +26,6 @@ const (
|
||||||
issueIndexerLatestVersion = 2
|
issueIndexerLatestVersion = 2
|
||||||
)
|
)
|
||||||
|
|
||||||
// indexerID a bleve-compatible unique identifier for an integer id
|
|
||||||
func indexerID(id int64) string {
|
|
||||||
return strconv.FormatInt(id, 36)
|
|
||||||
}
|
|
||||||
|
|
||||||
// idOfIndexerID the integer id associated with an indexer id
|
|
||||||
func idOfIndexerID(indexerID string) (int64, error) {
|
|
||||||
id, err := strconv.ParseInt(indexerID, 36, 64)
|
|
||||||
if err != nil {
|
|
||||||
return 0, fmt.Errorf("Unexpected indexer ID %s: %w", indexerID, err)
|
|
||||||
}
|
|
||||||
return id, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// numericEqualityQuery a numeric equality query for the given value and field
|
// numericEqualityQuery a numeric equality query for the given value and field
|
||||||
func numericEqualityQuery(value int64, field string) *query.NumericRangeQuery {
|
func numericEqualityQuery(value int64, field string) *query.NumericRangeQuery {
|
||||||
f := float64(value)
|
f := float64(value)
|
||||||
|
@ -72,49 +53,16 @@ func addUnicodeNormalizeTokenFilter(m *mapping.IndexMappingImpl) error {
|
||||||
|
|
||||||
const maxBatchSize = 16
|
const maxBatchSize = 16
|
||||||
|
|
||||||
// openIndexer open the index at the specified path, checking for metadata
|
// IndexerData an update to the issue indexer
|
||||||
// updates and bleve version updates. If index needs to be created (or
|
type IndexerData internal.IndexerData
|
||||||
// re-created), returns (nil, nil)
|
|
||||||
func openIndexer(path string, latestVersion int) (bleve.Index, error) {
|
|
||||||
_, err := os.Stat(path)
|
|
||||||
if err != nil && os.IsNotExist(err) {
|
|
||||||
return nil, nil
|
|
||||||
} else if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
|
|
||||||
metadata, err := rupture.ReadIndexMetadata(path)
|
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
if metadata.Version < latestVersion {
|
|
||||||
// the indexer is using a previous version, so we should delete it and
|
|
||||||
// re-populate
|
|
||||||
return nil, util.RemoveAll(path)
|
|
||||||
}
|
|
||||||
|
|
||||||
index, err := bleve.Open(path)
|
|
||||||
if err != nil && err == upsidedown.IncompatibleVersion {
|
|
||||||
// the indexer was built with a previous version of bleve, so we should
|
|
||||||
// delete it and re-populate
|
|
||||||
return nil, util.RemoveAll(path)
|
|
||||||
} else if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
|
|
||||||
return index, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// BleveIndexerData an update to the issue indexer
|
|
||||||
type BleveIndexerData IndexerData
|
|
||||||
|
|
||||||
// Type returns the document type, for bleve's mapping.Classifier interface.
|
// Type returns the document type, for bleve's mapping.Classifier interface.
|
||||||
func (i *BleveIndexerData) Type() string {
|
func (i *IndexerData) Type() string {
|
||||||
return issueIndexerDocType
|
return issueIndexerDocType
|
||||||
}
|
}
|
||||||
|
|
||||||
// createIssueIndexer create an issue indexer if one does not already exist
|
// generateIssueIndexMapping generates the bleve index mapping for issues
|
||||||
func createIssueIndexer(path string, latestVersion int) (bleve.Index, error) {
|
func generateIssueIndexMapping() (mapping.IndexMapping, error) {
|
||||||
mapping := bleve.NewIndexMapping()
|
mapping := bleve.NewIndexMapping()
|
||||||
docMapping := bleve.NewDocumentMapping()
|
docMapping := bleve.NewDocumentMapping()
|
||||||
|
|
||||||
|
@ -144,68 +92,31 @@ func createIssueIndexer(path string, latestVersion int) (bleve.Index, error) {
|
||||||
mapping.AddDocumentMapping(issueIndexerDocType, docMapping)
|
mapping.AddDocumentMapping(issueIndexerDocType, docMapping)
|
||||||
mapping.AddDocumentMapping("_all", bleve.NewDocumentDisabledMapping())
|
mapping.AddDocumentMapping("_all", bleve.NewDocumentDisabledMapping())
|
||||||
|
|
||||||
index, err := bleve.New(path, mapping)
|
return mapping, nil
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
|
|
||||||
if err = rupture.WriteIndexMetadata(path, &rupture.IndexMetadata{
|
|
||||||
Version: latestVersion,
|
|
||||||
}); err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
return index, nil
|
|
||||||
}
|
}
|
||||||
|
|
||||||
var _ Indexer = &BleveIndexer{}
|
var _ internal.Indexer = &Indexer{}
|
||||||
|
|
||||||
// BleveIndexer implements Indexer interface
|
// Indexer implements Indexer interface
|
||||||
type BleveIndexer struct {
|
type Indexer struct {
|
||||||
indexDir string
|
inner *inner_bleve.Indexer
|
||||||
indexer bleve.Index
|
indexer_internal.Indexer // do not composite inner_bleve.Indexer directly to avoid exposing too much
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewBleveIndexer creates a new bleve local indexer
|
// NewIndexer creates a new bleve local indexer
|
||||||
func NewBleveIndexer(indexDir string) *BleveIndexer {
|
func NewIndexer(indexDir string) *Indexer {
|
||||||
return &BleveIndexer{
|
inner := inner_bleve.NewIndexer(indexDir, issueIndexerLatestVersion, generateIssueIndexMapping)
|
||||||
indexDir: indexDir,
|
return &Indexer{
|
||||||
}
|
Indexer: inner,
|
||||||
}
|
inner: inner,
|
||||||
|
|
||||||
// Init will initialize the indexer
|
|
||||||
func (b *BleveIndexer) Init() (bool, error) {
|
|
||||||
var err error
|
|
||||||
b.indexer, err = openIndexer(b.indexDir, issueIndexerLatestVersion)
|
|
||||||
if err != nil {
|
|
||||||
return false, err
|
|
||||||
}
|
|
||||||
if b.indexer != nil {
|
|
||||||
return true, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
b.indexer, err = createIssueIndexer(b.indexDir, issueIndexerLatestVersion)
|
|
||||||
return false, err
|
|
||||||
}
|
|
||||||
|
|
||||||
// Ping does nothing
|
|
||||||
func (b *BleveIndexer) Ping() bool {
|
|
||||||
return true
|
|
||||||
}
|
|
||||||
|
|
||||||
// Close will close the bleve indexer
|
|
||||||
func (b *BleveIndexer) Close() {
|
|
||||||
if b.indexer != nil {
|
|
||||||
if err := b.indexer.Close(); err != nil {
|
|
||||||
log.Error("Error whilst closing indexer: %v", err)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Index will save the index data
|
// Index will save the index data
|
||||||
func (b *BleveIndexer) Index(issues []*IndexerData) error {
|
func (b *Indexer) Index(_ context.Context, issues []*internal.IndexerData) error {
|
||||||
batch := gitea_bleve.NewFlushingBatch(b.indexer, maxBatchSize)
|
batch := inner_bleve.NewFlushingBatch(b.inner.Indexer, maxBatchSize)
|
||||||
for _, issue := range issues {
|
for _, issue := range issues {
|
||||||
if err := batch.Index(indexerID(issue.ID), struct {
|
if err := batch.Index(indexer_internal.Base36(issue.ID), struct {
|
||||||
RepoID int64
|
RepoID int64
|
||||||
Title string
|
Title string
|
||||||
Content string
|
Content string
|
||||||
|
@ -223,10 +134,10 @@ func (b *BleveIndexer) Index(issues []*IndexerData) error {
|
||||||
}
|
}
|
||||||
|
|
||||||
// Delete deletes indexes by ids
|
// Delete deletes indexes by ids
|
||||||
func (b *BleveIndexer) Delete(ids ...int64) error {
|
func (b *Indexer) Delete(_ context.Context, ids ...int64) error {
|
||||||
batch := gitea_bleve.NewFlushingBatch(b.indexer, maxBatchSize)
|
batch := inner_bleve.NewFlushingBatch(b.inner.Indexer, maxBatchSize)
|
||||||
for _, id := range ids {
|
for _, id := range ids {
|
||||||
if err := batch.Delete(indexerID(id)); err != nil {
|
if err := batch.Delete(indexer_internal.Base36(id)); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -235,7 +146,7 @@ func (b *BleveIndexer) Delete(ids ...int64) error {
|
||||||
|
|
||||||
// Search searches for issues by given conditions.
|
// Search searches for issues by given conditions.
|
||||||
// Returns the matching issue IDs
|
// Returns the matching issue IDs
|
||||||
func (b *BleveIndexer) Search(ctx context.Context, keyword string, repoIDs []int64, limit, start int) (*SearchResult, error) {
|
func (b *Indexer) Search(ctx context.Context, keyword string, repoIDs []int64, limit, start int) (*internal.SearchResult, error) {
|
||||||
var repoQueriesP []*query.NumericRangeQuery
|
var repoQueriesP []*query.NumericRangeQuery
|
||||||
for _, repoID := range repoIDs {
|
for _, repoID := range repoIDs {
|
||||||
repoQueriesP = append(repoQueriesP, numericEqualityQuery(repoID, "RepoID"))
|
repoQueriesP = append(repoQueriesP, numericEqualityQuery(repoID, "RepoID"))
|
||||||
|
@ -255,20 +166,20 @@ func (b *BleveIndexer) Search(ctx context.Context, keyword string, repoIDs []int
|
||||||
search := bleve.NewSearchRequestOptions(indexerQuery, limit, start, false)
|
search := bleve.NewSearchRequestOptions(indexerQuery, limit, start, false)
|
||||||
search.SortBy([]string{"-_score"})
|
search.SortBy([]string{"-_score"})
|
||||||
|
|
||||||
result, err := b.indexer.SearchInContext(ctx, search)
|
result, err := b.inner.Indexer.SearchInContext(ctx, search)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
ret := SearchResult{
|
ret := internal.SearchResult{
|
||||||
Hits: make([]Match, 0, len(result.Hits)),
|
Hits: make([]internal.Match, 0, len(result.Hits)),
|
||||||
}
|
}
|
||||||
for _, hit := range result.Hits {
|
for _, hit := range result.Hits {
|
||||||
id, err := idOfIndexerID(hit.ID)
|
id, err := indexer_internal.ParseBase36(hit.ID)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
ret.Hits = append(ret.Hits, Match{
|
ret.Hits = append(ret.Hits, internal.Match{
|
||||||
ID: id,
|
ID: id,
|
||||||
})
|
})
|
||||||
}
|
}
|
|
@ -1,26 +1,28 @@
|
||||||
// Copyright 2018 The Gitea Authors. All rights reserved.
|
// Copyright 2018 The Gitea Authors. All rights reserved.
|
||||||
// SPDX-License-Identifier: MIT
|
// SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
package issues
|
package bleve
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
|
"code.gitea.io/gitea/modules/indexer/issues/internal"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestBleveIndexAndSearch(t *testing.T) {
|
func TestBleveIndexAndSearch(t *testing.T) {
|
||||||
dir := t.TempDir()
|
dir := t.TempDir()
|
||||||
indexer := NewBleveIndexer(dir)
|
indexer := NewIndexer(dir)
|
||||||
defer indexer.Close()
|
defer indexer.Close()
|
||||||
|
|
||||||
if _, err := indexer.Init(); err != nil {
|
if _, err := indexer.Init(context.Background()); err != nil {
|
||||||
assert.Fail(t, "Unable to initialize bleve indexer: %v", err)
|
assert.Fail(t, "Unable to initialize bleve indexer: %v", err)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
err := indexer.Index([]*IndexerData{
|
err := indexer.Index(context.Background(), []*internal.IndexerData{
|
||||||
{
|
{
|
||||||
ID: 1,
|
ID: 1,
|
||||||
RepoID: 2,
|
RepoID: 2,
|
|
@ -1,56 +0,0 @@
|
||||||
// Copyright 2019 The Gitea Authors. All rights reserved.
|
|
||||||
// SPDX-License-Identifier: MIT
|
|
||||||
|
|
||||||
package issues
|
|
||||||
|
|
||||||
import (
|
|
||||||
"context"
|
|
||||||
|
|
||||||
"code.gitea.io/gitea/models/db"
|
|
||||||
issues_model "code.gitea.io/gitea/models/issues"
|
|
||||||
)
|
|
||||||
|
|
||||||
// DBIndexer implements Indexer interface to use database's like search
|
|
||||||
type DBIndexer struct{}
|
|
||||||
|
|
||||||
// Init dummy function
|
|
||||||
func (i *DBIndexer) Init() (bool, error) {
|
|
||||||
return false, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Ping checks if database is available
|
|
||||||
func (i *DBIndexer) Ping() bool {
|
|
||||||
return db.GetEngine(db.DefaultContext).Ping() != nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Index dummy function
|
|
||||||
func (i *DBIndexer) Index(issue []*IndexerData) error {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Delete dummy function
|
|
||||||
func (i *DBIndexer) Delete(ids ...int64) error {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Close dummy function
|
|
||||||
func (i *DBIndexer) Close() {
|
|
||||||
}
|
|
||||||
|
|
||||||
// Search dummy function
|
|
||||||
func (i *DBIndexer) Search(ctx context.Context, kw string, repoIDs []int64, limit, start int) (*SearchResult, error) {
|
|
||||||
total, ids, err := issues_model.SearchIssueIDsByKeyword(ctx, kw, repoIDs, limit, start)
|
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
result := SearchResult{
|
|
||||||
Total: total,
|
|
||||||
Hits: make([]Match, 0, limit),
|
|
||||||
}
|
|
||||||
for _, id := range ids {
|
|
||||||
result.Hits = append(result.Hits, Match{
|
|
||||||
ID: id,
|
|
||||||
})
|
|
||||||
}
|
|
||||||
return &result, nil
|
|
||||||
}
|
|
54
modules/indexer/issues/db/db.go
Normal file
54
modules/indexer/issues/db/db.go
Normal file
|
@ -0,0 +1,54 @@
|
||||||
|
// Copyright 2019 The Gitea Authors. All rights reserved.
|
||||||
|
// SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
|
package db
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
|
||||||
|
issues_model "code.gitea.io/gitea/models/issues"
|
||||||
|
indexer_internal "code.gitea.io/gitea/modules/indexer/internal"
|
||||||
|
inner_db "code.gitea.io/gitea/modules/indexer/internal/db"
|
||||||
|
"code.gitea.io/gitea/modules/indexer/issues/internal"
|
||||||
|
)
|
||||||
|
|
||||||
|
var _ internal.Indexer = &Indexer{}
|
||||||
|
|
||||||
|
// Indexer implements Indexer interface to use database's like search
|
||||||
|
type Indexer struct {
|
||||||
|
indexer_internal.Indexer
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewIndexer() *Indexer {
|
||||||
|
return &Indexer{
|
||||||
|
Indexer: &inner_db.Indexer{},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Index dummy function
|
||||||
|
func (i *Indexer) Index(_ context.Context, _ []*internal.IndexerData) error {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Delete dummy function
|
||||||
|
func (i *Indexer) Delete(_ context.Context, _ ...int64) error {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Search searches for issues
|
||||||
|
func (i *Indexer) Search(ctx context.Context, kw string, repoIDs []int64, limit, start int) (*internal.SearchResult, error) {
|
||||||
|
total, ids, err := issues_model.SearchIssueIDsByKeyword(ctx, kw, repoIDs, limit, start)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
result := internal.SearchResult{
|
||||||
|
Total: total,
|
||||||
|
Hits: make([]internal.Match, 0, limit),
|
||||||
|
}
|
||||||
|
for _, id := range ids {
|
||||||
|
result.Hits = append(result.Hits, internal.Match{
|
||||||
|
ID: id,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
return &result, nil
|
||||||
|
}
|
|
@ -1,287 +0,0 @@
|
||||||
// Copyright 2019 The Gitea Authors. All rights reserved.
|
|
||||||
// SPDX-License-Identifier: MIT
|
|
||||||
|
|
||||||
package issues
|
|
||||||
|
|
||||||
import (
|
|
||||||
"context"
|
|
||||||
"errors"
|
|
||||||
"fmt"
|
|
||||||
"net"
|
|
||||||
"strconv"
|
|
||||||
"sync"
|
|
||||||
"time"
|
|
||||||
|
|
||||||
"code.gitea.io/gitea/modules/graceful"
|
|
||||||
"code.gitea.io/gitea/modules/log"
|
|
||||||
|
|
||||||
"github.com/olivere/elastic/v7"
|
|
||||||
)
|
|
||||||
|
|
||||||
var _ Indexer = &ElasticSearchIndexer{}
|
|
||||||
|
|
||||||
// ElasticSearchIndexer implements Indexer interface
|
|
||||||
type ElasticSearchIndexer struct {
|
|
||||||
client *elastic.Client
|
|
||||||
indexerName string
|
|
||||||
available bool
|
|
||||||
stopTimer chan struct{}
|
|
||||||
lock sync.RWMutex
|
|
||||||
}
|
|
||||||
|
|
||||||
// NewElasticSearchIndexer creates a new elasticsearch indexer
|
|
||||||
func NewElasticSearchIndexer(url, indexerName string) (*ElasticSearchIndexer, error) {
|
|
||||||
opts := []elastic.ClientOptionFunc{
|
|
||||||
elastic.SetURL(url),
|
|
||||||
elastic.SetSniff(false),
|
|
||||||
elastic.SetHealthcheckInterval(10 * time.Second),
|
|
||||||
elastic.SetGzip(false),
|
|
||||||
}
|
|
||||||
|
|
||||||
logger := log.GetLogger(log.DEFAULT)
|
|
||||||
opts = append(opts, elastic.SetTraceLog(&log.PrintfLogger{Logf: logger.Trace}))
|
|
||||||
opts = append(opts, elastic.SetInfoLog(&log.PrintfLogger{Logf: logger.Info}))
|
|
||||||
opts = append(opts, elastic.SetErrorLog(&log.PrintfLogger{Logf: logger.Error}))
|
|
||||||
|
|
||||||
client, err := elastic.NewClient(opts...)
|
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
|
|
||||||
indexer := &ElasticSearchIndexer{
|
|
||||||
client: client,
|
|
||||||
indexerName: indexerName,
|
|
||||||
available: true,
|
|
||||||
stopTimer: make(chan struct{}),
|
|
||||||
}
|
|
||||||
|
|
||||||
ticker := time.NewTicker(10 * time.Second)
|
|
||||||
go func() {
|
|
||||||
for {
|
|
||||||
select {
|
|
||||||
case <-ticker.C:
|
|
||||||
indexer.checkAvailability()
|
|
||||||
case <-indexer.stopTimer:
|
|
||||||
ticker.Stop()
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}()
|
|
||||||
|
|
||||||
return indexer, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
const (
|
|
||||||
defaultMapping = `{
|
|
||||||
"mappings": {
|
|
||||||
"properties": {
|
|
||||||
"id": {
|
|
||||||
"type": "integer",
|
|
||||||
"index": true
|
|
||||||
},
|
|
||||||
"repo_id": {
|
|
||||||
"type": "integer",
|
|
||||||
"index": true
|
|
||||||
},
|
|
||||||
"title": {
|
|
||||||
"type": "text",
|
|
||||||
"index": true
|
|
||||||
},
|
|
||||||
"content": {
|
|
||||||
"type": "text",
|
|
||||||
"index": true
|
|
||||||
},
|
|
||||||
"comments": {
|
|
||||||
"type" : "text",
|
|
||||||
"index": true
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}`
|
|
||||||
)
|
|
||||||
|
|
||||||
// Init will initialize the indexer
|
|
||||||
func (b *ElasticSearchIndexer) Init() (bool, error) {
|
|
||||||
ctx := graceful.GetManager().HammerContext()
|
|
||||||
exists, err := b.client.IndexExists(b.indexerName).Do(ctx)
|
|
||||||
if err != nil {
|
|
||||||
return false, b.checkError(err)
|
|
||||||
}
|
|
||||||
|
|
||||||
if !exists {
|
|
||||||
mapping := defaultMapping
|
|
||||||
|
|
||||||
createIndex, err := b.client.CreateIndex(b.indexerName).BodyString(mapping).Do(ctx)
|
|
||||||
if err != nil {
|
|
||||||
return false, b.checkError(err)
|
|
||||||
}
|
|
||||||
if !createIndex.Acknowledged {
|
|
||||||
return false, errors.New("init failed")
|
|
||||||
}
|
|
||||||
|
|
||||||
return false, nil
|
|
||||||
}
|
|
||||||
return true, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Ping checks if elastic is available
|
|
||||||
func (b *ElasticSearchIndexer) Ping() bool {
|
|
||||||
b.lock.RLock()
|
|
||||||
defer b.lock.RUnlock()
|
|
||||||
return b.available
|
|
||||||
}
|
|
||||||
|
|
||||||
// Index will save the index data
|
|
||||||
func (b *ElasticSearchIndexer) Index(issues []*IndexerData) error {
|
|
||||||
if len(issues) == 0 {
|
|
||||||
return nil
|
|
||||||
} else if len(issues) == 1 {
|
|
||||||
issue := issues[0]
|
|
||||||
_, err := b.client.Index().
|
|
||||||
Index(b.indexerName).
|
|
||||||
Id(fmt.Sprintf("%d", issue.ID)).
|
|
||||||
BodyJson(map[string]interface{}{
|
|
||||||
"id": issue.ID,
|
|
||||||
"repo_id": issue.RepoID,
|
|
||||||
"title": issue.Title,
|
|
||||||
"content": issue.Content,
|
|
||||||
"comments": issue.Comments,
|
|
||||||
}).
|
|
||||||
Do(graceful.GetManager().HammerContext())
|
|
||||||
return b.checkError(err)
|
|
||||||
}
|
|
||||||
|
|
||||||
reqs := make([]elastic.BulkableRequest, 0)
|
|
||||||
for _, issue := range issues {
|
|
||||||
reqs = append(reqs,
|
|
||||||
elastic.NewBulkIndexRequest().
|
|
||||||
Index(b.indexerName).
|
|
||||||
Id(fmt.Sprintf("%d", issue.ID)).
|
|
||||||
Doc(map[string]interface{}{
|
|
||||||
"id": issue.ID,
|
|
||||||
"repo_id": issue.RepoID,
|
|
||||||
"title": issue.Title,
|
|
||||||
"content": issue.Content,
|
|
||||||
"comments": issue.Comments,
|
|
||||||
}),
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
_, err := b.client.Bulk().
|
|
||||||
Index(b.indexerName).
|
|
||||||
Add(reqs...).
|
|
||||||
Do(graceful.GetManager().HammerContext())
|
|
||||||
return b.checkError(err)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Delete deletes indexes by ids
|
|
||||||
func (b *ElasticSearchIndexer) Delete(ids ...int64) error {
|
|
||||||
if len(ids) == 0 {
|
|
||||||
return nil
|
|
||||||
} else if len(ids) == 1 {
|
|
||||||
_, err := b.client.Delete().
|
|
||||||
Index(b.indexerName).
|
|
||||||
Id(fmt.Sprintf("%d", ids[0])).
|
|
||||||
Do(graceful.GetManager().HammerContext())
|
|
||||||
return b.checkError(err)
|
|
||||||
}
|
|
||||||
|
|
||||||
reqs := make([]elastic.BulkableRequest, 0)
|
|
||||||
for _, id := range ids {
|
|
||||||
reqs = append(reqs,
|
|
||||||
elastic.NewBulkDeleteRequest().
|
|
||||||
Index(b.indexerName).
|
|
||||||
Id(fmt.Sprintf("%d", id)),
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
_, err := b.client.Bulk().
|
|
||||||
Index(b.indexerName).
|
|
||||||
Add(reqs...).
|
|
||||||
Do(graceful.GetManager().HammerContext())
|
|
||||||
return b.checkError(err)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Search searches for issues by given conditions.
|
|
||||||
// Returns the matching issue IDs
|
|
||||||
func (b *ElasticSearchIndexer) Search(ctx context.Context, keyword string, repoIDs []int64, limit, start int) (*SearchResult, error) {
|
|
||||||
kwQuery := elastic.NewMultiMatchQuery(keyword, "title", "content", "comments")
|
|
||||||
query := elastic.NewBoolQuery()
|
|
||||||
query = query.Must(kwQuery)
|
|
||||||
if len(repoIDs) > 0 {
|
|
||||||
repoStrs := make([]interface{}, 0, len(repoIDs))
|
|
||||||
for _, repoID := range repoIDs {
|
|
||||||
repoStrs = append(repoStrs, repoID)
|
|
||||||
}
|
|
||||||
repoQuery := elastic.NewTermsQuery("repo_id", repoStrs...)
|
|
||||||
query = query.Must(repoQuery)
|
|
||||||
}
|
|
||||||
searchResult, err := b.client.Search().
|
|
||||||
Index(b.indexerName).
|
|
||||||
Query(query).
|
|
||||||
Sort("_score", false).
|
|
||||||
From(start).Size(limit).
|
|
||||||
Do(ctx)
|
|
||||||
if err != nil {
|
|
||||||
return nil, b.checkError(err)
|
|
||||||
}
|
|
||||||
|
|
||||||
hits := make([]Match, 0, limit)
|
|
||||||
for _, hit := range searchResult.Hits.Hits {
|
|
||||||
id, _ := strconv.ParseInt(hit.Id, 10, 64)
|
|
||||||
hits = append(hits, Match{
|
|
||||||
ID: id,
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
return &SearchResult{
|
|
||||||
Total: searchResult.TotalHits(),
|
|
||||||
Hits: hits,
|
|
||||||
}, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Close implements indexer
|
|
||||||
func (b *ElasticSearchIndexer) Close() {
|
|
||||||
select {
|
|
||||||
case <-b.stopTimer:
|
|
||||||
default:
|
|
||||||
close(b.stopTimer)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (b *ElasticSearchIndexer) checkError(err error) error {
|
|
||||||
var opErr *net.OpError
|
|
||||||
if !(elastic.IsConnErr(err) || (errors.As(err, &opErr) && (opErr.Op == "dial" || opErr.Op == "read"))) {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
b.setAvailability(false)
|
|
||||||
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
func (b *ElasticSearchIndexer) checkAvailability() {
|
|
||||||
if b.Ping() {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
// Request cluster state to check if elastic is available again
|
|
||||||
_, err := b.client.ClusterState().Do(graceful.GetManager().ShutdownContext())
|
|
||||||
if err != nil {
|
|
||||||
b.setAvailability(false)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
b.setAvailability(true)
|
|
||||||
}
|
|
||||||
|
|
||||||
func (b *ElasticSearchIndexer) setAvailability(available bool) {
|
|
||||||
b.lock.Lock()
|
|
||||||
defer b.lock.Unlock()
|
|
||||||
|
|
||||||
if b.available == available {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
b.available = available
|
|
||||||
}
|
|
177
modules/indexer/issues/elasticsearch/elasticsearch.go
Normal file
177
modules/indexer/issues/elasticsearch/elasticsearch.go
Normal file
|
@ -0,0 +1,177 @@
|
||||||
|
// Copyright 2019 The Gitea Authors. All rights reserved.
|
||||||
|
// SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
|
package elasticsearch
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"fmt"
|
||||||
|
"strconv"
|
||||||
|
|
||||||
|
"code.gitea.io/gitea/modules/graceful"
|
||||||
|
indexer_internal "code.gitea.io/gitea/modules/indexer/internal"
|
||||||
|
inner_elasticsearch "code.gitea.io/gitea/modules/indexer/internal/elasticsearch"
|
||||||
|
"code.gitea.io/gitea/modules/indexer/issues/internal"
|
||||||
|
|
||||||
|
"github.com/olivere/elastic/v7"
|
||||||
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
issueIndexerLatestVersion = 0
|
||||||
|
)
|
||||||
|
|
||||||
|
var _ internal.Indexer = &Indexer{}
|
||||||
|
|
||||||
|
// Indexer implements Indexer interface
|
||||||
|
type Indexer struct {
|
||||||
|
inner *inner_elasticsearch.Indexer
|
||||||
|
indexer_internal.Indexer // do not composite inner_elasticsearch.Indexer directly to avoid exposing too much
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewIndexer creates a new elasticsearch indexer
|
||||||
|
func NewIndexer(url, indexerName string) *Indexer {
|
||||||
|
inner := inner_elasticsearch.NewIndexer(url, indexerName, issueIndexerLatestVersion, defaultMapping)
|
||||||
|
indexer := &Indexer{
|
||||||
|
inner: inner,
|
||||||
|
Indexer: inner,
|
||||||
|
}
|
||||||
|
return indexer
|
||||||
|
}
|
||||||
|
|
||||||
|
const (
|
||||||
|
defaultMapping = `{
|
||||||
|
"mappings": {
|
||||||
|
"properties": {
|
||||||
|
"id": {
|
||||||
|
"type": "integer",
|
||||||
|
"index": true
|
||||||
|
},
|
||||||
|
"repo_id": {
|
||||||
|
"type": "integer",
|
||||||
|
"index": true
|
||||||
|
},
|
||||||
|
"title": {
|
||||||
|
"type": "text",
|
||||||
|
"index": true
|
||||||
|
},
|
||||||
|
"content": {
|
||||||
|
"type": "text",
|
||||||
|
"index": true
|
||||||
|
},
|
||||||
|
"comments": {
|
||||||
|
"type" : "text",
|
||||||
|
"index": true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}`
|
||||||
|
)
|
||||||
|
|
||||||
|
// Index will save the index data
|
||||||
|
func (b *Indexer) Index(ctx context.Context, issues []*internal.IndexerData) error {
|
||||||
|
if len(issues) == 0 {
|
||||||
|
return nil
|
||||||
|
} else if len(issues) == 1 {
|
||||||
|
issue := issues[0]
|
||||||
|
_, err := b.inner.Client.Index().
|
||||||
|
Index(b.inner.VersionedIndexName()).
|
||||||
|
Id(fmt.Sprintf("%d", issue.ID)).
|
||||||
|
BodyJson(map[string]interface{}{
|
||||||
|
"id": issue.ID,
|
||||||
|
"repo_id": issue.RepoID,
|
||||||
|
"title": issue.Title,
|
||||||
|
"content": issue.Content,
|
||||||
|
"comments": issue.Comments,
|
||||||
|
}).
|
||||||
|
Do(ctx)
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
reqs := make([]elastic.BulkableRequest, 0)
|
||||||
|
for _, issue := range issues {
|
||||||
|
reqs = append(reqs,
|
||||||
|
elastic.NewBulkIndexRequest().
|
||||||
|
Index(b.inner.VersionedIndexName()).
|
||||||
|
Id(fmt.Sprintf("%d", issue.ID)).
|
||||||
|
Doc(map[string]interface{}{
|
||||||
|
"id": issue.ID,
|
||||||
|
"repo_id": issue.RepoID,
|
||||||
|
"title": issue.Title,
|
||||||
|
"content": issue.Content,
|
||||||
|
"comments": issue.Comments,
|
||||||
|
}),
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
_, err := b.inner.Client.Bulk().
|
||||||
|
Index(b.inner.VersionedIndexName()).
|
||||||
|
Add(reqs...).
|
||||||
|
Do(graceful.GetManager().HammerContext())
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
// Delete deletes indexes by ids
|
||||||
|
func (b *Indexer) Delete(ctx context.Context, ids ...int64) error {
|
||||||
|
if len(ids) == 0 {
|
||||||
|
return nil
|
||||||
|
} else if len(ids) == 1 {
|
||||||
|
_, err := b.inner.Client.Delete().
|
||||||
|
Index(b.inner.VersionedIndexName()).
|
||||||
|
Id(fmt.Sprintf("%d", ids[0])).
|
||||||
|
Do(ctx)
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
reqs := make([]elastic.BulkableRequest, 0)
|
||||||
|
for _, id := range ids {
|
||||||
|
reqs = append(reqs,
|
||||||
|
elastic.NewBulkDeleteRequest().
|
||||||
|
Index(b.inner.VersionedIndexName()).
|
||||||
|
Id(fmt.Sprintf("%d", id)),
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
_, err := b.inner.Client.Bulk().
|
||||||
|
Index(b.inner.VersionedIndexName()).
|
||||||
|
Add(reqs...).
|
||||||
|
Do(graceful.GetManager().HammerContext())
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
// Search searches for issues by given conditions.
|
||||||
|
// Returns the matching issue IDs
|
||||||
|
func (b *Indexer) Search(ctx context.Context, keyword string, repoIDs []int64, limit, start int) (*internal.SearchResult, error) {
|
||||||
|
kwQuery := elastic.NewMultiMatchQuery(keyword, "title", "content", "comments")
|
||||||
|
query := elastic.NewBoolQuery()
|
||||||
|
query = query.Must(kwQuery)
|
||||||
|
if len(repoIDs) > 0 {
|
||||||
|
repoStrs := make([]interface{}, 0, len(repoIDs))
|
||||||
|
for _, repoID := range repoIDs {
|
||||||
|
repoStrs = append(repoStrs, repoID)
|
||||||
|
}
|
||||||
|
repoQuery := elastic.NewTermsQuery("repo_id", repoStrs...)
|
||||||
|
query = query.Must(repoQuery)
|
||||||
|
}
|
||||||
|
searchResult, err := b.inner.Client.Search().
|
||||||
|
Index(b.inner.VersionedIndexName()).
|
||||||
|
Query(query).
|
||||||
|
Sort("_score", false).
|
||||||
|
From(start).Size(limit).
|
||||||
|
Do(ctx)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
hits := make([]internal.Match, 0, limit)
|
||||||
|
for _, hit := range searchResult.Hits.Hits {
|
||||||
|
id, _ := strconv.ParseInt(hit.Id, 10, 64)
|
||||||
|
hits = append(hits, internal.Match{
|
||||||
|
ID: id,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
return &internal.SearchResult{
|
||||||
|
Total: searchResult.TotalHits(),
|
||||||
|
Hits: hits,
|
||||||
|
}, nil
|
||||||
|
}
|
|
@ -5,16 +5,20 @@ package issues
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"fmt"
|
|
||||||
"os"
|
"os"
|
||||||
"runtime/pprof"
|
"runtime/pprof"
|
||||||
"sync"
|
"sync/atomic"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"code.gitea.io/gitea/models/db"
|
db_model "code.gitea.io/gitea/models/db"
|
||||||
issues_model "code.gitea.io/gitea/models/issues"
|
issues_model "code.gitea.io/gitea/models/issues"
|
||||||
repo_model "code.gitea.io/gitea/models/repo"
|
repo_model "code.gitea.io/gitea/models/repo"
|
||||||
"code.gitea.io/gitea/modules/graceful"
|
"code.gitea.io/gitea/modules/graceful"
|
||||||
|
"code.gitea.io/gitea/modules/indexer/issues/bleve"
|
||||||
|
"code.gitea.io/gitea/modules/indexer/issues/db"
|
||||||
|
"code.gitea.io/gitea/modules/indexer/issues/elasticsearch"
|
||||||
|
"code.gitea.io/gitea/modules/indexer/issues/internal"
|
||||||
|
"code.gitea.io/gitea/modules/indexer/issues/meilisearch"
|
||||||
"code.gitea.io/gitea/modules/log"
|
"code.gitea.io/gitea/modules/log"
|
||||||
"code.gitea.io/gitea/modules/process"
|
"code.gitea.io/gitea/modules/process"
|
||||||
"code.gitea.io/gitea/modules/queue"
|
"code.gitea.io/gitea/modules/queue"
|
||||||
|
@ -22,81 +26,22 @@ import (
|
||||||
"code.gitea.io/gitea/modules/util"
|
"code.gitea.io/gitea/modules/util"
|
||||||
)
|
)
|
||||||
|
|
||||||
// IndexerData data stored in the issue indexer
|
|
||||||
type IndexerData struct {
|
|
||||||
ID int64 `json:"id"`
|
|
||||||
RepoID int64 `json:"repo_id"`
|
|
||||||
Title string `json:"title"`
|
|
||||||
Content string `json:"content"`
|
|
||||||
Comments []string `json:"comments"`
|
|
||||||
IsDelete bool `json:"is_delete"`
|
|
||||||
IDs []int64 `json:"ids"`
|
|
||||||
}
|
|
||||||
|
|
||||||
// Match represents on search result
|
|
||||||
type Match struct {
|
|
||||||
ID int64 `json:"id"`
|
|
||||||
Score float64 `json:"score"`
|
|
||||||
}
|
|
||||||
|
|
||||||
// SearchResult represents search results
|
|
||||||
type SearchResult struct {
|
|
||||||
Total int64
|
|
||||||
Hits []Match
|
|
||||||
}
|
|
||||||
|
|
||||||
// Indexer defines an interface to indexer issues contents
|
|
||||||
type Indexer interface {
|
|
||||||
Init() (bool, error)
|
|
||||||
Ping() bool
|
|
||||||
Index(issue []*IndexerData) error
|
|
||||||
Delete(ids ...int64) error
|
|
||||||
Search(ctx context.Context, kw string, repoIDs []int64, limit, start int) (*SearchResult, error)
|
|
||||||
Close()
|
|
||||||
}
|
|
||||||
|
|
||||||
type indexerHolder struct {
|
|
||||||
indexer Indexer
|
|
||||||
mutex sync.RWMutex
|
|
||||||
cond *sync.Cond
|
|
||||||
cancelled bool
|
|
||||||
}
|
|
||||||
|
|
||||||
func newIndexerHolder() *indexerHolder {
|
|
||||||
h := &indexerHolder{}
|
|
||||||
h.cond = sync.NewCond(h.mutex.RLocker())
|
|
||||||
return h
|
|
||||||
}
|
|
||||||
|
|
||||||
func (h *indexerHolder) cancel() {
|
|
||||||
h.mutex.Lock()
|
|
||||||
defer h.mutex.Unlock()
|
|
||||||
h.cancelled = true
|
|
||||||
h.cond.Broadcast()
|
|
||||||
}
|
|
||||||
|
|
||||||
func (h *indexerHolder) set(indexer Indexer) {
|
|
||||||
h.mutex.Lock()
|
|
||||||
defer h.mutex.Unlock()
|
|
||||||
h.indexer = indexer
|
|
||||||
h.cond.Broadcast()
|
|
||||||
}
|
|
||||||
|
|
||||||
func (h *indexerHolder) get() Indexer {
|
|
||||||
h.mutex.RLock()
|
|
||||||
defer h.mutex.RUnlock()
|
|
||||||
if h.indexer == nil && !h.cancelled {
|
|
||||||
h.cond.Wait()
|
|
||||||
}
|
|
||||||
return h.indexer
|
|
||||||
}
|
|
||||||
|
|
||||||
var (
|
var (
|
||||||
// issueIndexerQueue queue of issue ids to be updated
|
// issueIndexerQueue queue of issue ids to be updated
|
||||||
issueIndexerQueue *queue.WorkerPoolQueue[*IndexerData]
|
issueIndexerQueue *queue.WorkerPoolQueue[*internal.IndexerData]
|
||||||
holder = newIndexerHolder()
|
// globalIndexer is the global indexer, it cannot be nil.
|
||||||
|
// When the real indexer is not ready, it will be a dummy indexer which will return error to explain it's not ready.
|
||||||
|
// So it's always safe use it as *globalIndexer.Load() and call its methods.
|
||||||
|
globalIndexer atomic.Pointer[internal.Indexer]
|
||||||
|
dummyIndexer *internal.Indexer
|
||||||
)
|
)
|
||||||
|
|
||||||
|
func init() {
|
||||||
|
i := internal.NewDummyIndexer()
|
||||||
|
dummyIndexer = &i
|
||||||
|
globalIndexer.Store(dummyIndexer)
|
||||||
|
}
|
||||||
|
|
||||||
// InitIssueIndexer initialize issue indexer, syncReindex is true then reindex until
|
// InitIssueIndexer initialize issue indexer, syncReindex is true then reindex until
|
||||||
// all issue index done.
|
// all issue index done.
|
||||||
func InitIssueIndexer(syncReindex bool) {
|
func InitIssueIndexer(syncReindex bool) {
|
||||||
|
@ -107,33 +52,23 @@ func InitIssueIndexer(syncReindex bool) {
|
||||||
// Create the Queue
|
// Create the Queue
|
||||||
switch setting.Indexer.IssueType {
|
switch setting.Indexer.IssueType {
|
||||||
case "bleve", "elasticsearch", "meilisearch":
|
case "bleve", "elasticsearch", "meilisearch":
|
||||||
handler := func(items ...*IndexerData) (unhandled []*IndexerData) {
|
handler := func(items ...*internal.IndexerData) (unhandled []*internal.IndexerData) {
|
||||||
indexer := holder.get()
|
indexer := *globalIndexer.Load()
|
||||||
if indexer == nil {
|
toIndex := make([]*internal.IndexerData, 0, len(items))
|
||||||
log.Warn("Issue indexer handler: indexer is not ready, retry later.")
|
|
||||||
return items
|
|
||||||
}
|
|
||||||
toIndex := make([]*IndexerData, 0, len(items))
|
|
||||||
for _, indexerData := range items {
|
for _, indexerData := range items {
|
||||||
log.Trace("IndexerData Process: %d %v %t", indexerData.ID, indexerData.IDs, indexerData.IsDelete)
|
log.Trace("IndexerData Process: %d %v %t", indexerData.ID, indexerData.IDs, indexerData.IsDelete)
|
||||||
if indexerData.IsDelete {
|
if indexerData.IsDelete {
|
||||||
if err := indexer.Delete(indexerData.IDs...); err != nil {
|
if err := indexer.Delete(ctx, indexerData.IDs...); err != nil {
|
||||||
log.Error("Issue indexer handler: failed to from index: %v Error: %v", indexerData.IDs, err)
|
log.Error("Issue indexer handler: failed to from index: %v Error: %v", indexerData.IDs, err)
|
||||||
if !indexer.Ping() {
|
unhandled = append(unhandled, indexerData)
|
||||||
log.Error("Issue indexer handler: indexer is unavailable when deleting")
|
|
||||||
unhandled = append(unhandled, indexerData)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
toIndex = append(toIndex, indexerData)
|
toIndex = append(toIndex, indexerData)
|
||||||
}
|
}
|
||||||
if err := indexer.Index(toIndex); err != nil {
|
if err := indexer.Index(ctx, toIndex); err != nil {
|
||||||
log.Error("Error whilst indexing: %v Error: %v", toIndex, err)
|
log.Error("Error whilst indexing: %v Error: %v", toIndex, err)
|
||||||
if !indexer.Ping() {
|
unhandled = append(unhandled, toIndex...)
|
||||||
log.Error("Issue indexer handler: indexer is unavailable when indexing")
|
|
||||||
unhandled = append(unhandled, toIndex...)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
return unhandled
|
return unhandled
|
||||||
}
|
}
|
||||||
|
@ -144,7 +79,7 @@ func InitIssueIndexer(syncReindex bool) {
|
||||||
log.Fatal("Unable to create issue indexer queue")
|
log.Fatal("Unable to create issue indexer queue")
|
||||||
}
|
}
|
||||||
default:
|
default:
|
||||||
issueIndexerQueue = queue.CreateSimpleQueue[*IndexerData](ctx, "issue_indexer", nil)
|
issueIndexerQueue = queue.CreateSimpleQueue[*internal.IndexerData](ctx, "issue_indexer", nil)
|
||||||
}
|
}
|
||||||
|
|
||||||
graceful.GetManager().RunAtTerminate(finished)
|
graceful.GetManager().RunAtTerminate(finished)
|
||||||
|
@ -154,7 +89,11 @@ func InitIssueIndexer(syncReindex bool) {
|
||||||
pprof.SetGoroutineLabels(ctx)
|
pprof.SetGoroutineLabels(ctx)
|
||||||
start := time.Now()
|
start := time.Now()
|
||||||
log.Info("PID %d: Initializing Issue Indexer: %s", os.Getpid(), setting.Indexer.IssueType)
|
log.Info("PID %d: Initializing Issue Indexer: %s", os.Getpid(), setting.Indexer.IssueType)
|
||||||
var populate bool
|
var (
|
||||||
|
issueIndexer internal.Indexer
|
||||||
|
existed bool
|
||||||
|
err error
|
||||||
|
)
|
||||||
switch setting.Indexer.IssueType {
|
switch setting.Indexer.IssueType {
|
||||||
case "bleve":
|
case "bleve":
|
||||||
defer func() {
|
defer func() {
|
||||||
|
@ -162,62 +101,45 @@ func InitIssueIndexer(syncReindex bool) {
|
||||||
log.Error("PANIC whilst initializing issue indexer: %v\nStacktrace: %s", err, log.Stack(2))
|
log.Error("PANIC whilst initializing issue indexer: %v\nStacktrace: %s", err, log.Stack(2))
|
||||||
log.Error("The indexer files are likely corrupted and may need to be deleted")
|
log.Error("The indexer files are likely corrupted and may need to be deleted")
|
||||||
log.Error("You can completely remove the %q directory to make Gitea recreate the indexes", setting.Indexer.IssuePath)
|
log.Error("You can completely remove the %q directory to make Gitea recreate the indexes", setting.Indexer.IssuePath)
|
||||||
holder.cancel()
|
globalIndexer.Store(dummyIndexer)
|
||||||
log.Fatal("PID: %d Unable to initialize the Bleve Issue Indexer at path: %s Error: %v", os.Getpid(), setting.Indexer.IssuePath, err)
|
log.Fatal("PID: %d Unable to initialize the Bleve Issue Indexer at path: %s Error: %v", os.Getpid(), setting.Indexer.IssuePath, err)
|
||||||
}
|
}
|
||||||
}()
|
}()
|
||||||
issueIndexer := NewBleveIndexer(setting.Indexer.IssuePath)
|
issueIndexer = bleve.NewIndexer(setting.Indexer.IssuePath)
|
||||||
exist, err := issueIndexer.Init()
|
existed, err = issueIndexer.Init(ctx)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
holder.cancel()
|
|
||||||
log.Fatal("Unable to initialize Bleve Issue Indexer at path: %s Error: %v", setting.Indexer.IssuePath, err)
|
log.Fatal("Unable to initialize Bleve Issue Indexer at path: %s Error: %v", setting.Indexer.IssuePath, err)
|
||||||
}
|
}
|
||||||
populate = !exist
|
|
||||||
holder.set(issueIndexer)
|
|
||||||
graceful.GetManager().RunAtTerminate(func() {
|
|
||||||
log.Debug("Closing issue indexer")
|
|
||||||
issueIndexer := holder.get()
|
|
||||||
if issueIndexer != nil {
|
|
||||||
issueIndexer.Close()
|
|
||||||
}
|
|
||||||
log.Info("PID: %d Issue Indexer closed", os.Getpid())
|
|
||||||
})
|
|
||||||
log.Debug("Created Bleve Indexer")
|
|
||||||
case "elasticsearch":
|
case "elasticsearch":
|
||||||
issueIndexer, err := NewElasticSearchIndexer(setting.Indexer.IssueConnStr, setting.Indexer.IssueIndexerName)
|
issueIndexer = elasticsearch.NewIndexer(setting.Indexer.IssueConnStr, setting.Indexer.IssueIndexerName)
|
||||||
if err != nil {
|
existed, err = issueIndexer.Init(ctx)
|
||||||
log.Fatal("Unable to initialize Elastic Search Issue Indexer at connection: %s Error: %v", setting.Indexer.IssueConnStr, err)
|
|
||||||
}
|
|
||||||
exist, err := issueIndexer.Init()
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Fatal("Unable to issueIndexer.Init with connection %s Error: %v", setting.Indexer.IssueConnStr, err)
|
log.Fatal("Unable to issueIndexer.Init with connection %s Error: %v", setting.Indexer.IssueConnStr, err)
|
||||||
}
|
}
|
||||||
populate = !exist
|
|
||||||
holder.set(issueIndexer)
|
|
||||||
case "db":
|
case "db":
|
||||||
issueIndexer := &DBIndexer{}
|
issueIndexer = db.NewIndexer()
|
||||||
holder.set(issueIndexer)
|
|
||||||
case "meilisearch":
|
case "meilisearch":
|
||||||
issueIndexer, err := NewMeilisearchIndexer(setting.Indexer.IssueConnStr, setting.Indexer.IssueConnAuth, setting.Indexer.IssueIndexerName)
|
issueIndexer = meilisearch.NewIndexer(setting.Indexer.IssueConnStr, setting.Indexer.IssueConnAuth, setting.Indexer.IssueIndexerName)
|
||||||
if err != nil {
|
existed, err = issueIndexer.Init(ctx)
|
||||||
log.Fatal("Unable to initialize Meilisearch Issue Indexer at connection: %s Error: %v", setting.Indexer.IssueConnStr, err)
|
|
||||||
}
|
|
||||||
exist, err := issueIndexer.Init()
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Fatal("Unable to issueIndexer.Init with connection %s Error: %v", setting.Indexer.IssueConnStr, err)
|
log.Fatal("Unable to issueIndexer.Init with connection %s Error: %v", setting.Indexer.IssueConnStr, err)
|
||||||
}
|
}
|
||||||
populate = !exist
|
|
||||||
holder.set(issueIndexer)
|
|
||||||
default:
|
default:
|
||||||
holder.cancel()
|
|
||||||
log.Fatal("Unknown issue indexer type: %s", setting.Indexer.IssueType)
|
log.Fatal("Unknown issue indexer type: %s", setting.Indexer.IssueType)
|
||||||
}
|
}
|
||||||
|
globalIndexer.Store(&issueIndexer)
|
||||||
|
|
||||||
|
graceful.GetManager().RunAtTerminate(func() {
|
||||||
|
log.Debug("Closing issue indexer")
|
||||||
|
(*globalIndexer.Load()).Close()
|
||||||
|
log.Info("PID: %d Issue Indexer closed", os.Getpid())
|
||||||
|
})
|
||||||
|
|
||||||
// Start processing the queue
|
// Start processing the queue
|
||||||
go graceful.GetManager().RunWithCancel(issueIndexerQueue)
|
go graceful.GetManager().RunWithCancel(issueIndexerQueue)
|
||||||
|
|
||||||
// Populate the index
|
// Populate the index
|
||||||
if populate {
|
if !existed {
|
||||||
if syncReindex {
|
if syncReindex {
|
||||||
graceful.GetManager().RunWithShutdownContext(populateIssueIndexer)
|
graceful.GetManager().RunWithShutdownContext(populateIssueIndexer)
|
||||||
} else {
|
} else {
|
||||||
|
@ -266,8 +188,8 @@ func populateIssueIndexer(ctx context.Context) {
|
||||||
default:
|
default:
|
||||||
}
|
}
|
||||||
repos, _, err := repo_model.SearchRepositoryByName(ctx, &repo_model.SearchRepoOptions{
|
repos, _, err := repo_model.SearchRepositoryByName(ctx, &repo_model.SearchRepoOptions{
|
||||||
ListOptions: db.ListOptions{Page: page, PageSize: repo_model.RepositoryListDefaultPageSize},
|
ListOptions: db_model.ListOptions{Page: page, PageSize: repo_model.RepositoryListDefaultPageSize},
|
||||||
OrderBy: db.SearchOrderByID,
|
OrderBy: db_model.SearchOrderByID,
|
||||||
Private: true,
|
Private: true,
|
||||||
Collaborate: util.OptionalBoolFalse,
|
Collaborate: util.OptionalBoolFalse,
|
||||||
})
|
})
|
||||||
|
@ -320,7 +242,7 @@ func UpdateIssueIndexer(issue *issues_model.Issue) {
|
||||||
comments = append(comments, comment.Content)
|
comments = append(comments, comment.Content)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
indexerData := &IndexerData{
|
indexerData := &internal.IndexerData{
|
||||||
ID: issue.ID,
|
ID: issue.ID,
|
||||||
RepoID: issue.RepoID,
|
RepoID: issue.RepoID,
|
||||||
Title: issue.Title,
|
Title: issue.Title,
|
||||||
|
@ -345,7 +267,7 @@ func DeleteRepoIssueIndexer(ctx context.Context, repo *repo_model.Repository) {
|
||||||
if len(ids) == 0 {
|
if len(ids) == 0 {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
indexerData := &IndexerData{
|
indexerData := &internal.IndexerData{
|
||||||
IDs: ids,
|
IDs: ids,
|
||||||
IsDelete: true,
|
IsDelete: true,
|
||||||
}
|
}
|
||||||
|
@ -358,12 +280,7 @@ func DeleteRepoIssueIndexer(ctx context.Context, repo *repo_model.Repository) {
|
||||||
// WARNNING: You have to ensure user have permission to visit repoIDs' issues
|
// WARNNING: You have to ensure user have permission to visit repoIDs' issues
|
||||||
func SearchIssuesByKeyword(ctx context.Context, repoIDs []int64, keyword string) ([]int64, error) {
|
func SearchIssuesByKeyword(ctx context.Context, repoIDs []int64, keyword string) ([]int64, error) {
|
||||||
var issueIDs []int64
|
var issueIDs []int64
|
||||||
indexer := holder.get()
|
indexer := *globalIndexer.Load()
|
||||||
|
|
||||||
if indexer == nil {
|
|
||||||
log.Error("SearchIssuesByKeyword(): unable to get indexer!")
|
|
||||||
return nil, fmt.Errorf("unable to get issue indexer")
|
|
||||||
}
|
|
||||||
res, err := indexer.Search(ctx, keyword, repoIDs, 50, 0)
|
res, err := indexer.Search(ctx, keyword, repoIDs, 50, 0)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
|
@ -375,12 +292,6 @@ func SearchIssuesByKeyword(ctx context.Context, repoIDs []int64, keyword string)
|
||||||
}
|
}
|
||||||
|
|
||||||
// IsAvailable checks if issue indexer is available
|
// IsAvailable checks if issue indexer is available
|
||||||
func IsAvailable() bool {
|
func IsAvailable(ctx context.Context) bool {
|
||||||
indexer := holder.get()
|
return (*globalIndexer.Load()).Ping(ctx) == nil
|
||||||
if indexer == nil {
|
|
||||||
log.Error("IsAvailable(): unable to get indexer!")
|
|
||||||
return false
|
|
||||||
}
|
|
||||||
|
|
||||||
return indexer.Ping()
|
|
||||||
}
|
}
|
||||||
|
|
|
@ -11,6 +11,7 @@ import (
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"code.gitea.io/gitea/models/unittest"
|
"code.gitea.io/gitea/models/unittest"
|
||||||
|
"code.gitea.io/gitea/modules/indexer/issues/bleve"
|
||||||
"code.gitea.io/gitea/modules/setting"
|
"code.gitea.io/gitea/modules/setting"
|
||||||
|
|
||||||
_ "code.gitea.io/gitea/models"
|
_ "code.gitea.io/gitea/models"
|
||||||
|
@ -42,8 +43,7 @@ func TestBleveSearchIssues(t *testing.T) {
|
||||||
setting.LoadQueueSettings()
|
setting.LoadQueueSettings()
|
||||||
InitIssueIndexer(true)
|
InitIssueIndexer(true)
|
||||||
defer func() {
|
defer func() {
|
||||||
indexer := holder.get()
|
if bleveIndexer, ok := (*globalIndexer.Load()).(*bleve.Indexer); ok {
|
||||||
if bleveIndexer, ok := indexer.(*BleveIndexer); ok {
|
|
||||||
bleveIndexer.Close()
|
bleveIndexer.Close()
|
||||||
}
|
}
|
||||||
}()
|
}()
|
||||||
|
|
42
modules/indexer/issues/internal/indexer.go
Normal file
42
modules/indexer/issues/internal/indexer.go
Normal file
|
@ -0,0 +1,42 @@
|
||||||
|
// Copyright 2023 The Gitea Authors. All rights reserved.
|
||||||
|
// SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
|
package internal
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
"code.gitea.io/gitea/modules/indexer/internal"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Indexer defines an interface to indexer issues contents
|
||||||
|
type Indexer interface {
|
||||||
|
internal.Indexer
|
||||||
|
Index(ctx context.Context, issue []*IndexerData) error
|
||||||
|
Delete(ctx context.Context, ids ...int64) error
|
||||||
|
Search(ctx context.Context, kw string, repoIDs []int64, limit, start int) (*SearchResult, error)
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewDummyIndexer returns a dummy indexer
|
||||||
|
func NewDummyIndexer() Indexer {
|
||||||
|
return &dummyIndexer{
|
||||||
|
Indexer: internal.NewDummyIndexer(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
type dummyIndexer struct {
|
||||||
|
internal.Indexer
|
||||||
|
}
|
||||||
|
|
||||||
|
func (d *dummyIndexer) Index(ctx context.Context, issue []*IndexerData) error {
|
||||||
|
return fmt.Errorf("indexer is not ready")
|
||||||
|
}
|
||||||
|
|
||||||
|
func (d *dummyIndexer) Delete(ctx context.Context, ids ...int64) error {
|
||||||
|
return fmt.Errorf("indexer is not ready")
|
||||||
|
}
|
||||||
|
|
||||||
|
func (d *dummyIndexer) Search(ctx context.Context, kw string, repoIDs []int64, limit, start int) (*SearchResult, error) {
|
||||||
|
return nil, fmt.Errorf("indexer is not ready")
|
||||||
|
}
|
27
modules/indexer/issues/internal/model.go
Normal file
27
modules/indexer/issues/internal/model.go
Normal file
|
@ -0,0 +1,27 @@
|
||||||
|
// Copyright 2023 The Gitea Authors. All rights reserved.
|
||||||
|
// SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
|
package internal
|
||||||
|
|
||||||
|
// IndexerData data stored in the issue indexer
|
||||||
|
type IndexerData struct {
|
||||||
|
ID int64 `json:"id"`
|
||||||
|
RepoID int64 `json:"repo_id"`
|
||||||
|
Title string `json:"title"`
|
||||||
|
Content string `json:"content"`
|
||||||
|
Comments []string `json:"comments"`
|
||||||
|
IsDelete bool `json:"is_delete"`
|
||||||
|
IDs []int64 `json:"ids"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// Match represents on search result
|
||||||
|
type Match struct {
|
||||||
|
ID int64 `json:"id"`
|
||||||
|
Score float64 `json:"score"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// SearchResult represents search results
|
||||||
|
type SearchResult struct {
|
||||||
|
Total int64
|
||||||
|
Hits []Match
|
||||||
|
}
|
|
@ -1,173 +0,0 @@
|
||||||
// Copyright 2023 The Gitea Authors. All rights reserved.
|
|
||||||
// SPDX-License-Identifier: MIT
|
|
||||||
|
|
||||||
package issues
|
|
||||||
|
|
||||||
import (
|
|
||||||
"context"
|
|
||||||
"strconv"
|
|
||||||
"strings"
|
|
||||||
"sync"
|
|
||||||
"time"
|
|
||||||
|
|
||||||
"github.com/meilisearch/meilisearch-go"
|
|
||||||
)
|
|
||||||
|
|
||||||
var _ Indexer = &MeilisearchIndexer{}
|
|
||||||
|
|
||||||
// MeilisearchIndexer implements Indexer interface
|
|
||||||
type MeilisearchIndexer struct {
|
|
||||||
client *meilisearch.Client
|
|
||||||
indexerName string
|
|
||||||
available bool
|
|
||||||
stopTimer chan struct{}
|
|
||||||
lock sync.RWMutex
|
|
||||||
}
|
|
||||||
|
|
||||||
// MeilisearchIndexer creates a new meilisearch indexer
|
|
||||||
func NewMeilisearchIndexer(url, apiKey, indexerName string) (*MeilisearchIndexer, error) {
|
|
||||||
client := meilisearch.NewClient(meilisearch.ClientConfig{
|
|
||||||
Host: url,
|
|
||||||
APIKey: apiKey,
|
|
||||||
})
|
|
||||||
|
|
||||||
indexer := &MeilisearchIndexer{
|
|
||||||
client: client,
|
|
||||||
indexerName: indexerName,
|
|
||||||
available: true,
|
|
||||||
stopTimer: make(chan struct{}),
|
|
||||||
}
|
|
||||||
|
|
||||||
ticker := time.NewTicker(10 * time.Second)
|
|
||||||
go func() {
|
|
||||||
for {
|
|
||||||
select {
|
|
||||||
case <-ticker.C:
|
|
||||||
indexer.checkAvailability()
|
|
||||||
case <-indexer.stopTimer:
|
|
||||||
ticker.Stop()
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}()
|
|
||||||
|
|
||||||
return indexer, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Init will initialize the indexer
|
|
||||||
func (b *MeilisearchIndexer) Init() (bool, error) {
|
|
||||||
_, err := b.client.GetIndex(b.indexerName)
|
|
||||||
if err == nil {
|
|
||||||
return true, nil
|
|
||||||
}
|
|
||||||
_, err = b.client.CreateIndex(&meilisearch.IndexConfig{
|
|
||||||
Uid: b.indexerName,
|
|
||||||
PrimaryKey: "id",
|
|
||||||
})
|
|
||||||
if err != nil {
|
|
||||||
return false, b.checkError(err)
|
|
||||||
}
|
|
||||||
|
|
||||||
_, err = b.client.Index(b.indexerName).UpdateFilterableAttributes(&[]string{"repo_id"})
|
|
||||||
return false, b.checkError(err)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Ping checks if meilisearch is available
|
|
||||||
func (b *MeilisearchIndexer) Ping() bool {
|
|
||||||
b.lock.RLock()
|
|
||||||
defer b.lock.RUnlock()
|
|
||||||
return b.available
|
|
||||||
}
|
|
||||||
|
|
||||||
// Index will save the index data
|
|
||||||
func (b *MeilisearchIndexer) Index(issues []*IndexerData) error {
|
|
||||||
if len(issues) == 0 {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
for _, issue := range issues {
|
|
||||||
_, err := b.client.Index(b.indexerName).AddDocuments(issue)
|
|
||||||
if err != nil {
|
|
||||||
return b.checkError(err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
// TODO: bulk send index data
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Delete deletes indexes by ids
|
|
||||||
func (b *MeilisearchIndexer) Delete(ids ...int64) error {
|
|
||||||
if len(ids) == 0 {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, id := range ids {
|
|
||||||
_, err := b.client.Index(b.indexerName).DeleteDocument(strconv.FormatInt(id, 10))
|
|
||||||
if err != nil {
|
|
||||||
return b.checkError(err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
// TODO: bulk send deletes
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Search searches for issues by given conditions.
|
|
||||||
// Returns the matching issue IDs
|
|
||||||
func (b *MeilisearchIndexer) Search(ctx context.Context, keyword string, repoIDs []int64, limit, start int) (*SearchResult, error) {
|
|
||||||
repoFilters := make([]string, 0, len(repoIDs))
|
|
||||||
for _, repoID := range repoIDs {
|
|
||||||
repoFilters = append(repoFilters, "repo_id = "+strconv.FormatInt(repoID, 10))
|
|
||||||
}
|
|
||||||
filter := strings.Join(repoFilters, " OR ")
|
|
||||||
searchRes, err := b.client.Index(b.indexerName).Search(keyword, &meilisearch.SearchRequest{
|
|
||||||
Filter: filter,
|
|
||||||
Limit: int64(limit),
|
|
||||||
Offset: int64(start),
|
|
||||||
})
|
|
||||||
if err != nil {
|
|
||||||
return nil, b.checkError(err)
|
|
||||||
}
|
|
||||||
|
|
||||||
hits := make([]Match, 0, len(searchRes.Hits))
|
|
||||||
for _, hit := range searchRes.Hits {
|
|
||||||
hits = append(hits, Match{
|
|
||||||
ID: int64(hit.(map[string]interface{})["id"].(float64)),
|
|
||||||
})
|
|
||||||
}
|
|
||||||
return &SearchResult{
|
|
||||||
Total: searchRes.TotalHits,
|
|
||||||
Hits: hits,
|
|
||||||
}, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Close implements indexer
|
|
||||||
func (b *MeilisearchIndexer) Close() {
|
|
||||||
select {
|
|
||||||
case <-b.stopTimer:
|
|
||||||
default:
|
|
||||||
close(b.stopTimer)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (b *MeilisearchIndexer) checkError(err error) error {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
func (b *MeilisearchIndexer) checkAvailability() {
|
|
||||||
_, err := b.client.Health()
|
|
||||||
if err != nil {
|
|
||||||
b.setAvailability(false)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
b.setAvailability(true)
|
|
||||||
}
|
|
||||||
|
|
||||||
func (b *MeilisearchIndexer) setAvailability(available bool) {
|
|
||||||
b.lock.Lock()
|
|
||||||
defer b.lock.Unlock()
|
|
||||||
|
|
||||||
if b.available == available {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
b.available = available
|
|
||||||
}
|
|
98
modules/indexer/issues/meilisearch/meilisearch.go
Normal file
98
modules/indexer/issues/meilisearch/meilisearch.go
Normal file
|
@ -0,0 +1,98 @@
|
||||||
|
// Copyright 2023 The Gitea Authors. All rights reserved.
|
||||||
|
// SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
|
package meilisearch
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"strconv"
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
indexer_internal "code.gitea.io/gitea/modules/indexer/internal"
|
||||||
|
inner_meilisearch "code.gitea.io/gitea/modules/indexer/internal/meilisearch"
|
||||||
|
"code.gitea.io/gitea/modules/indexer/issues/internal"
|
||||||
|
|
||||||
|
"github.com/meilisearch/meilisearch-go"
|
||||||
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
issueIndexerLatestVersion = 0
|
||||||
|
)
|
||||||
|
|
||||||
|
var _ internal.Indexer = &Indexer{}
|
||||||
|
|
||||||
|
// Indexer implements Indexer interface
|
||||||
|
type Indexer struct {
|
||||||
|
inner *inner_meilisearch.Indexer
|
||||||
|
indexer_internal.Indexer // do not composite inner_meilisearch.Indexer directly to avoid exposing too much
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewIndexer creates a new meilisearch indexer
|
||||||
|
func NewIndexer(url, apiKey, indexerName string) *Indexer {
|
||||||
|
inner := inner_meilisearch.NewIndexer(url, apiKey, indexerName, issueIndexerLatestVersion)
|
||||||
|
indexer := &Indexer{
|
||||||
|
inner: inner,
|
||||||
|
Indexer: inner,
|
||||||
|
}
|
||||||
|
return indexer
|
||||||
|
}
|
||||||
|
|
||||||
|
// Index will save the index data
|
||||||
|
func (b *Indexer) Index(_ context.Context, issues []*internal.IndexerData) error {
|
||||||
|
if len(issues) == 0 {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
for _, issue := range issues {
|
||||||
|
_, err := b.inner.Client.Index(b.inner.VersionedIndexName()).AddDocuments(issue)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// TODO: bulk send index data
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Delete deletes indexes by ids
|
||||||
|
func (b *Indexer) Delete(_ context.Context, ids ...int64) error {
|
||||||
|
if len(ids) == 0 {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, id := range ids {
|
||||||
|
_, err := b.inner.Client.Index(b.inner.VersionedIndexName()).DeleteDocument(strconv.FormatInt(id, 10))
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// TODO: bulk send deletes
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Search searches for issues by given conditions.
|
||||||
|
// Returns the matching issue IDs
|
||||||
|
func (b *Indexer) Search(ctx context.Context, keyword string, repoIDs []int64, limit, start int) (*internal.SearchResult, error) {
|
||||||
|
repoFilters := make([]string, 0, len(repoIDs))
|
||||||
|
for _, repoID := range repoIDs {
|
||||||
|
repoFilters = append(repoFilters, "repo_id = "+strconv.FormatInt(repoID, 10))
|
||||||
|
}
|
||||||
|
filter := strings.Join(repoFilters, " OR ")
|
||||||
|
searchRes, err := b.inner.Client.Index(b.inner.VersionedIndexName()).Search(keyword, &meilisearch.SearchRequest{
|
||||||
|
Filter: filter,
|
||||||
|
Limit: int64(limit),
|
||||||
|
Offset: int64(start),
|
||||||
|
})
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
hits := make([]internal.Match, 0, len(searchRes.Hits))
|
||||||
|
for _, hit := range searchRes.Hits {
|
||||||
|
hits = append(hits, internal.Match{
|
||||||
|
ID: int64(hit.(map[string]interface{})["id"].(float64)),
|
||||||
|
})
|
||||||
|
}
|
||||||
|
return &internal.SearchResult{
|
||||||
|
Total: searchRes.TotalHits,
|
||||||
|
Hits: hits,
|
||||||
|
}, nil
|
||||||
|
}
|
|
@ -11,6 +11,7 @@ import (
|
||||||
)
|
)
|
||||||
|
|
||||||
// Indexer defines an interface to index repository stats
|
// Indexer defines an interface to index repository stats
|
||||||
|
// TODO: this indexer is quite different from the others, maybe this package should be moved out from module/indexer
|
||||||
type Indexer interface {
|
type Indexer interface {
|
||||||
Index(id int64) error
|
Index(id int64) error
|
||||||
Close()
|
Close()
|
||||||
|
|
|
@ -79,13 +79,13 @@ func Code(ctx *context.Context) {
|
||||||
if (len(repoIDs) > 0) || isAdmin {
|
if (len(repoIDs) > 0) || isAdmin {
|
||||||
total, searchResults, searchResultLanguages, err = code_indexer.PerformSearch(ctx, repoIDs, language, keyword, page, setting.UI.RepoSearchPagingNum, isMatch)
|
total, searchResults, searchResultLanguages, err = code_indexer.PerformSearch(ctx, repoIDs, language, keyword, page, setting.UI.RepoSearchPagingNum, isMatch)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
if code_indexer.IsAvailable() {
|
if code_indexer.IsAvailable(ctx) {
|
||||||
ctx.ServerError("SearchResults", err)
|
ctx.ServerError("SearchResults", err)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
ctx.Data["CodeIndexerUnavailable"] = true
|
ctx.Data["CodeIndexerUnavailable"] = true
|
||||||
} else {
|
} else {
|
||||||
ctx.Data["CodeIndexerUnavailable"] = !code_indexer.IsAvailable()
|
ctx.Data["CodeIndexerUnavailable"] = !code_indexer.IsAvailable(ctx)
|
||||||
}
|
}
|
||||||
|
|
||||||
loadRepoIDs := make([]int64, 0, len(searchResults))
|
loadRepoIDs := make([]int64, 0, len(searchResults))
|
||||||
|
|
|
@ -191,7 +191,7 @@ func issues(ctx *context.Context, milestoneID, projectID int64, isPullOption uti
|
||||||
if len(keyword) > 0 {
|
if len(keyword) > 0 {
|
||||||
issueIDs, err = issue_indexer.SearchIssuesByKeyword(ctx, []int64{repo.ID}, keyword)
|
issueIDs, err = issue_indexer.SearchIssuesByKeyword(ctx, []int64{repo.ID}, keyword)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
if issue_indexer.IsAvailable() {
|
if issue_indexer.IsAvailable(ctx) {
|
||||||
ctx.ServerError("issueIndexer.Search", err)
|
ctx.ServerError("issueIndexer.Search", err)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
|
@ -45,13 +45,13 @@ func Search(ctx *context.Context) {
|
||||||
total, searchResults, searchResultLanguages, err := code_indexer.PerformSearch(ctx, []int64{ctx.Repo.Repository.ID},
|
total, searchResults, searchResultLanguages, err := code_indexer.PerformSearch(ctx, []int64{ctx.Repo.Repository.ID},
|
||||||
language, keyword, page, setting.UI.RepoSearchPagingNum, isMatch)
|
language, keyword, page, setting.UI.RepoSearchPagingNum, isMatch)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
if code_indexer.IsAvailable() {
|
if code_indexer.IsAvailable(ctx) {
|
||||||
ctx.ServerError("SearchResults", err)
|
ctx.ServerError("SearchResults", err)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
ctx.Data["CodeIndexerUnavailable"] = true
|
ctx.Data["CodeIndexerUnavailable"] = true
|
||||||
} else {
|
} else {
|
||||||
ctx.Data["CodeIndexerUnavailable"] = !code_indexer.IsAvailable()
|
ctx.Data["CodeIndexerUnavailable"] = !code_indexer.IsAvailable(ctx)
|
||||||
}
|
}
|
||||||
|
|
||||||
ctx.Data["SourcePath"] = ctx.Repo.Repository.Link()
|
ctx.Data["SourcePath"] = ctx.Repo.Repository.Link()
|
||||||
|
|
|
@ -71,13 +71,13 @@ func CodeSearch(ctx *context.Context) {
|
||||||
if len(repoIDs) > 0 {
|
if len(repoIDs) > 0 {
|
||||||
total, searchResults, searchResultLanguages, err = code_indexer.PerformSearch(ctx, repoIDs, language, keyword, page, setting.UI.RepoSearchPagingNum, isMatch)
|
total, searchResults, searchResultLanguages, err = code_indexer.PerformSearch(ctx, repoIDs, language, keyword, page, setting.UI.RepoSearchPagingNum, isMatch)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
if code_indexer.IsAvailable() {
|
if code_indexer.IsAvailable(ctx) {
|
||||||
ctx.ServerError("SearchResults", err)
|
ctx.ServerError("SearchResults", err)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
ctx.Data["CodeIndexerUnavailable"] = true
|
ctx.Data["CodeIndexerUnavailable"] = true
|
||||||
} else {
|
} else {
|
||||||
ctx.Data["CodeIndexerUnavailable"] = !code_indexer.IsAvailable()
|
ctx.Data["CodeIndexerUnavailable"] = !code_indexer.IsAvailable(ctx)
|
||||||
}
|
}
|
||||||
|
|
||||||
loadRepoIDs := make([]int64, 0, len(searchResults))
|
loadRepoIDs := make([]int64, 0, len(searchResults))
|
||||||
|
|
Loading…
Reference in a new issue