mirror of
https://github.com/go-gitea/gitea.git
synced 2024-12-05 00:44:12 +08:00
900ac62251
This is a large and complex PR, so let me explain in detail its changes. First, I had to create new index mappings for Bleve and ElasticSerach as the current ones do not support search by filename. This requires Gitea to recreate the code search indexes (I do not know if this is a breaking change, but I feel it deserves a heads-up). I've used [this approach](https://www.elastic.co/guide/en/elasticsearch/reference/7.17/analysis-pathhierarchy-tokenizer.html) to model the filename index. It allows us to efficiently search for both the full path and the name of a file. Bleve, however, does not support this out-of-box, so I had to code a brand new [token filter](https://blevesearch.com/docs/Token-Filters/) to generate the search terms. I also did an overhaul in the `indexer_test.go` file. It now asserts the order of the expected results (this is important since matches based on the name of a file are more relevant than those based on its content). I've added new test scenarios that deal with searching by filename. They use a new repo included in the Gitea fixture. The screenshot below depicts how Gitea shows the search results. It shows results based on content in the same way as the current version does. In matches based on the filename, the first seven lines of the file contents are shown (BTW, this is how GitHub does it). ![image](https://github.com/user-attachments/assets/9d938d86-1a8d-4f89-8644-1921a473e858) Resolves #32096 --------- Signed-off-by: Bruno Sofiato <bruno.sofiato@gmail.com>
77 lines
1.6 KiB
Go
77 lines
1.6 KiB
Go
// Copyright 2024 The Gitea Authors. All rights reserved.
|
|
// SPDX-License-Identifier: MIT
|
|
|
|
package path
|
|
|
|
import (
|
|
"fmt"
|
|
"testing"
|
|
|
|
"github.com/blevesearch/bleve/v2/analysis"
|
|
"github.com/blevesearch/bleve/v2/analysis/tokenizer/unicode"
|
|
"github.com/stretchr/testify/assert"
|
|
)
|
|
|
|
type Scenario struct {
|
|
Input string
|
|
Tokens []string
|
|
}
|
|
|
|
func TestTokenFilter(t *testing.T) {
|
|
scenarios := []struct {
|
|
Input string
|
|
Terms []string
|
|
}{
|
|
{
|
|
Input: "Dockerfile",
|
|
Terms: []string{"Dockerfile"},
|
|
},
|
|
{
|
|
Input: "Dockerfile.rootless",
|
|
Terms: []string{"Dockerfile.rootless"},
|
|
},
|
|
{
|
|
Input: "a/b/c/Dockerfile.rootless",
|
|
Terms: []string{"a", "a/b", "a/b/c", "a/b/c/Dockerfile.rootless", "Dockerfile.rootless", "Dockerfile.rootless/c", "Dockerfile.rootless/c/b", "Dockerfile.rootless/c/b/a"},
|
|
},
|
|
{
|
|
Input: "",
|
|
Terms: []string{},
|
|
},
|
|
}
|
|
|
|
for _, scenario := range scenarios {
|
|
t.Run(fmt.Sprintf("ensure terms of '%s'", scenario.Input), func(t *testing.T) {
|
|
terms := extractTerms(scenario.Input)
|
|
|
|
assert.Len(t, terms, len(scenario.Terms))
|
|
|
|
for _, term := range terms {
|
|
assert.Contains(t, scenario.Terms, term)
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
func extractTerms(input string) []string {
|
|
tokens := tokenize(input)
|
|
filteredTokens := filter(tokens)
|
|
terms := make([]string, 0, len(filteredTokens))
|
|
|
|
for _, token := range filteredTokens {
|
|
terms = append(terms, string(token.Term))
|
|
}
|
|
|
|
return terms
|
|
}
|
|
|
|
func filter(input analysis.TokenStream) analysis.TokenStream {
|
|
filter := NewTokenFilter()
|
|
return filter.Filter(input)
|
|
}
|
|
|
|
func tokenize(input string) analysis.TokenStream {
|
|
tokenizer := unicode.NewUnicodeTokenizer()
|
|
return tokenizer.Tokenize([]byte(input))
|
|
}
|