gitea/vendor/github.com/olivere/elastic/v7/scroll.go
Lunny Xiao 5dbf36f356
Issue search support elasticsearch (#9428)
* Issue search support elasticsearch

* Fix lint

* Add indexer name on app.ini

* add a warnning on SearchIssuesByKeyword

* improve code
2020-02-13 14:06:17 +08:00

591 lines
16 KiB
Go

// Copyright 2012-present Oliver Eilhard. All rights reserved.
// Use of this source code is governed by a MIT-license.
// See http://olivere.mit-license.org/license.txt for details.
package elastic
import (
"context"
"fmt"
"io"
"net/http"
"net/url"
"strings"
"sync"
"github.com/olivere/elastic/v7/uritemplates"
)
const (
// DefaultScrollKeepAlive is the default time a scroll cursor will be kept alive.
DefaultScrollKeepAlive = "5m"
)
// ScrollService iterates over pages of search results from Elasticsearch.
type ScrollService struct {
client *Client
retrier Retrier
pretty *bool // pretty format the returned JSON response
human *bool // return human readable values for statistics
errorTrace *bool // include the stack trace of returned errors
filterPath []string // list of filters used to reduce the response
headers http.Header // custom request-level HTTP headers
indices []string
types []string
keepAlive string
body interface{}
ss *SearchSource
size *int
routing string
preference string
ignoreUnavailable *bool
allowNoIndices *bool
expandWildcards string
maxResponseSize int64
mu sync.RWMutex
scrollId string
}
// NewScrollService initializes and returns a new ScrollService.
func NewScrollService(client *Client) *ScrollService {
builder := &ScrollService{
client: client,
ss: NewSearchSource(),
keepAlive: DefaultScrollKeepAlive,
}
return builder
}
// Pretty tells Elasticsearch whether to return a formatted JSON response.
func (s *ScrollService) Pretty(pretty bool) *ScrollService {
s.pretty = &pretty
return s
}
// Human specifies whether human readable values should be returned in
// the JSON response, e.g. "7.5mb".
func (s *ScrollService) Human(human bool) *ScrollService {
s.human = &human
return s
}
// ErrorTrace specifies whether to include the stack trace of returned errors.
func (s *ScrollService) ErrorTrace(errorTrace bool) *ScrollService {
s.errorTrace = &errorTrace
return s
}
// FilterPath specifies a list of filters used to reduce the response.
func (s *ScrollService) FilterPath(filterPath ...string) *ScrollService {
s.filterPath = filterPath
return s
}
// Header adds a header to the request.
func (s *ScrollService) Header(name string, value string) *ScrollService {
if s.headers == nil {
s.headers = http.Header{}
}
s.headers.Add(name, value)
return s
}
// Headers specifies the headers of the request.
func (s *ScrollService) Headers(headers http.Header) *ScrollService {
s.headers = headers
return s
}
// Retrier allows to set specific retry logic for this ScrollService.
// If not specified, it will use the client's default retrier.
func (s *ScrollService) Retrier(retrier Retrier) *ScrollService {
s.retrier = retrier
return s
}
// Index sets the name of one or more indices to iterate over.
func (s *ScrollService) Index(indices ...string) *ScrollService {
if s.indices == nil {
s.indices = make([]string, 0)
}
s.indices = append(s.indices, indices...)
return s
}
// Type sets the name of one or more types to iterate over.
func (s *ScrollService) Type(types ...string) *ScrollService {
if s.types == nil {
s.types = make([]string, 0)
}
s.types = append(s.types, types...)
return s
}
// Scroll is an alias for KeepAlive, the time to keep
// the cursor alive (e.g. "5m" for 5 minutes).
func (s *ScrollService) Scroll(keepAlive string) *ScrollService {
s.keepAlive = keepAlive
return s
}
// KeepAlive sets the maximum time after which the cursor will expire.
// It is "5m" by default.
func (s *ScrollService) KeepAlive(keepAlive string) *ScrollService {
s.keepAlive = keepAlive
return s
}
// Size specifies the number of documents Elasticsearch should return
// from each shard, per page.
func (s *ScrollService) Size(size int) *ScrollService {
s.size = &size
return s
}
// Highlight allows to highlight search results on one or more fields
func (s *ScrollService) Highlight(highlight *Highlight) *ScrollService {
s.ss = s.ss.Highlight(highlight)
return s
}
// Body sets the raw body to send to Elasticsearch. This can be e.g. a string,
// a map[string]interface{} or anything that can be serialized into JSON.
// Notice that setting the body disables the use of SearchSource and many
// other properties of the ScanService.
func (s *ScrollService) Body(body interface{}) *ScrollService {
s.body = body
return s
}
// SearchSource sets the search source builder to use with this iterator.
// Notice that only a certain number of properties can be used when scrolling,
// e.g. query and sorting.
func (s *ScrollService) SearchSource(searchSource *SearchSource) *ScrollService {
s.ss = searchSource
if s.ss == nil {
s.ss = NewSearchSource()
}
return s
}
// Query sets the query to perform, e.g. a MatchAllQuery.
func (s *ScrollService) Query(query Query) *ScrollService {
s.ss = s.ss.Query(query)
return s
}
// PostFilter is executed as the last filter. It only affects the
// search hits but not facets. See
// https://www.elastic.co/guide/en/elasticsearch/reference/7.0/search-request-post-filter.html
// for details.
func (s *ScrollService) PostFilter(postFilter Query) *ScrollService {
s.ss = s.ss.PostFilter(postFilter)
return s
}
// Slice allows slicing the scroll request into several batches.
// This is supported in Elasticsearch 5.0 or later.
// See https://www.elastic.co/guide/en/elasticsearch/reference/7.0/search-request-scroll.html#sliced-scroll
// for details.
func (s *ScrollService) Slice(sliceQuery Query) *ScrollService {
s.ss = s.ss.Slice(sliceQuery)
return s
}
// FetchSource indicates whether the response should contain the stored
// _source for every hit.
func (s *ScrollService) FetchSource(fetchSource bool) *ScrollService {
s.ss = s.ss.FetchSource(fetchSource)
return s
}
// FetchSourceContext indicates how the _source should be fetched.
func (s *ScrollService) FetchSourceContext(fetchSourceContext *FetchSourceContext) *ScrollService {
s.ss = s.ss.FetchSourceContext(fetchSourceContext)
return s
}
// Version can be set to true to return a version for each search hit.
// See https://www.elastic.co/guide/en/elasticsearch/reference/7.0/search-request-version.html.
func (s *ScrollService) Version(version bool) *ScrollService {
s.ss = s.ss.Version(version)
return s
}
// Sort adds a sort order. This can have negative effects on the performance
// of the scroll operation as Elasticsearch needs to sort first.
func (s *ScrollService) Sort(field string, ascending bool) *ScrollService {
s.ss = s.ss.Sort(field, ascending)
return s
}
// SortWithInfo specifies a sort order. Notice that sorting can have a
// negative impact on scroll performance.
func (s *ScrollService) SortWithInfo(info SortInfo) *ScrollService {
s.ss = s.ss.SortWithInfo(info)
return s
}
// SortBy specifies a sort order. Notice that sorting can have a
// negative impact on scroll performance.
func (s *ScrollService) SortBy(sorter ...Sorter) *ScrollService {
s.ss = s.ss.SortBy(sorter...)
return s
}
// TrackTotalHits controls if the total hit count for the query should be tracked.
//
// See https://www.elastic.co/guide/en/elasticsearch/reference/7.1/search-request-track-total-hits.html
// for details.
func (s *ScrollService) TrackTotalHits(trackTotalHits interface{}) *ScrollService {
s.ss = s.ss.TrackTotalHits(trackTotalHits)
return s
}
// Routing is a list of specific routing values to control the shards
// the search will be executed on.
func (s *ScrollService) Routing(routings ...string) *ScrollService {
s.routing = strings.Join(routings, ",")
return s
}
// Preference sets the preference to execute the search. Defaults to
// randomize across shards ("random"). Can be set to "_local" to prefer
// local shards, "_primary" to execute on primary shards only,
// or a custom value which guarantees that the same order will be used
// across different requests.
func (s *ScrollService) Preference(preference string) *ScrollService {
s.preference = preference
return s
}
// IgnoreUnavailable indicates whether the specified concrete indices
// should be ignored when unavailable (missing or closed).
func (s *ScrollService) IgnoreUnavailable(ignoreUnavailable bool) *ScrollService {
s.ignoreUnavailable = &ignoreUnavailable
return s
}
// AllowNoIndices indicates whether to ignore if a wildcard indices
// expression resolves into no concrete indices. (This includes `_all` string
// or when no indices have been specified).
func (s *ScrollService) AllowNoIndices(allowNoIndices bool) *ScrollService {
s.allowNoIndices = &allowNoIndices
return s
}
// ExpandWildcards indicates whether to expand wildcard expression to
// concrete indices that are open, closed or both.
func (s *ScrollService) ExpandWildcards(expandWildcards string) *ScrollService {
s.expandWildcards = expandWildcards
return s
}
// MaxResponseSize sets an upper limit on the response body size that we accept,
// to guard against OOM situations.
func (s *ScrollService) MaxResponseSize(maxResponseSize int64) *ScrollService {
s.maxResponseSize = maxResponseSize
return s
}
// ScrollId specifies the identifier of a scroll in action.
func (s *ScrollService) ScrollId(scrollId string) *ScrollService {
s.mu.Lock()
s.scrollId = scrollId
s.mu.Unlock()
return s
}
// Do returns the next search result. It will return io.EOF as error if there
// are no more search results.
func (s *ScrollService) Do(ctx context.Context) (*SearchResult, error) {
s.mu.RLock()
nextScrollId := s.scrollId
s.mu.RUnlock()
if len(nextScrollId) == 0 {
return s.first(ctx)
}
return s.next(ctx)
}
// Clear cancels the current scroll operation. If you don't do this manually,
// the scroll will be expired automatically by Elasticsearch. You can control
// how long a scroll cursor is kept alive with the KeepAlive func.
func (s *ScrollService) Clear(ctx context.Context) error {
s.mu.RLock()
scrollId := s.scrollId
s.mu.RUnlock()
if len(scrollId) == 0 {
return nil
}
path := "/_search/scroll"
params := url.Values{}
if v := s.pretty; v != nil {
params.Set("pretty", fmt.Sprint(*v))
}
if v := s.human; v != nil {
params.Set("human", fmt.Sprint(*v))
}
if v := s.errorTrace; v != nil {
params.Set("error_trace", fmt.Sprint(*v))
}
if len(s.filterPath) > 0 {
params.Set("filter_path", strings.Join(s.filterPath, ","))
}
body := struct {
ScrollId []string `json:"scroll_id,omitempty"`
}{
ScrollId: []string{scrollId},
}
_, err := s.client.PerformRequest(ctx, PerformRequestOptions{
Method: "DELETE",
Path: path,
Params: params,
Body: body,
Retrier: s.retrier,
})
if err != nil {
return err
}
return nil
}
// -- First --
// first takes the first page of search results.
func (s *ScrollService) first(ctx context.Context) (*SearchResult, error) {
// Get URL and parameters for request
path, params, err := s.buildFirstURL()
if err != nil {
return nil, err
}
// Get HTTP request body
body, err := s.bodyFirst()
if err != nil {
return nil, err
}
// Get HTTP response
res, err := s.client.PerformRequest(ctx, PerformRequestOptions{
Method: "POST",
Path: path,
Params: params,
Body: body,
Retrier: s.retrier,
Headers: s.headers,
MaxResponseSize: s.maxResponseSize,
})
if err != nil {
return nil, err
}
// Return operation response
ret := new(SearchResult)
if err := s.client.decoder.Decode(res.Body, ret); err != nil {
return nil, err
}
s.mu.Lock()
s.scrollId = ret.ScrollId
s.mu.Unlock()
if ret.Hits == nil || len(ret.Hits.Hits) == 0 {
return ret, io.EOF
}
return ret, nil
}
// buildFirstURL builds the URL for retrieving the first page.
func (s *ScrollService) buildFirstURL() (string, url.Values, error) {
// Build URL
var err error
var path string
if len(s.indices) == 0 && len(s.types) == 0 {
path = "/_search"
} else if len(s.indices) > 0 && len(s.types) == 0 {
path, err = uritemplates.Expand("/{index}/_search", map[string]string{
"index": strings.Join(s.indices, ","),
})
} else if len(s.indices) == 0 && len(s.types) > 0 {
path, err = uritemplates.Expand("/_all/{typ}/_search", map[string]string{
"typ": strings.Join(s.types, ","),
})
} else {
path, err = uritemplates.Expand("/{index}/{typ}/_search", map[string]string{
"index": strings.Join(s.indices, ","),
"typ": strings.Join(s.types, ","),
})
}
if err != nil {
return "", url.Values{}, err
}
// Add query string parameters
params := url.Values{}
if v := s.pretty; v != nil {
params.Set("pretty", fmt.Sprint(*v))
}
if v := s.human; v != nil {
params.Set("human", fmt.Sprint(*v))
}
if v := s.errorTrace; v != nil {
params.Set("error_trace", fmt.Sprint(*v))
}
if len(s.filterPath) > 0 {
// Always add "hits._scroll_id", otherwise we cannot scroll
var found bool
for _, path := range s.filterPath {
if path == "_scroll_id" {
found = true
break
}
}
if !found {
s.filterPath = append(s.filterPath, "_scroll_id")
}
params.Set("filter_path", strings.Join(s.filterPath, ","))
}
if s.size != nil && *s.size > 0 {
params.Set("size", fmt.Sprintf("%d", *s.size))
}
if len(s.keepAlive) > 0 {
params.Set("scroll", s.keepAlive)
}
if len(s.routing) > 0 {
params.Set("routing", s.routing)
}
if len(s.preference) > 0 {
params.Set("preference", s.preference)
}
if s.allowNoIndices != nil {
params.Set("allow_no_indices", fmt.Sprintf("%v", *s.allowNoIndices))
}
if len(s.expandWildcards) > 0 {
params.Set("expand_wildcards", s.expandWildcards)
}
if s.ignoreUnavailable != nil {
params.Set("ignore_unavailable", fmt.Sprintf("%v", *s.ignoreUnavailable))
}
return path, params, nil
}
// bodyFirst returns the request to fetch the first batch of results.
func (s *ScrollService) bodyFirst() (interface{}, error) {
var err error
var body interface{}
if s.body != nil {
body = s.body
} else {
// Use _doc sort by default if none is specified
if !s.ss.hasSort() {
// Use efficient sorting when no user-defined query/body is specified
s.ss = s.ss.SortBy(SortByDoc{})
}
// Body from search source
body, err = s.ss.Source()
if err != nil {
return nil, err
}
}
return body, nil
}
// -- Next --
func (s *ScrollService) next(ctx context.Context) (*SearchResult, error) {
// Get URL for request
path, params, err := s.buildNextURL()
if err != nil {
return nil, err
}
// Setup HTTP request body
body, err := s.bodyNext()
if err != nil {
return nil, err
}
// Get HTTP response
res, err := s.client.PerformRequest(ctx, PerformRequestOptions{
Method: "POST",
Path: path,
Params: params,
Body: body,
Retrier: s.retrier,
Headers: s.headers,
MaxResponseSize: s.maxResponseSize,
})
if err != nil {
return nil, err
}
// Return operation response
ret := new(SearchResult)
if err := s.client.decoder.Decode(res.Body, ret); err != nil {
return nil, err
}
s.mu.Lock()
s.scrollId = ret.ScrollId
s.mu.Unlock()
if ret.Hits == nil || len(ret.Hits.Hits) == 0 {
return ret, io.EOF
}
return ret, nil
}
// buildNextURL builds the URL for the operation.
func (s *ScrollService) buildNextURL() (string, url.Values, error) {
path := "/_search/scroll"
// Add query string parameters
params := url.Values{}
if v := s.pretty; v != nil {
params.Set("pretty", fmt.Sprint(*v))
}
if v := s.human; v != nil {
params.Set("human", fmt.Sprint(*v))
}
if v := s.errorTrace; v != nil {
params.Set("error_trace", fmt.Sprint(*v))
}
if len(s.filterPath) > 0 {
// Always add "hits._scroll_id", otherwise we cannot scroll
var found bool
for _, path := range s.filterPath {
if path == "_scroll_id" {
found = true
break
}
}
if !found {
s.filterPath = append(s.filterPath, "_scroll_id")
}
params.Set("filter_path", strings.Join(s.filterPath, ","))
}
return path, params, nil
}
// body returns the request to fetch the next batch of results.
func (s *ScrollService) bodyNext() (interface{}, error) {
s.mu.RLock()
body := struct {
Scroll string `json:"scroll"`
ScrollId string `json:"scroll_id,omitempty"`
}{
Scroll: s.keepAlive,
ScrollId: s.scrollId,
}
s.mu.RUnlock()
return body, nil
}