mirror of
				https://github.com/go-gitea/gitea.git
				synced 2025-10-29 10:57:44 +09:00 
			
		
		
		
	
		
			
				
	
	
		
			90 lines
		
	
	
		
			2.6 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			90 lines
		
	
	
		
			2.6 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| //  Copyright (c) 2014 Couchbase, Inc.
 | |
| //
 | |
| // Licensed under the Apache License, Version 2.0 (the "License");
 | |
| // you may not use this file except in compliance with the License.
 | |
| // You may obtain a copy of the License at
 | |
| //
 | |
| // 		http://www.apache.org/licenses/LICENSE-2.0
 | |
| //
 | |
| // Unless required by applicable law or agreed to in writing, software
 | |
| // distributed under the License is distributed on an "AS IS" BASIS,
 | |
| // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | |
| // See the License for the specific language governing permissions and
 | |
| // limitations under the License.
 | |
| 
 | |
| package registry
 | |
| 
 | |
| import (
 | |
| 	"fmt"
 | |
| 
 | |
| 	"github.com/blevesearch/bleve/analysis"
 | |
| )
 | |
| 
 | |
| func RegisterTokenizer(name string, constructor TokenizerConstructor) {
 | |
| 	_, exists := tokenizers[name]
 | |
| 	if exists {
 | |
| 		panic(fmt.Errorf("attempted to register duplicate tokenizer named '%s'", name))
 | |
| 	}
 | |
| 	tokenizers[name] = constructor
 | |
| }
 | |
| 
 | |
| type TokenizerConstructor func(config map[string]interface{}, cache *Cache) (analysis.Tokenizer, error)
 | |
| type TokenizerRegistry map[string]TokenizerConstructor
 | |
| 
 | |
| type TokenizerCache struct {
 | |
| 	*ConcurrentCache
 | |
| }
 | |
| 
 | |
| func NewTokenizerCache() *TokenizerCache {
 | |
| 	return &TokenizerCache{
 | |
| 		NewConcurrentCache(),
 | |
| 	}
 | |
| }
 | |
| 
 | |
| func TokenizerBuild(name string, config map[string]interface{}, cache *Cache) (interface{}, error) {
 | |
| 	cons, registered := tokenizers[name]
 | |
| 	if !registered {
 | |
| 		return nil, fmt.Errorf("no tokenizer with name or type '%s' registered", name)
 | |
| 	}
 | |
| 	tokenizer, err := cons(config, cache)
 | |
| 	if err != nil {
 | |
| 		return nil, fmt.Errorf("error building tokenizer: %v", err)
 | |
| 	}
 | |
| 	return tokenizer, nil
 | |
| }
 | |
| 
 | |
| func (c *TokenizerCache) TokenizerNamed(name string, cache *Cache) (analysis.Tokenizer, error) {
 | |
| 	item, err := c.ItemNamed(name, cache, TokenizerBuild)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 	return item.(analysis.Tokenizer), nil
 | |
| }
 | |
| 
 | |
| func (c *TokenizerCache) DefineTokenizer(name string, typ string, config map[string]interface{}, cache *Cache) (analysis.Tokenizer, error) {
 | |
| 	item, err := c.DefineItem(name, typ, config, cache, TokenizerBuild)
 | |
| 	if err != nil {
 | |
| 		if err == ErrAlreadyDefined {
 | |
| 			return nil, fmt.Errorf("tokenizer named '%s' already defined", name)
 | |
| 		}
 | |
| 		return nil, err
 | |
| 	}
 | |
| 	return item.(analysis.Tokenizer), nil
 | |
| }
 | |
| 
 | |
| func TokenizerTypesAndInstances() ([]string, []string) {
 | |
| 	emptyConfig := map[string]interface{}{}
 | |
| 	emptyCache := NewCache()
 | |
| 	var types []string
 | |
| 	var instances []string
 | |
| 	for name, cons := range tokenizers {
 | |
| 		_, err := cons(emptyConfig, emptyCache)
 | |
| 		if err == nil {
 | |
| 			instances = append(instances, name)
 | |
| 		} else {
 | |
| 			types = append(types, name)
 | |
| 		}
 | |
| 	}
 | |
| 	return types, instances
 | |
| }
 |