mirror of
				https://github.com/go-gitea/gitea.git
				synced 2025-10-29 10:57:44 +09:00 
			
		
		
		
	chore: update github.com/couchbase/gomemcached and github.com/couchbase/go-couchbase (#9419)
This commit is contained in:
		
				
					committed by
					
						 techknowlogick
						techknowlogick
					
				
			
			
				
	
			
			
			
						parent
						
							8873a80276
						
					
				
				
					commit
					559fb6ccf0
				
			
							
								
								
									
										2
									
								
								go.mod
									
									
									
									
									
								
							
							
						
						
									
										2
									
								
								go.mod
									
									
									
									
									
								
							| @@ -24,6 +24,8 @@ require ( | |||||||
| 	github.com/blevesearch/go-porterstemmer v1.0.2 // indirect | 	github.com/blevesearch/go-porterstemmer v1.0.2 // indirect | ||||||
| 	github.com/blevesearch/segment v0.0.0-20160915185041-762005e7a34f // indirect | 	github.com/blevesearch/segment v0.0.0-20160915185041-762005e7a34f // indirect | ||||||
| 	github.com/boombuler/barcode v0.0.0-20161226211916-fe0f26ff6d26 // indirect | 	github.com/boombuler/barcode v0.0.0-20161226211916-fe0f26ff6d26 // indirect | ||||||
|  | 	github.com/couchbase/gomemcached v0.0.0-20191004160342-7b5da2ec40b2 // indirect | ||||||
|  | 	github.com/couchbase/goutils v0.0.0-20191018232750-b49639060d85 // indirect | ||||||
| 	github.com/couchbase/vellum v0.0.0-20190829182332-ef2e028c01fd // indirect | 	github.com/couchbase/vellum v0.0.0-20190829182332-ef2e028c01fd // indirect | ||||||
| 	github.com/cznic/b v0.0.0-20181122101859-a26611c4d92d // indirect | 	github.com/cznic/b v0.0.0-20181122101859-a26611c4d92d // indirect | ||||||
| 	github.com/cznic/mathutil v0.0.0-20181122101859-297441e03548 // indirect | 	github.com/cznic/mathutil v0.0.0-20181122101859-297441e03548 // indirect | ||||||
|   | |||||||
							
								
								
									
										4
									
								
								go.sum
									
									
									
									
									
								
							
							
						
						
									
										4
									
								
								go.sum
									
									
									
									
									
								
							| @@ -97,8 +97,12 @@ github.com/coreos/go-systemd v0.0.0-20190321100706-95778dfbb74e/go.mod h1:F5haX7 | |||||||
| github.com/coreos/pkg v0.0.0-20180928190104-399ea9e2e55f/go.mod h1:E3G3o1h8I7cfcXa63jLwjI0eiQQMgzzUDFVpN/nH/eA= | github.com/coreos/pkg v0.0.0-20180928190104-399ea9e2e55f/go.mod h1:E3G3o1h8I7cfcXa63jLwjI0eiQQMgzzUDFVpN/nH/eA= | ||||||
| github.com/couchbase/gomemcached v0.0.0-20190515232915-c4b4ca0eb21d h1:XMf4E1U+b9E3ElF0mjvfXZdflBRZz4gLp16nQ/QSHQM= | github.com/couchbase/gomemcached v0.0.0-20190515232915-c4b4ca0eb21d h1:XMf4E1U+b9E3ElF0mjvfXZdflBRZz4gLp16nQ/QSHQM= | ||||||
| github.com/couchbase/gomemcached v0.0.0-20190515232915-c4b4ca0eb21d/go.mod h1:srVSlQLB8iXBVXHgnqemxUXqN6FCvClgCMPCsjBDR7c= | github.com/couchbase/gomemcached v0.0.0-20190515232915-c4b4ca0eb21d/go.mod h1:srVSlQLB8iXBVXHgnqemxUXqN6FCvClgCMPCsjBDR7c= | ||||||
|  | github.com/couchbase/gomemcached v0.0.0-20191004160342-7b5da2ec40b2 h1:vZryARwW4PSFXd9arwegEywvMTvPuXL3/oa+4L5NTe8= | ||||||
|  | github.com/couchbase/gomemcached v0.0.0-20191004160342-7b5da2ec40b2/go.mod h1:srVSlQLB8iXBVXHgnqemxUXqN6FCvClgCMPCsjBDR7c= | ||||||
| github.com/couchbase/goutils v0.0.0-20190315194238-f9d42b11473b h1:bZ9rKU2/V8sY+NulSfxDOnXTWcs1rySqdF1sVepihvo= | github.com/couchbase/goutils v0.0.0-20190315194238-f9d42b11473b h1:bZ9rKU2/V8sY+NulSfxDOnXTWcs1rySqdF1sVepihvo= | ||||||
| github.com/couchbase/goutils v0.0.0-20190315194238-f9d42b11473b/go.mod h1:BQwMFlJzDjFDG3DJUdU0KORxn88UlsOULuxLExMh3Hs= | github.com/couchbase/goutils v0.0.0-20190315194238-f9d42b11473b/go.mod h1:BQwMFlJzDjFDG3DJUdU0KORxn88UlsOULuxLExMh3Hs= | ||||||
|  | github.com/couchbase/goutils v0.0.0-20191018232750-b49639060d85 h1:0WMIDtuXCKEm4wtAJgAAXa/qtM5O9MariLwgHaRlYmk= | ||||||
|  | github.com/couchbase/goutils v0.0.0-20191018232750-b49639060d85/go.mod h1:BQwMFlJzDjFDG3DJUdU0KORxn88UlsOULuxLExMh3Hs= | ||||||
| github.com/couchbase/vellum v0.0.0-20190829182332-ef2e028c01fd h1:zeuJhcG3f8eePshH3KxkNE+Xtl53pVln9MOUPMyr/1w= | github.com/couchbase/vellum v0.0.0-20190829182332-ef2e028c01fd h1:zeuJhcG3f8eePshH3KxkNE+Xtl53pVln9MOUPMyr/1w= | ||||||
| github.com/couchbase/vellum v0.0.0-20190829182332-ef2e028c01fd/go.mod h1:xbc8Ff/oG7h2ejd7AlwOpfd+6QZntc92ygpAOfGwcKY= | github.com/couchbase/vellum v0.0.0-20190829182332-ef2e028c01fd/go.mod h1:xbc8Ff/oG7h2ejd7AlwOpfd+6QZntc92ygpAOfGwcKY= | ||||||
| github.com/couchbaselabs/go-couchbase v0.0.0-20190708161019-23e7ca2ce2b7 h1:1XjEY/gnjQ+AfXef2U6dxCquhiRzkEpxZuWqs+QxTL8= | github.com/couchbaselabs/go-couchbase v0.0.0-20190708161019-23e7ca2ce2b7 h1:1XjEY/gnjQ+AfXef2U6dxCquhiRzkEpxZuWqs+QxTL8= | ||||||
|   | |||||||
							
								
								
									
										123
									
								
								vendor/github.com/couchbase/gomemcached/client/collections_filter.go
									
									
									
										generated
									
									
										vendored
									
									
										Normal file
									
								
							
							
						
						
									
										123
									
								
								vendor/github.com/couchbase/gomemcached/client/collections_filter.go
									
									
									
										generated
									
									
										vendored
									
									
										Normal file
									
								
							| @@ -0,0 +1,123 @@ | |||||||
|  | package memcached | ||||||
|  |  | ||||||
|  | import ( | ||||||
|  | 	"encoding/json" | ||||||
|  | 	"fmt" | ||||||
|  | ) | ||||||
|  |  | ||||||
|  | // Collection based filter | ||||||
|  | type CollectionsFilter struct { | ||||||
|  | 	ManifestUid    uint64 | ||||||
|  | 	UseManifestUid bool | ||||||
|  | 	StreamId       uint16 | ||||||
|  | 	UseStreamId    bool | ||||||
|  |  | ||||||
|  | 	// Use either ScopeId OR CollectionsList, not both | ||||||
|  | 	CollectionsList []uint32 | ||||||
|  | 	ScopeId         uint32 | ||||||
|  | } | ||||||
|  |  | ||||||
|  | type nonStreamIdNonResumeScopeMeta struct { | ||||||
|  | 	ScopeId string `json:"scope"` | ||||||
|  | } | ||||||
|  |  | ||||||
|  | type nonStreamIdResumeScopeMeta struct { | ||||||
|  | 	ManifestId string `json:"uid"` | ||||||
|  | } | ||||||
|  |  | ||||||
|  | type nonStreamIdNonResumeCollectionsMeta struct { | ||||||
|  | 	CollectionsList []string `json:"collections"` | ||||||
|  | } | ||||||
|  |  | ||||||
|  | type nonStreamIdResumeCollectionsMeta struct { | ||||||
|  | 	ManifestId      string   `json:"uid"` | ||||||
|  | 	CollectionsList []string `json:"collections"` | ||||||
|  | } | ||||||
|  |  | ||||||
|  | type streamIdNonResumeCollectionsMeta struct { | ||||||
|  | 	CollectionsList []string `json:"collections"` | ||||||
|  | 	StreamId        uint16   `json:"sid"` | ||||||
|  | } | ||||||
|  |  | ||||||
|  | type streamIdNonResumeScopeMeta struct { | ||||||
|  | 	ScopeId  string `json:"scope"` | ||||||
|  | 	StreamId uint16 `json:"sid"` | ||||||
|  | } | ||||||
|  |  | ||||||
|  | func (c *CollectionsFilter) IsValid() error { | ||||||
|  | 	if c.UseManifestUid { | ||||||
|  | 		return fmt.Errorf("Not implemented yet") | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	if len(c.CollectionsList) > 0 && c.ScopeId > 0 { | ||||||
|  | 		return fmt.Errorf("Collection list is specified but scope ID is also specified") | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	return nil | ||||||
|  | } | ||||||
|  |  | ||||||
|  | func (c *CollectionsFilter) outputCollectionsFilterColList() (outputList []string) { | ||||||
|  | 	for _, collectionUint := range c.CollectionsList { | ||||||
|  | 		outputList = append(outputList, fmt.Sprintf("%x", collectionUint)) | ||||||
|  | 	} | ||||||
|  | 	return | ||||||
|  | } | ||||||
|  |  | ||||||
|  | func (c *CollectionsFilter) outputScopeId() string { | ||||||
|  | 	return fmt.Sprintf("%x", c.ScopeId) | ||||||
|  | } | ||||||
|  |  | ||||||
|  | func (c *CollectionsFilter) ToStreamReqBody() ([]byte, error) { | ||||||
|  | 	if err := c.IsValid(); err != nil { | ||||||
|  | 		return nil, err | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	var output interface{} | ||||||
|  |  | ||||||
|  | 	switch c.UseStreamId { | ||||||
|  | 	case true: | ||||||
|  | 		switch c.UseManifestUid { | ||||||
|  | 		case true: | ||||||
|  | 			// TODO | ||||||
|  | 			return nil, fmt.Errorf("NotImplemented0") | ||||||
|  | 		case false: | ||||||
|  | 			switch len(c.CollectionsList) > 0 { | ||||||
|  | 			case true: | ||||||
|  | 				filter := &streamIdNonResumeCollectionsMeta{ | ||||||
|  | 					StreamId:        c.StreamId, | ||||||
|  | 					CollectionsList: c.outputCollectionsFilterColList(), | ||||||
|  | 				} | ||||||
|  | 				output = *filter | ||||||
|  | 			case false: | ||||||
|  | 				filter := &streamIdNonResumeScopeMeta{ | ||||||
|  | 					StreamId: c.StreamId, | ||||||
|  | 					ScopeId:  c.outputScopeId(), | ||||||
|  | 				} | ||||||
|  | 				output = *filter | ||||||
|  | 			} | ||||||
|  | 		} | ||||||
|  | 	case false: | ||||||
|  | 		switch c.UseManifestUid { | ||||||
|  | 		case true: | ||||||
|  | 			// TODO | ||||||
|  | 			return nil, fmt.Errorf("NotImplemented1") | ||||||
|  | 		case false: | ||||||
|  | 			switch len(c.CollectionsList) > 0 { | ||||||
|  | 			case true: | ||||||
|  | 				filter := &nonStreamIdNonResumeCollectionsMeta{ | ||||||
|  | 					CollectionsList: c.outputCollectionsFilterColList(), | ||||||
|  | 				} | ||||||
|  | 				output = *filter | ||||||
|  | 			case false: | ||||||
|  | 				output = nonStreamIdNonResumeScopeMeta{ScopeId: c.outputScopeId()} | ||||||
|  | 			} | ||||||
|  | 		} | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	data, err := json.Marshal(output) | ||||||
|  | 	if err != nil { | ||||||
|  | 		return nil, err | ||||||
|  | 	} else { | ||||||
|  | 		return data, nil | ||||||
|  | 	} | ||||||
|  | } | ||||||
							
								
								
									
										30
									
								
								vendor/github.com/couchbase/gomemcached/client/mc.go
									
									
									
										generated
									
									
										vendored
									
									
								
							
							
						
						
									
										30
									
								
								vendor/github.com/couchbase/gomemcached/client/mc.go
									
									
									
										generated
									
									
										vendored
									
									
								
							| @@ -28,10 +28,12 @@ type ClientIface interface { | |||||||
| 	CASNext(vb uint16, k string, exp int, state *CASState) bool | 	CASNext(vb uint16, k string, exp int, state *CASState) bool | ||||||
| 	CAS(vb uint16, k string, f CasFunc, initexp int) (*gomemcached.MCResponse, error) | 	CAS(vb uint16, k string, f CasFunc, initexp int) (*gomemcached.MCResponse, error) | ||||||
| 	CollectionsGetCID(scope string, collection string) (*gomemcached.MCResponse, error) | 	CollectionsGetCID(scope string, collection string) (*gomemcached.MCResponse, error) | ||||||
|  | 	CollectionEnabled() bool | ||||||
| 	Close() error | 	Close() error | ||||||
| 	Decr(vb uint16, key string, amt, def uint64, exp int) (uint64, error) | 	Decr(vb uint16, key string, amt, def uint64, exp int) (uint64, error) | ||||||
| 	Del(vb uint16, key string) (*gomemcached.MCResponse, error) | 	Del(vb uint16, key string) (*gomemcached.MCResponse, error) | ||||||
| 	EnableMutationToken() (*gomemcached.MCResponse, error) | 	EnableMutationToken() (*gomemcached.MCResponse, error) | ||||||
|  | 	EnableFeatures(features Features) (*gomemcached.MCResponse, error) | ||||||
| 	Get(vb uint16, key string) (*gomemcached.MCResponse, error) | 	Get(vb uint16, key string) (*gomemcached.MCResponse, error) | ||||||
| 	GetCollectionsManifest() (*gomemcached.MCResponse, error) | 	GetCollectionsManifest() (*gomemcached.MCResponse, error) | ||||||
| 	GetFromCollection(vb uint16, cid uint32, key string) (*gomemcached.MCResponse, error) | 	GetFromCollection(vb uint16, cid uint32, key string) (*gomemcached.MCResponse, error) | ||||||
| @@ -76,9 +78,12 @@ var Healthy uint32 = 1 | |||||||
| type Features []Feature | type Features []Feature | ||||||
| type Feature uint16 | type Feature uint16 | ||||||
|  |  | ||||||
| const FeatureMutationToken = Feature(0x04) | const FeatureTcpNoDelay = Feature(0x03) | ||||||
|  | const FeatureMutationToken = Feature(0x04) // XATTR bit in data type field with dcp mutations | ||||||
| const FeatureXattr = Feature(0x06) | const FeatureXattr = Feature(0x06) | ||||||
|  | const FeatureXerror = Feature(0x07) | ||||||
| const FeatureCollections = Feature(0x12) | const FeatureCollections = Feature(0x12) | ||||||
|  | const FeatureSnappyCompression = Feature(0x0a) | ||||||
| const FeatureDataType = Feature(0x0b) | const FeatureDataType = Feature(0x0b) | ||||||
|  |  | ||||||
| type memcachedConnection interface { | type memcachedConnection interface { | ||||||
| @@ -96,6 +101,9 @@ type Client struct { | |||||||
| 	opaque  uint32 | 	opaque  uint32 | ||||||
|  |  | ||||||
| 	hdrBuf []byte | 	hdrBuf []byte | ||||||
|  |  | ||||||
|  | 	featureMtx       sync.RWMutex | ||||||
|  | 	sentHeloFeatures Features | ||||||
| } | } | ||||||
|  |  | ||||||
| var ( | var ( | ||||||
| @@ -285,6 +293,10 @@ func (c *Client) EnableFeatures(features Features) (*gomemcached.MCResponse, err | |||||||
| 		binary.BigEndian.PutUint16(payload[len(payload)-2:], uint16(feature)) | 		binary.BigEndian.PutUint16(payload[len(payload)-2:], uint16(feature)) | ||||||
| 	} | 	} | ||||||
|  |  | ||||||
|  | 	c.featureMtx.Lock() | ||||||
|  | 	c.sentHeloFeatures = features | ||||||
|  | 	c.featureMtx.Unlock() | ||||||
|  |  | ||||||
| 	return c.Send(&gomemcached.MCRequest{ | 	return c.Send(&gomemcached.MCRequest{ | ||||||
| 		Opcode: gomemcached.HELLO, | 		Opcode: gomemcached.HELLO, | ||||||
| 		Key:    []byte("GoMemcached"), | 		Key:    []byte("GoMemcached"), | ||||||
| @@ -363,6 +375,18 @@ func (c *Client) CollectionsGetCID(scope string, collection string) (*gomemcache | |||||||
| 	return res, nil | 	return res, nil | ||||||
| } | } | ||||||
|  |  | ||||||
|  | func (c *Client) CollectionEnabled() bool { | ||||||
|  | 	c.featureMtx.RLock() | ||||||
|  | 	defer c.featureMtx.RUnlock() | ||||||
|  |  | ||||||
|  | 	for _, feature := range c.sentHeloFeatures { | ||||||
|  | 		if feature == FeatureCollections { | ||||||
|  | 			return true | ||||||
|  | 		} | ||||||
|  | 	} | ||||||
|  | 	return false | ||||||
|  | } | ||||||
|  |  | ||||||
| // Get the value for a key, and update expiry | // Get the value for a key, and update expiry | ||||||
| func (c *Client) GetAndTouch(vb uint16, key string, exp int) (*gomemcached.MCResponse, error) { | func (c *Client) GetAndTouch(vb uint16, key string, exp int) (*gomemcached.MCResponse, error) { | ||||||
| 	extraBuf := make([]byte, 4) | 	extraBuf := make([]byte, 4) | ||||||
| @@ -1138,3 +1162,7 @@ func IfResStatusError(response *gomemcached.MCResponse) bool { | |||||||
| 			response.Status != gomemcached.SUBDOC_PATH_NOT_FOUND && | 			response.Status != gomemcached.SUBDOC_PATH_NOT_FOUND && | ||||||
| 			response.Status != gomemcached.SUBDOC_MULTI_PATH_FAILURE_DELETED) | 			response.Status != gomemcached.SUBDOC_MULTI_PATH_FAILURE_DELETED) | ||||||
| } | } | ||||||
|  |  | ||||||
|  | func (c *Client) Conn() io.ReadWriteCloser { | ||||||
|  | 	return c.conn | ||||||
|  | } | ||||||
|   | |||||||
							
								
								
									
										346
									
								
								vendor/github.com/couchbase/gomemcached/client/upr_event.go
									
									
									
										generated
									
									
										vendored
									
									
										Normal file
									
								
							
							
						
						
									
										346
									
								
								vendor/github.com/couchbase/gomemcached/client/upr_event.go
									
									
									
										generated
									
									
										vendored
									
									
										Normal file
									
								
							| @@ -0,0 +1,346 @@ | |||||||
|  | package memcached | ||||||
|  |  | ||||||
|  | import ( | ||||||
|  | 	"encoding/binary" | ||||||
|  | 	"fmt" | ||||||
|  | 	"github.com/couchbase/gomemcached" | ||||||
|  | 	"math" | ||||||
|  | ) | ||||||
|  |  | ||||||
|  | type SystemEventType int | ||||||
|  |  | ||||||
|  | const InvalidSysEvent SystemEventType = -1 | ||||||
|  |  | ||||||
|  | const ( | ||||||
|  | 	CollectionCreate  SystemEventType = 0 | ||||||
|  | 	CollectionDrop    SystemEventType = iota | ||||||
|  | 	CollectionFlush   SystemEventType = iota // KV did not implement | ||||||
|  | 	ScopeCreate       SystemEventType = iota | ||||||
|  | 	ScopeDrop         SystemEventType = iota | ||||||
|  | 	CollectionChanged SystemEventType = iota | ||||||
|  | ) | ||||||
|  |  | ||||||
|  | type ScopeCreateEvent interface { | ||||||
|  | 	GetSystemEventName() (string, error) | ||||||
|  | 	GetScopeId() (uint32, error) | ||||||
|  | 	GetManifestId() (uint64, error) | ||||||
|  | } | ||||||
|  |  | ||||||
|  | type CollectionCreateEvent interface { | ||||||
|  | 	GetSystemEventName() (string, error) | ||||||
|  | 	GetScopeId() (uint32, error) | ||||||
|  | 	GetCollectionId() (uint32, error) | ||||||
|  | 	GetManifestId() (uint64, error) | ||||||
|  | 	GetMaxTTL() (uint32, error) | ||||||
|  | } | ||||||
|  |  | ||||||
|  | type CollectionDropEvent interface { | ||||||
|  | 	GetScopeId() (uint32, error) | ||||||
|  | 	GetCollectionId() (uint32, error) | ||||||
|  | 	GetManifestId() (uint64, error) | ||||||
|  | } | ||||||
|  |  | ||||||
|  | type ScopeDropEvent interface { | ||||||
|  | 	GetScopeId() (uint32, error) | ||||||
|  | 	GetManifestId() (uint64, error) | ||||||
|  | } | ||||||
|  |  | ||||||
|  | type CollectionChangedEvent interface { | ||||||
|  | 	GetCollectionId() (uint32, error) | ||||||
|  | 	GetManifestId() (uint64, error) | ||||||
|  | 	GetMaxTTL() (uint32, error) | ||||||
|  | } | ||||||
|  |  | ||||||
|  | var ErrorInvalidOp error = fmt.Errorf("Invalid Operation") | ||||||
|  | var ErrorInvalidVersion error = fmt.Errorf("Invalid version for parsing") | ||||||
|  | var ErrorValueTooShort error = fmt.Errorf("Value length is too short") | ||||||
|  | var ErrorNoMaxTTL error = fmt.Errorf("This event has no max TTL") | ||||||
|  |  | ||||||
|  | // UprEvent memcached events for UPR streams. | ||||||
|  | type UprEvent struct { | ||||||
|  | 	Opcode          gomemcached.CommandCode // Type of event | ||||||
|  | 	Status          gomemcached.Status      // Response status | ||||||
|  | 	VBucket         uint16                  // VBucket this event applies to | ||||||
|  | 	DataType        uint8                   // data type | ||||||
|  | 	Opaque          uint16                  // 16 MSB of opaque | ||||||
|  | 	VBuuid          uint64                  // This field is set by downstream | ||||||
|  | 	Flags           uint32                  // Item flags | ||||||
|  | 	Expiry          uint32                  // Item expiration time | ||||||
|  | 	Key, Value      []byte                  // Item key/value | ||||||
|  | 	OldValue        []byte                  // TODO: TBD: old document value | ||||||
|  | 	Cas             uint64                  // CAS value of the item | ||||||
|  | 	Seqno           uint64                  // sequence number of the mutation | ||||||
|  | 	RevSeqno        uint64                  // rev sequence number : deletions | ||||||
|  | 	LockTime        uint32                  // Lock time | ||||||
|  | 	MetadataSize    uint16                  // Metadata size | ||||||
|  | 	SnapstartSeq    uint64                  // start sequence number of this snapshot | ||||||
|  | 	SnapendSeq      uint64                  // End sequence number of the snapshot | ||||||
|  | 	SnapshotType    uint32                  // 0: disk 1: memory | ||||||
|  | 	FailoverLog     *FailoverLog            // Failover log containing vvuid and sequnce number | ||||||
|  | 	Error           error                   // Error value in case of a failure | ||||||
|  | 	ExtMeta         []byte                  // Extended Metadata | ||||||
|  | 	AckSize         uint32                  // The number of bytes that can be Acked to DCP | ||||||
|  | 	SystemEvent     SystemEventType         // Only valid if IsSystemEvent() is true | ||||||
|  | 	SysEventVersion uint8                   // Based on the version, the way Extra bytes is parsed is different | ||||||
|  | 	ValueLen        int                     // Cache it to avoid len() calls for performance | ||||||
|  | 	CollectionId    uint64                  // Valid if Collection is in use | ||||||
|  | } | ||||||
|  |  | ||||||
|  | // FailoverLog containing vvuid and sequnce number | ||||||
|  | type FailoverLog [][2]uint64 | ||||||
|  |  | ||||||
|  | func makeUprEvent(rq gomemcached.MCRequest, stream *UprStream, bytesReceivedFromDCP int) *UprEvent { | ||||||
|  | 	event := &UprEvent{ | ||||||
|  | 		Opcode:       rq.Opcode, | ||||||
|  | 		VBucket:      stream.Vbucket, | ||||||
|  | 		VBuuid:       stream.Vbuuid, | ||||||
|  | 		Value:        rq.Body, | ||||||
|  | 		Cas:          rq.Cas, | ||||||
|  | 		ExtMeta:      rq.ExtMeta, | ||||||
|  | 		DataType:     rq.DataType, | ||||||
|  | 		ValueLen:     len(rq.Body), | ||||||
|  | 		SystemEvent:  InvalidSysEvent, | ||||||
|  | 		CollectionId: math.MaxUint64, | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	event.PopulateFieldsBasedOnStreamType(rq, stream.StreamType) | ||||||
|  |  | ||||||
|  | 	// set AckSize for events that need to be acked to DCP, | ||||||
|  | 	// i.e., events with CommandCodes that need to be buffered in DCP | ||||||
|  | 	if _, ok := gomemcached.BufferedCommandCodeMap[rq.Opcode]; ok { | ||||||
|  | 		event.AckSize = uint32(bytesReceivedFromDCP) | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	// 16 LSBits are used by client library to encode vbucket number. | ||||||
|  | 	// 16 MSBits are left for application to multiplex on opaque value. | ||||||
|  | 	event.Opaque = appOpaque(rq.Opaque) | ||||||
|  |  | ||||||
|  | 	if len(rq.Extras) >= uprMutationExtraLen && | ||||||
|  | 		event.Opcode == gomemcached.UPR_MUTATION { | ||||||
|  |  | ||||||
|  | 		event.Seqno = binary.BigEndian.Uint64(rq.Extras[:8]) | ||||||
|  | 		event.RevSeqno = binary.BigEndian.Uint64(rq.Extras[8:16]) | ||||||
|  | 		event.Flags = binary.BigEndian.Uint32(rq.Extras[16:20]) | ||||||
|  | 		event.Expiry = binary.BigEndian.Uint32(rq.Extras[20:24]) | ||||||
|  | 		event.LockTime = binary.BigEndian.Uint32(rq.Extras[24:28]) | ||||||
|  | 		event.MetadataSize = binary.BigEndian.Uint16(rq.Extras[28:30]) | ||||||
|  |  | ||||||
|  | 	} else if len(rq.Extras) >= uprDeletetionWithDeletionTimeExtraLen && | ||||||
|  | 		event.Opcode == gomemcached.UPR_DELETION { | ||||||
|  |  | ||||||
|  | 		event.Seqno = binary.BigEndian.Uint64(rq.Extras[:8]) | ||||||
|  | 		event.RevSeqno = binary.BigEndian.Uint64(rq.Extras[8:16]) | ||||||
|  | 		event.Expiry = binary.BigEndian.Uint32(rq.Extras[16:20]) | ||||||
|  |  | ||||||
|  | 	} else if len(rq.Extras) >= uprDeletetionExtraLen && | ||||||
|  | 		event.Opcode == gomemcached.UPR_DELETION || | ||||||
|  | 		event.Opcode == gomemcached.UPR_EXPIRATION { | ||||||
|  |  | ||||||
|  | 		event.Seqno = binary.BigEndian.Uint64(rq.Extras[:8]) | ||||||
|  | 		event.RevSeqno = binary.BigEndian.Uint64(rq.Extras[8:16]) | ||||||
|  | 		event.MetadataSize = binary.BigEndian.Uint16(rq.Extras[16:18]) | ||||||
|  |  | ||||||
|  | 	} else if len(rq.Extras) >= uprSnapshotExtraLen && | ||||||
|  | 		event.Opcode == gomemcached.UPR_SNAPSHOT { | ||||||
|  |  | ||||||
|  | 		event.SnapstartSeq = binary.BigEndian.Uint64(rq.Extras[:8]) | ||||||
|  | 		event.SnapendSeq = binary.BigEndian.Uint64(rq.Extras[8:16]) | ||||||
|  | 		event.SnapshotType = binary.BigEndian.Uint32(rq.Extras[16:20]) | ||||||
|  | 	} else if event.IsSystemEvent() { | ||||||
|  | 		event.PopulateEvent(rq.Extras) | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	return event | ||||||
|  | } | ||||||
|  |  | ||||||
|  | func (event *UprEvent) PopulateFieldsBasedOnStreamType(rq gomemcached.MCRequest, streamType DcpStreamType) { | ||||||
|  | 	switch streamType { | ||||||
|  | 	case CollectionsNonStreamId: | ||||||
|  | 		switch rq.Opcode { | ||||||
|  | 		// Only these will have CID encoded within the key | ||||||
|  | 		case gomemcached.UPR_MUTATION, | ||||||
|  | 			gomemcached.UPR_DELETION, | ||||||
|  | 			gomemcached.UPR_EXPIRATION: | ||||||
|  | 			uleb128 := Uleb128(rq.Key) | ||||||
|  | 			result, bytesShifted := uleb128.ToUint64(rq.Keylen) | ||||||
|  | 			event.CollectionId = result | ||||||
|  | 			event.Key = rq.Key[bytesShifted:] | ||||||
|  | 		default: | ||||||
|  | 			event.Key = rq.Key | ||||||
|  | 		} | ||||||
|  | 	case CollectionsStreamId: | ||||||
|  | 		// TODO - not implemented | ||||||
|  | 		fallthrough | ||||||
|  | 	case NonCollectionStream: | ||||||
|  | 		// Let default behavior be legacy stream type | ||||||
|  | 		fallthrough | ||||||
|  | 	default: | ||||||
|  | 		event.Key = rq.Key | ||||||
|  | 	} | ||||||
|  | } | ||||||
|  |  | ||||||
|  | func (event *UprEvent) String() string { | ||||||
|  | 	name := gomemcached.CommandNames[event.Opcode] | ||||||
|  | 	if name == "" { | ||||||
|  | 		name = fmt.Sprintf("#%d", event.Opcode) | ||||||
|  | 	} | ||||||
|  | 	return name | ||||||
|  | } | ||||||
|  |  | ||||||
|  | func (event *UprEvent) IsSnappyDataType() bool { | ||||||
|  | 	return event.Opcode == gomemcached.UPR_MUTATION && (event.DataType&SnappyDataType > 0) | ||||||
|  | } | ||||||
|  |  | ||||||
|  | func (event *UprEvent) IsCollectionType() bool { | ||||||
|  | 	return event.IsSystemEvent() || event.CollectionId <= math.MaxUint32 | ||||||
|  | } | ||||||
|  |  | ||||||
|  | func (event *UprEvent) IsSystemEvent() bool { | ||||||
|  | 	return event.Opcode == gomemcached.DCP_SYSTEM_EVENT | ||||||
|  | } | ||||||
|  |  | ||||||
|  | func (event *UprEvent) PopulateEvent(extras []byte) { | ||||||
|  | 	if len(extras) < dcpSystemEventExtraLen { | ||||||
|  | 		// Wrong length, don't parse | ||||||
|  | 		return | ||||||
|  | 	} | ||||||
|  | 	event.Seqno = binary.BigEndian.Uint64(extras[:8]) | ||||||
|  | 	event.SystemEvent = SystemEventType(binary.BigEndian.Uint32(extras[8:12])) | ||||||
|  | 	var versionTemp uint16 = binary.BigEndian.Uint16(extras[12:14]) | ||||||
|  | 	event.SysEventVersion = uint8(versionTemp >> 8) | ||||||
|  | } | ||||||
|  |  | ||||||
|  | func (event *UprEvent) GetSystemEventName() (string, error) { | ||||||
|  | 	switch event.SystemEvent { | ||||||
|  | 	case CollectionCreate: | ||||||
|  | 		fallthrough | ||||||
|  | 	case ScopeCreate: | ||||||
|  | 		return string(event.Key), nil | ||||||
|  | 	default: | ||||||
|  | 		return "", ErrorInvalidOp | ||||||
|  | 	} | ||||||
|  | } | ||||||
|  |  | ||||||
|  | func (event *UprEvent) GetManifestId() (uint64, error) { | ||||||
|  | 	switch event.SystemEvent { | ||||||
|  | 	// Version 0 only checks | ||||||
|  | 	case CollectionChanged: | ||||||
|  | 		fallthrough | ||||||
|  | 	case ScopeDrop: | ||||||
|  | 		fallthrough | ||||||
|  | 	case ScopeCreate: | ||||||
|  | 		fallthrough | ||||||
|  | 	case CollectionDrop: | ||||||
|  | 		if event.SysEventVersion > 0 { | ||||||
|  | 			return 0, ErrorInvalidVersion | ||||||
|  | 		} | ||||||
|  | 		fallthrough | ||||||
|  | 	case CollectionCreate: | ||||||
|  | 		// CollectionCreate supports version 1 | ||||||
|  | 		if event.SysEventVersion > 1 { | ||||||
|  | 			return 0, ErrorInvalidVersion | ||||||
|  | 		} | ||||||
|  | 		if event.ValueLen < 8 { | ||||||
|  | 			return 0, ErrorValueTooShort | ||||||
|  | 		} | ||||||
|  | 		return binary.BigEndian.Uint64(event.Value[0:8]), nil | ||||||
|  | 	default: | ||||||
|  | 		return 0, ErrorInvalidOp | ||||||
|  | 	} | ||||||
|  | } | ||||||
|  |  | ||||||
|  | func (event *UprEvent) GetCollectionId() (uint32, error) { | ||||||
|  | 	switch event.SystemEvent { | ||||||
|  | 	case CollectionDrop: | ||||||
|  | 		if event.SysEventVersion > 0 { | ||||||
|  | 			return 0, ErrorInvalidVersion | ||||||
|  | 		} | ||||||
|  | 		fallthrough | ||||||
|  | 	case CollectionCreate: | ||||||
|  | 		if event.SysEventVersion > 1 { | ||||||
|  | 			return 0, ErrorInvalidVersion | ||||||
|  | 		} | ||||||
|  | 		if event.ValueLen < 16 { | ||||||
|  | 			return 0, ErrorValueTooShort | ||||||
|  | 		} | ||||||
|  | 		return binary.BigEndian.Uint32(event.Value[12:16]), nil | ||||||
|  | 	case CollectionChanged: | ||||||
|  | 		if event.SysEventVersion > 0 { | ||||||
|  | 			return 0, ErrorInvalidVersion | ||||||
|  | 		} | ||||||
|  | 		if event.ValueLen < 12 { | ||||||
|  | 			return 0, ErrorValueTooShort | ||||||
|  | 		} | ||||||
|  | 		return binary.BigEndian.Uint32(event.Value[8:12]), nil | ||||||
|  | 	default: | ||||||
|  | 		return 0, ErrorInvalidOp | ||||||
|  | 	} | ||||||
|  | } | ||||||
|  |  | ||||||
|  | func (event *UprEvent) GetScopeId() (uint32, error) { | ||||||
|  | 	switch event.SystemEvent { | ||||||
|  | 	// version 0 checks | ||||||
|  | 	case ScopeCreate: | ||||||
|  | 		fallthrough | ||||||
|  | 	case ScopeDrop: | ||||||
|  | 		fallthrough | ||||||
|  | 	case CollectionDrop: | ||||||
|  | 		if event.SysEventVersion > 0 { | ||||||
|  | 			return 0, ErrorInvalidVersion | ||||||
|  | 		} | ||||||
|  | 		fallthrough | ||||||
|  | 	case CollectionCreate: | ||||||
|  | 		// CollectionCreate could be either 0 or 1 | ||||||
|  | 		if event.SysEventVersion > 1 { | ||||||
|  | 			return 0, ErrorInvalidVersion | ||||||
|  | 		} | ||||||
|  | 		if event.ValueLen < 12 { | ||||||
|  | 			return 0, ErrorValueTooShort | ||||||
|  | 		} | ||||||
|  | 		return binary.BigEndian.Uint32(event.Value[8:12]), nil | ||||||
|  | 	default: | ||||||
|  | 		return 0, ErrorInvalidOp | ||||||
|  | 	} | ||||||
|  | } | ||||||
|  |  | ||||||
|  | func (event *UprEvent) GetMaxTTL() (uint32, error) { | ||||||
|  | 	switch event.SystemEvent { | ||||||
|  | 	case CollectionCreate: | ||||||
|  | 		if event.SysEventVersion < 1 { | ||||||
|  | 			return 0, ErrorNoMaxTTL | ||||||
|  | 		} | ||||||
|  | 		if event.ValueLen < 20 { | ||||||
|  | 			return 0, ErrorValueTooShort | ||||||
|  | 		} | ||||||
|  | 		return binary.BigEndian.Uint32(event.Value[16:20]), nil | ||||||
|  | 	case CollectionChanged: | ||||||
|  | 		if event.SysEventVersion > 0 { | ||||||
|  | 			return 0, ErrorInvalidVersion | ||||||
|  | 		} | ||||||
|  | 		if event.ValueLen < 16 { | ||||||
|  | 			return 0, ErrorValueTooShort | ||||||
|  | 		} | ||||||
|  | 		return binary.BigEndian.Uint32(event.Value[12:16]), nil | ||||||
|  | 	default: | ||||||
|  | 		return 0, ErrorInvalidOp | ||||||
|  | 	} | ||||||
|  | } | ||||||
|  |  | ||||||
|  | type Uleb128 []byte | ||||||
|  |  | ||||||
|  | func (u Uleb128) ToUint64(cachedLen int) (result uint64, bytesShifted int) { | ||||||
|  | 	var shift uint = 0 | ||||||
|  |  | ||||||
|  | 	for curByte := 0; curByte < cachedLen; curByte++ { | ||||||
|  | 		oneByte := u[curByte] | ||||||
|  | 		last7Bits := 0x7f & oneByte | ||||||
|  | 		result |= uint64(last7Bits) << shift | ||||||
|  | 		bytesShifted++ | ||||||
|  | 		if oneByte&0x80 == 0 { | ||||||
|  | 			break | ||||||
|  | 		} | ||||||
|  | 		shift += 7 | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	return | ||||||
|  | } | ||||||
							
								
								
									
										248
									
								
								vendor/github.com/couchbase/gomemcached/client/upr_feed.go
									
									
									
										generated
									
									
										vendored
									
									
								
							
							
						
						
									
										248
									
								
								vendor/github.com/couchbase/gomemcached/client/upr_feed.go
									
									
									
										generated
									
									
										vendored
									
									
								
							| @@ -19,6 +19,7 @@ const uprMutationExtraLen = 30 | |||||||
| const uprDeletetionExtraLen = 18 | const uprDeletetionExtraLen = 18 | ||||||
| const uprDeletetionWithDeletionTimeExtraLen = 21 | const uprDeletetionWithDeletionTimeExtraLen = 21 | ||||||
| const uprSnapshotExtraLen = 20 | const uprSnapshotExtraLen = 20 | ||||||
|  | const dcpSystemEventExtraLen = 13 | ||||||
| const bufferAckThreshold = 0.2 | const bufferAckThreshold = 0.2 | ||||||
| const opaqueOpen = 0xBEAF0001 | const opaqueOpen = 0xBEAF0001 | ||||||
| const opaqueFailover = 0xDEADBEEF | const opaqueFailover = 0xDEADBEEF | ||||||
| @@ -27,32 +28,6 @@ const uprDefaultNoopInterval = 120 | |||||||
| // Counter on top of opaqueOpen that others can draw from for open and control msgs | // Counter on top of opaqueOpen that others can draw from for open and control msgs | ||||||
| var opaqueOpenCtrlWell uint32 = opaqueOpen | var opaqueOpenCtrlWell uint32 = opaqueOpen | ||||||
|  |  | ||||||
| // UprEvent memcached events for UPR streams. |  | ||||||
| type UprEvent struct { |  | ||||||
| 	Opcode       gomemcached.CommandCode // Type of event |  | ||||||
| 	Status       gomemcached.Status      // Response status |  | ||||||
| 	VBucket      uint16                  // VBucket this event applies to |  | ||||||
| 	DataType     uint8                   // data type |  | ||||||
| 	Opaque       uint16                  // 16 MSB of opaque |  | ||||||
| 	VBuuid       uint64                  // This field is set by downstream |  | ||||||
| 	Flags        uint32                  // Item flags |  | ||||||
| 	Expiry       uint32                  // Item expiration time |  | ||||||
| 	Key, Value   []byte                  // Item key/value |  | ||||||
| 	OldValue     []byte                  // TODO: TBD: old document value |  | ||||||
| 	Cas          uint64                  // CAS value of the item |  | ||||||
| 	Seqno        uint64                  // sequence number of the mutation |  | ||||||
| 	RevSeqno     uint64                  // rev sequence number : deletions |  | ||||||
| 	LockTime     uint32                  // Lock time |  | ||||||
| 	MetadataSize uint16                  // Metadata size |  | ||||||
| 	SnapstartSeq uint64                  // start sequence number of this snapshot |  | ||||||
| 	SnapendSeq   uint64                  // End sequence number of the snapshot |  | ||||||
| 	SnapshotType uint32                  // 0: disk 1: memory |  | ||||||
| 	FailoverLog  *FailoverLog            // Failover log containing vvuid and sequnce number |  | ||||||
| 	Error        error                   // Error value in case of a failure |  | ||||||
| 	ExtMeta      []byte |  | ||||||
| 	AckSize      uint32 // The number of bytes that can be Acked to DCP |  | ||||||
| } |  | ||||||
|  |  | ||||||
| type PriorityType string | type PriorityType string | ||||||
|  |  | ||||||
| // high > medium > disabled > low | // high > medium > disabled > low | ||||||
| @@ -63,13 +38,39 @@ const ( | |||||||
| 	PriorityHigh     PriorityType = "high" | 	PriorityHigh     PriorityType = "high" | ||||||
| ) | ) | ||||||
|  |  | ||||||
|  | type DcpStreamType int32 | ||||||
|  |  | ||||||
|  | var UninitializedStream DcpStreamType = -1 | ||||||
|  |  | ||||||
|  | const ( | ||||||
|  | 	NonCollectionStream    DcpStreamType = 0 | ||||||
|  | 	CollectionsNonStreamId DcpStreamType = iota | ||||||
|  | 	CollectionsStreamId    DcpStreamType = iota | ||||||
|  | ) | ||||||
|  |  | ||||||
|  | func (t DcpStreamType) String() string { | ||||||
|  | 	switch t { | ||||||
|  | 	case UninitializedStream: | ||||||
|  | 		return "Un-Initialized Stream" | ||||||
|  | 	case NonCollectionStream: | ||||||
|  | 		return "Traditional Non-Collection Stream" | ||||||
|  | 	case CollectionsNonStreamId: | ||||||
|  | 		return "Collections Stream without StreamID" | ||||||
|  | 	case CollectionsStreamId: | ||||||
|  | 		return "Collection Stream with StreamID" | ||||||
|  | 	default: | ||||||
|  | 		return "Unknown Stream Type" | ||||||
|  | 	} | ||||||
|  | } | ||||||
|  |  | ||||||
| // UprStream is per stream data structure over an UPR Connection. | // UprStream is per stream data structure over an UPR Connection. | ||||||
| type UprStream struct { | type UprStream struct { | ||||||
| 	Vbucket   uint16 // Vbucket id | 	Vbucket    uint16 // Vbucket id | ||||||
| 	Vbuuid    uint64 // vbucket uuid | 	Vbuuid     uint64 // vbucket uuid | ||||||
| 	StartSeq  uint64 // start sequence number | 	StartSeq   uint64 // start sequence number | ||||||
| 	EndSeq    uint64 // end sequence number | 	EndSeq     uint64 // end sequence number | ||||||
| 	connected bool | 	connected  bool | ||||||
|  | 	StreamType DcpStreamType | ||||||
| } | } | ||||||
|  |  | ||||||
| type FeedState int | type FeedState int | ||||||
| @@ -113,6 +114,7 @@ type UprFeatures struct { | |||||||
| 	IncludeDeletionTime bool | 	IncludeDeletionTime bool | ||||||
| 	DcpPriority         PriorityType | 	DcpPriority         PriorityType | ||||||
| 	EnableExpiry        bool | 	EnableExpiry        bool | ||||||
|  | 	EnableStreamId      bool | ||||||
| } | } | ||||||
|  |  | ||||||
| /** | /** | ||||||
| @@ -274,9 +276,15 @@ type UprFeed struct { | |||||||
| 	// if flag is true, upr feed will use ack from client to determine whether/when to send ack to DCP | 	// if flag is true, upr feed will use ack from client to determine whether/when to send ack to DCP | ||||||
| 	// if flag is false, upr feed will track how many bytes it has sent to client | 	// if flag is false, upr feed will track how many bytes it has sent to client | ||||||
| 	// and use that to determine whether/when to send ack to DCP | 	// and use that to determine whether/when to send ack to DCP | ||||||
| 	ackByClient bool | 	ackByClient       bool | ||||||
| 	feedState   FeedState | 	feedState         FeedState | ||||||
| 	muFeedState sync.RWMutex | 	muFeedState       sync.RWMutex | ||||||
|  | 	activatedFeatures UprFeatures | ||||||
|  | 	collectionEnabled bool // This is needed separately because parsing depends on this | ||||||
|  | 	// DCP StreamID allows multiple filtered collection streams to share a single DCP Stream | ||||||
|  | 	// It is not allowed once a regular/legacy stream was started originally | ||||||
|  | 	streamsType        DcpStreamType | ||||||
|  | 	initStreamTypeOnce sync.Once | ||||||
| } | } | ||||||
|  |  | ||||||
| // Exported interface - to allow for mocking | // Exported interface - to allow for mocking | ||||||
| @@ -296,6 +304,9 @@ type UprFeedIface interface { | |||||||
| 	UprRequestStream(vbno, opaqueMSB uint16, flags uint32, vuuid, startSequence, endSequence, snapStart, snapEnd uint64) error | 	UprRequestStream(vbno, opaqueMSB uint16, flags uint32, vuuid, startSequence, endSequence, snapStart, snapEnd uint64) error | ||||||
| 	// Set DCP priority on an existing DCP connection. The command is sent asynchronously without waiting for a response | 	// Set DCP priority on an existing DCP connection. The command is sent asynchronously without waiting for a response | ||||||
| 	SetPriorityAsync(p PriorityType) error | 	SetPriorityAsync(p PriorityType) error | ||||||
|  |  | ||||||
|  | 	// Various Collection-Type RequestStreams | ||||||
|  | 	UprRequestCollectionsStream(vbno, opaqueMSB uint16, flags uint32, vbuuid, startSeq, endSeq, snapStart, snapEnd uint64, filter *CollectionsFilter) error | ||||||
| } | } | ||||||
|  |  | ||||||
| type UprStats struct { | type UprStats struct { | ||||||
| @@ -305,9 +316,6 @@ type UprStats struct { | |||||||
| 	TotalSnapShot      uint64 | 	TotalSnapShot      uint64 | ||||||
| } | } | ||||||
|  |  | ||||||
| // FailoverLog containing vvuid and sequnce number |  | ||||||
| type FailoverLog [][2]uint64 |  | ||||||
|  |  | ||||||
| // error codes | // error codes | ||||||
| var ErrorInvalidLog = errors.New("couchbase.errorInvalidLog") | var ErrorInvalidLog = errors.New("couchbase.errorInvalidLog") | ||||||
|  |  | ||||||
| @@ -320,76 +328,6 @@ func (flogp *FailoverLog) Latest() (vbuuid, seqno uint64, err error) { | |||||||
| 	return vbuuid, seqno, ErrorInvalidLog | 	return vbuuid, seqno, ErrorInvalidLog | ||||||
| } | } | ||||||
|  |  | ||||||
| func makeUprEvent(rq gomemcached.MCRequest, stream *UprStream, bytesReceivedFromDCP int) *UprEvent { |  | ||||||
| 	event := &UprEvent{ |  | ||||||
| 		Opcode:   rq.Opcode, |  | ||||||
| 		VBucket:  stream.Vbucket, |  | ||||||
| 		VBuuid:   stream.Vbuuid, |  | ||||||
| 		Key:      rq.Key, |  | ||||||
| 		Value:    rq.Body, |  | ||||||
| 		Cas:      rq.Cas, |  | ||||||
| 		ExtMeta:  rq.ExtMeta, |  | ||||||
| 		DataType: rq.DataType, |  | ||||||
| 	} |  | ||||||
|  |  | ||||||
| 	// set AckSize for events that need to be acked to DCP, |  | ||||||
| 	// i.e., events with CommandCodes that need to be buffered in DCP |  | ||||||
| 	if _, ok := gomemcached.BufferedCommandCodeMap[rq.Opcode]; ok { |  | ||||||
| 		event.AckSize = uint32(bytesReceivedFromDCP) |  | ||||||
| 	} |  | ||||||
|  |  | ||||||
| 	// 16 LSBits are used by client library to encode vbucket number. |  | ||||||
| 	// 16 MSBits are left for application to multiplex on opaque value. |  | ||||||
| 	event.Opaque = appOpaque(rq.Opaque) |  | ||||||
|  |  | ||||||
| 	if len(rq.Extras) >= uprMutationExtraLen && |  | ||||||
| 		event.Opcode == gomemcached.UPR_MUTATION { |  | ||||||
|  |  | ||||||
| 		event.Seqno = binary.BigEndian.Uint64(rq.Extras[:8]) |  | ||||||
| 		event.RevSeqno = binary.BigEndian.Uint64(rq.Extras[8:16]) |  | ||||||
| 		event.Flags = binary.BigEndian.Uint32(rq.Extras[16:20]) |  | ||||||
| 		event.Expiry = binary.BigEndian.Uint32(rq.Extras[20:24]) |  | ||||||
| 		event.LockTime = binary.BigEndian.Uint32(rq.Extras[24:28]) |  | ||||||
| 		event.MetadataSize = binary.BigEndian.Uint16(rq.Extras[28:30]) |  | ||||||
|  |  | ||||||
| 	} else if len(rq.Extras) >= uprDeletetionWithDeletionTimeExtraLen && |  | ||||||
| 		event.Opcode == gomemcached.UPR_DELETION { |  | ||||||
|  |  | ||||||
| 		event.Seqno = binary.BigEndian.Uint64(rq.Extras[:8]) |  | ||||||
| 		event.RevSeqno = binary.BigEndian.Uint64(rq.Extras[8:16]) |  | ||||||
| 		event.Expiry = binary.BigEndian.Uint32(rq.Extras[16:20]) |  | ||||||
|  |  | ||||||
| 	} else if len(rq.Extras) >= uprDeletetionExtraLen && |  | ||||||
| 		event.Opcode == gomemcached.UPR_DELETION || |  | ||||||
| 		event.Opcode == gomemcached.UPR_EXPIRATION { |  | ||||||
|  |  | ||||||
| 		event.Seqno = binary.BigEndian.Uint64(rq.Extras[:8]) |  | ||||||
| 		event.RevSeqno = binary.BigEndian.Uint64(rq.Extras[8:16]) |  | ||||||
| 		event.MetadataSize = binary.BigEndian.Uint16(rq.Extras[16:18]) |  | ||||||
|  |  | ||||||
| 	} else if len(rq.Extras) >= uprSnapshotExtraLen && |  | ||||||
| 		event.Opcode == gomemcached.UPR_SNAPSHOT { |  | ||||||
|  |  | ||||||
| 		event.SnapstartSeq = binary.BigEndian.Uint64(rq.Extras[:8]) |  | ||||||
| 		event.SnapendSeq = binary.BigEndian.Uint64(rq.Extras[8:16]) |  | ||||||
| 		event.SnapshotType = binary.BigEndian.Uint32(rq.Extras[16:20]) |  | ||||||
| 	} |  | ||||||
|  |  | ||||||
| 	return event |  | ||||||
| } |  | ||||||
|  |  | ||||||
| func (event *UprEvent) String() string { |  | ||||||
| 	name := gomemcached.CommandNames[event.Opcode] |  | ||||||
| 	if name == "" { |  | ||||||
| 		name = fmt.Sprintf("#%d", event.Opcode) |  | ||||||
| 	} |  | ||||||
| 	return name |  | ||||||
| } |  | ||||||
|  |  | ||||||
| func (event *UprEvent) IsSnappyDataType() bool { |  | ||||||
| 	return event.Opcode == gomemcached.UPR_MUTATION && (event.DataType&SnappyDataType > 0) |  | ||||||
| } |  | ||||||
|  |  | ||||||
| func (feed *UprFeed) sendCommands(mc *Client) { | func (feed *UprFeed) sendCommands(mc *Client) { | ||||||
| 	transmitCh := feed.transmitCh | 	transmitCh := feed.transmitCh | ||||||
| 	transmitCl := feed.transmitCl | 	transmitCl := feed.transmitCl | ||||||
| @@ -420,6 +358,10 @@ func (feed *UprFeed) activateStream(vbno, opaque uint16, stream *UprStream) erro | |||||||
| 	feed.muVbstreams.Lock() | 	feed.muVbstreams.Lock() | ||||||
| 	defer feed.muVbstreams.Unlock() | 	defer feed.muVbstreams.Unlock() | ||||||
|  |  | ||||||
|  | 	if feed.collectionEnabled { | ||||||
|  | 		stream.StreamType = feed.streamsType | ||||||
|  | 	} | ||||||
|  |  | ||||||
| 	// Set this stream as the officially connected stream for this vb | 	// Set this stream as the officially connected stream for this vb | ||||||
| 	stream.connected = true | 	stream.connected = true | ||||||
| 	feed.vbstreams[vbno] = stream | 	feed.vbstreams[vbno] = stream | ||||||
| @@ -440,14 +382,15 @@ func (mc *Client) NewUprFeed() (*UprFeed, error) { | |||||||
| } | } | ||||||
|  |  | ||||||
| func (mc *Client) NewUprFeedWithConfig(ackByClient bool) (*UprFeed, error) { | func (mc *Client) NewUprFeedWithConfig(ackByClient bool) (*UprFeed, error) { | ||||||
|  |  | ||||||
| 	feed := &UprFeed{ | 	feed := &UprFeed{ | ||||||
| 		conn:        mc, | 		conn:              mc, | ||||||
| 		closer:      make(chan bool, 1), | 		closer:            make(chan bool, 1), | ||||||
| 		vbstreams:   make(map[uint16]*UprStream), | 		vbstreams:         make(map[uint16]*UprStream), | ||||||
| 		transmitCh:  make(chan *gomemcached.MCRequest), | 		transmitCh:        make(chan *gomemcached.MCRequest), | ||||||
| 		transmitCl:  make(chan bool), | 		transmitCl:        make(chan bool), | ||||||
| 		ackByClient: ackByClient, | 		ackByClient:       ackByClient, | ||||||
|  | 		collectionEnabled: mc.CollectionEnabled(), | ||||||
|  | 		streamsType:       UninitializedStream, | ||||||
| 	} | 	} | ||||||
|  |  | ||||||
| 	feed.negotiator.initialize() | 	feed.negotiator.initialize() | ||||||
| @@ -642,7 +585,22 @@ func (feed *UprFeed) uprOpen(name string, sequence uint32, bufSize uint32, featu | |||||||
| 		activatedFeatures.EnableExpiry = true | 		activatedFeatures.EnableExpiry = true | ||||||
| 	} | 	} | ||||||
|  |  | ||||||
|  | 	if features.EnableStreamId { | ||||||
|  | 		rq := &gomemcached.MCRequest{ | ||||||
|  | 			Opcode: gomemcached.UPR_CONTROL, | ||||||
|  | 			Key:    []byte("enable_stream_id"), | ||||||
|  | 			Body:   []byte("true"), | ||||||
|  | 			Opaque: getUprOpenCtrlOpaque(), | ||||||
|  | 		} | ||||||
|  | 		err = sendMcRequestSync(feed.conn, rq) | ||||||
|  | 		if err != nil { | ||||||
|  | 			return | ||||||
|  | 		} | ||||||
|  | 		activatedFeatures.EnableStreamId = true | ||||||
|  | 	} | ||||||
|  |  | ||||||
| 	// everything is ok so far, set upr feed to open state | 	// everything is ok so far, set upr feed to open state | ||||||
|  | 	feed.activatedFeatures = activatedFeatures | ||||||
| 	feed.setOpen() | 	feed.setOpen() | ||||||
| 	return | 	return | ||||||
| } | } | ||||||
| @@ -689,10 +647,60 @@ func (mc *Client) UprGetFailoverLog( | |||||||
| func (feed *UprFeed) UprRequestStream(vbno, opaqueMSB uint16, flags uint32, | func (feed *UprFeed) UprRequestStream(vbno, opaqueMSB uint16, flags uint32, | ||||||
| 	vuuid, startSequence, endSequence, snapStart, snapEnd uint64) error { | 	vuuid, startSequence, endSequence, snapStart, snapEnd uint64) error { | ||||||
|  |  | ||||||
|  | 	return feed.UprRequestCollectionsStream(vbno, opaqueMSB, flags, vuuid, startSequence, endSequence, snapStart, snapEnd, nil) | ||||||
|  | } | ||||||
|  |  | ||||||
|  | func (feed *UprFeed) initStreamType(filter *CollectionsFilter) (err error) { | ||||||
|  | 	if filter != nil && filter.UseStreamId && !feed.activatedFeatures.EnableStreamId { | ||||||
|  | 		err = fmt.Errorf("Cannot use streamID based filter if the feed was not started with the streamID feature") | ||||||
|  | 		return | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	streamInitFunc := func() { | ||||||
|  | 		if feed.streamsType != UninitializedStream { | ||||||
|  | 			// Shouldn't happen | ||||||
|  | 			err = fmt.Errorf("The current feed has already been started in %v mode", feed.streamsType.String()) | ||||||
|  | 		} else { | ||||||
|  | 			if !feed.collectionEnabled { | ||||||
|  | 				feed.streamsType = NonCollectionStream | ||||||
|  | 			} else { | ||||||
|  | 				if filter != nil && filter.UseStreamId { | ||||||
|  | 					feed.streamsType = CollectionsStreamId | ||||||
|  | 				} else { | ||||||
|  | 					feed.streamsType = CollectionsNonStreamId | ||||||
|  | 				} | ||||||
|  | 			} | ||||||
|  | 		} | ||||||
|  | 	} | ||||||
|  | 	feed.initStreamTypeOnce.Do(streamInitFunc) | ||||||
|  | 	return | ||||||
|  | } | ||||||
|  |  | ||||||
|  | func (feed *UprFeed) UprRequestCollectionsStream(vbno, opaqueMSB uint16, flags uint32, | ||||||
|  | 	vbuuid, startSequence, endSequence, snapStart, snapEnd uint64, filter *CollectionsFilter) error { | ||||||
|  |  | ||||||
|  | 	err := feed.initStreamType(filter) | ||||||
|  | 	if err != nil { | ||||||
|  | 		return err | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	var mcRequestBody []byte | ||||||
|  | 	if filter != nil { | ||||||
|  | 		err = filter.IsValid() | ||||||
|  | 		if err != nil { | ||||||
|  | 			return err | ||||||
|  | 		} | ||||||
|  | 		mcRequestBody, err = filter.ToStreamReqBody() | ||||||
|  | 		if err != nil { | ||||||
|  | 			return err | ||||||
|  | 		} | ||||||
|  | 	} | ||||||
|  |  | ||||||
| 	rq := &gomemcached.MCRequest{ | 	rq := &gomemcached.MCRequest{ | ||||||
| 		Opcode:  gomemcached.UPR_STREAMREQ, | 		Opcode:  gomemcached.UPR_STREAMREQ, | ||||||
| 		VBucket: vbno, | 		VBucket: vbno, | ||||||
| 		Opaque:  composeOpaque(vbno, opaqueMSB), | 		Opaque:  composeOpaque(vbno, opaqueMSB), | ||||||
|  | 		Body:    mcRequestBody, | ||||||
| 	} | 	} | ||||||
|  |  | ||||||
| 	rq.Extras = make([]byte, 48) // #Extras | 	rq.Extras = make([]byte, 48) // #Extras | ||||||
| @@ -700,15 +708,15 @@ func (feed *UprFeed) UprRequestStream(vbno, opaqueMSB uint16, flags uint32, | |||||||
| 	binary.BigEndian.PutUint32(rq.Extras[4:8], uint32(0)) | 	binary.BigEndian.PutUint32(rq.Extras[4:8], uint32(0)) | ||||||
| 	binary.BigEndian.PutUint64(rq.Extras[8:16], startSequence) | 	binary.BigEndian.PutUint64(rq.Extras[8:16], startSequence) | ||||||
| 	binary.BigEndian.PutUint64(rq.Extras[16:24], endSequence) | 	binary.BigEndian.PutUint64(rq.Extras[16:24], endSequence) | ||||||
| 	binary.BigEndian.PutUint64(rq.Extras[24:32], vuuid) | 	binary.BigEndian.PutUint64(rq.Extras[24:32], vbuuid) | ||||||
| 	binary.BigEndian.PutUint64(rq.Extras[32:40], snapStart) | 	binary.BigEndian.PutUint64(rq.Extras[32:40], snapStart) | ||||||
| 	binary.BigEndian.PutUint64(rq.Extras[40:48], snapEnd) | 	binary.BigEndian.PutUint64(rq.Extras[40:48], snapEnd) | ||||||
|  |  | ||||||
| 	feed.negotiator.registerRequest(vbno, opaqueMSB, vuuid, startSequence, endSequence) | 	feed.negotiator.registerRequest(vbno, opaqueMSB, vbuuid, startSequence, endSequence) | ||||||
| 	// Any client that has ever called this method, regardless of return code, | 	// Any client that has ever called this method, regardless of return code, | ||||||
| 	// should expect a potential UPR_CLOSESTREAM message due to this new map entry prior to Transmit. | 	// should expect a potential UPR_CLOSESTREAM message due to this new map entry prior to Transmit. | ||||||
|  |  | ||||||
| 	if err := feed.conn.Transmit(rq); err != nil { | 	if err = feed.conn.Transmit(rq); err != nil { | ||||||
| 		logging.Errorf("Error in StreamRequest %s", err.Error()) | 		logging.Errorf("Error in StreamRequest %s", err.Error()) | ||||||
| 		// If an error occurs during transmit, then the UPRFeed will keep the stream | 		// If an error occurs during transmit, then the UPRFeed will keep the stream | ||||||
| 		// in the vbstreams map. This is to prevent nil lookup from any previously | 		// in the vbstreams map. This is to prevent nil lookup from any previously | ||||||
| @@ -973,6 +981,12 @@ loop: | |||||||
| 					if err := feed.conn.TransmitResponse(noop); err != nil { | 					if err := feed.conn.TransmitResponse(noop); err != nil { | ||||||
| 						logging.Warnf("failed to transmit command %s. Error %s", noop.Opcode.String(), err.Error()) | 						logging.Warnf("failed to transmit command %s. Error %s", noop.Opcode.String(), err.Error()) | ||||||
| 					} | 					} | ||||||
|  | 				case gomemcached.DCP_SYSTEM_EVENT: | ||||||
|  | 					if stream == nil { | ||||||
|  | 						logging.Infof("Stream not found for vb %d: %#v", vb, pkt) | ||||||
|  | 						break loop | ||||||
|  | 					} | ||||||
|  | 					event = makeUprEvent(pkt, stream, bytes) | ||||||
| 				default: | 				default: | ||||||
| 					logging.Infof("Recived an unknown response for vbucket %d", vb) | 					logging.Infof("Recived an unknown response for vbucket %d", vb) | ||||||
| 				} | 				} | ||||||
|   | |||||||
							
								
								
									
										381
									
								
								vendor/github.com/couchbase/gomemcached/flexibleFraming.go
									
									
									
										generated
									
									
										vendored
									
									
										Normal file
									
								
							
							
						
						
									
										381
									
								
								vendor/github.com/couchbase/gomemcached/flexibleFraming.go
									
									
									
										generated
									
									
										vendored
									
									
										Normal file
									
								
							| @@ -0,0 +1,381 @@ | |||||||
|  | package gomemcached | ||||||
|  |  | ||||||
|  | import ( | ||||||
|  | 	"encoding/binary" | ||||||
|  | 	"fmt" | ||||||
|  | ) | ||||||
|  |  | ||||||
|  | type FrameObjType int | ||||||
|  |  | ||||||
|  | const ( | ||||||
|  | 	FrameBarrier     FrameObjType = iota | ||||||
|  | 	FrameDurability  FrameObjType = iota | ||||||
|  | 	FrameDcpStreamId FrameObjType = iota | ||||||
|  | 	FrameOpenTracing FrameObjType = iota | ||||||
|  | ) | ||||||
|  |  | ||||||
|  | type FrameInfo struct { | ||||||
|  | 	ObjId   FrameObjType | ||||||
|  | 	ObjLen  int | ||||||
|  | 	ObjData []byte | ||||||
|  | } | ||||||
|  |  | ||||||
|  | var ErrorInvalidOp error = fmt.Errorf("Specified method is not applicable") | ||||||
|  | var ErrorObjLenNotMatch error = fmt.Errorf("Object length does not match data") | ||||||
|  |  | ||||||
|  | func (f *FrameInfo) Validate() error { | ||||||
|  | 	switch f.ObjId { | ||||||
|  | 	case FrameBarrier: | ||||||
|  | 		if f.ObjLen != 0 { | ||||||
|  | 			return fmt.Errorf("Invalid FrameBarrier - length is %v\n", f.ObjLen) | ||||||
|  | 		} else if f.ObjLen != len(f.ObjData) { | ||||||
|  | 			return ErrorObjLenNotMatch | ||||||
|  | 		} | ||||||
|  | 	case FrameDurability: | ||||||
|  | 		if f.ObjLen != 1 && f.ObjLen != 3 { | ||||||
|  | 			return fmt.Errorf("Invalid FrameDurability - length is %v\n", f.ObjLen) | ||||||
|  | 		} else if f.ObjLen != len(f.ObjData) { | ||||||
|  | 			return ErrorObjLenNotMatch | ||||||
|  | 		} | ||||||
|  | 	case FrameDcpStreamId: | ||||||
|  | 		if f.ObjLen != 2 { | ||||||
|  | 			return fmt.Errorf("Invalid FrameDcpStreamId - length is %v\n", f.ObjLen) | ||||||
|  | 		} else if f.ObjLen != len(f.ObjData) { | ||||||
|  | 			return ErrorObjLenNotMatch | ||||||
|  | 		} | ||||||
|  | 	case FrameOpenTracing: | ||||||
|  | 		if f.ObjLen == 0 { | ||||||
|  | 			return fmt.Errorf("Invalid FrameOpenTracing - length must be > 0") | ||||||
|  | 		} else if f.ObjLen != len(f.ObjData) { | ||||||
|  | 			return ErrorObjLenNotMatch | ||||||
|  | 		} | ||||||
|  | 	default: | ||||||
|  | 		return fmt.Errorf("Unknown FrameInfo type") | ||||||
|  | 	} | ||||||
|  | 	return nil | ||||||
|  | } | ||||||
|  |  | ||||||
|  | func (f *FrameInfo) GetStreamId() (uint16, error) { | ||||||
|  | 	if f.ObjId != FrameDcpStreamId { | ||||||
|  | 		return 0, ErrorInvalidOp | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	var output uint16 | ||||||
|  | 	output = uint16(f.ObjData[0]) | ||||||
|  | 	output = output << 8 | ||||||
|  | 	output |= uint16(f.ObjData[1]) | ||||||
|  | 	return output, nil | ||||||
|  | } | ||||||
|  |  | ||||||
|  | type DurabilityLvl uint8 | ||||||
|  |  | ||||||
|  | const ( | ||||||
|  | 	DuraInvalid                    DurabilityLvl = iota // Not used (0x0) | ||||||
|  | 	DuraMajority                   DurabilityLvl = iota // (0x01) | ||||||
|  | 	DuraMajorityAndPersistOnMaster DurabilityLvl = iota // (0x02) | ||||||
|  | 	DuraPersistToMajority          DurabilityLvl = iota // (0x03) | ||||||
|  | ) | ||||||
|  |  | ||||||
|  | func (f *FrameInfo) GetDurabilityRequirements() (lvl DurabilityLvl, timeoutProvided bool, timeoutMs uint16, err error) { | ||||||
|  | 	if f.ObjId != FrameDurability { | ||||||
|  | 		err = ErrorInvalidOp | ||||||
|  | 		return | ||||||
|  | 	} | ||||||
|  | 	if f.ObjLen != 1 && f.ObjLen != 3 { | ||||||
|  | 		err = ErrorObjLenNotMatch | ||||||
|  | 		return | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	lvl = DurabilityLvl(uint8(f.ObjData[0])) | ||||||
|  |  | ||||||
|  | 	if f.ObjLen == 3 { | ||||||
|  | 		timeoutProvided = true | ||||||
|  | 		timeoutMs = binary.BigEndian.Uint16(f.ObjData[1:2]) | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	return | ||||||
|  | } | ||||||
|  |  | ||||||
|  | func incrementMarker(bitsToBeIncremented, byteIncrementCnt *int, framingElen, curObjIdx int) (int, error) { | ||||||
|  | 	for *bitsToBeIncremented >= 8 { | ||||||
|  | 		*byteIncrementCnt++ | ||||||
|  | 		*bitsToBeIncremented -= 8 | ||||||
|  | 	} | ||||||
|  | 	marker := curObjIdx + *byteIncrementCnt | ||||||
|  | 	if marker > framingElen { | ||||||
|  | 		return -1, fmt.Errorf("Out of bounds") | ||||||
|  | 	} | ||||||
|  | 	return marker, nil | ||||||
|  | } | ||||||
|  |  | ||||||
|  | // Right now, halfByteRemaining will always be false, because ObjID and Len haven't gotten that large yet | ||||||
|  | func (f *FrameInfo) Bytes() (output []byte, halfByteRemaining bool) { | ||||||
|  | 	// ObjIdentifier - 4 bits + ObjLength - 4 bits | ||||||
|  | 	var idAndLen uint8 | ||||||
|  | 	idAndLen |= uint8(f.ObjId) << 4 | ||||||
|  | 	idAndLen |= uint8(f.ObjLen) | ||||||
|  | 	output = append(output, byte(idAndLen)) | ||||||
|  |  | ||||||
|  | 	// Rest is Data | ||||||
|  | 	output = append(output, f.ObjData...) | ||||||
|  | 	return | ||||||
|  | } | ||||||
|  |  | ||||||
|  | func parseFrameInfoObjects(buf []byte, framingElen int) (objs []FrameInfo, err error, halfByteRemaining bool) { | ||||||
|  | 	var curObjIdx int | ||||||
|  | 	var byteIncrementCnt int | ||||||
|  | 	var bitsToBeIncremented int | ||||||
|  | 	var marker int | ||||||
|  |  | ||||||
|  | 	// Parse frameInfo objects | ||||||
|  | 	for curObjIdx = 0; curObjIdx < framingElen; curObjIdx += byteIncrementCnt { | ||||||
|  | 		byteIncrementCnt = 0 | ||||||
|  | 		var oneFrameObj FrameInfo | ||||||
|  |  | ||||||
|  | 		// First get the objId | ||||||
|  | 		// ------------------------- | ||||||
|  | 		var objId int | ||||||
|  | 		var objHeader uint8 = buf[curObjIdx] | ||||||
|  | 		var objIdentifierRaw uint8 | ||||||
|  | 		if bitsToBeIncremented == 0 { | ||||||
|  | 			// ObjHeader | ||||||
|  | 			// 0 1 2 3 4 5 6 7 | ||||||
|  | 			// ^-----^ | ||||||
|  | 			// ObjIdentifierRaw | ||||||
|  | 			objIdentifierRaw = (objHeader & 0xf0) >> 4 | ||||||
|  | 		} else { | ||||||
|  | 			// ObjHeader | ||||||
|  | 			// 0 1 2 3 4 5 6 7 | ||||||
|  | 			//         ^-----^ | ||||||
|  | 			//           ObjIdentifierRaw | ||||||
|  | 			objIdentifierRaw = (objHeader & 0x0f) | ||||||
|  | 		} | ||||||
|  | 		bitsToBeIncremented += 4 | ||||||
|  |  | ||||||
|  | 		marker, err = incrementMarker(&bitsToBeIncremented, &byteIncrementCnt, framingElen, curObjIdx) | ||||||
|  | 		if err != nil { | ||||||
|  | 			return | ||||||
|  | 		} | ||||||
|  |  | ||||||
|  | 		// Value is 0-14 | ||||||
|  | 		objId = int(objIdentifierRaw & 0xe) | ||||||
|  | 		// If bit 15 is set, ID is 15 + value of next byte | ||||||
|  | 		if objIdentifierRaw&0x1 > 0 { | ||||||
|  | 			if bitsToBeIncremented > 0 { | ||||||
|  | 				// ObjHeader | ||||||
|  | 				// 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 | ||||||
|  | 				// ^-----^ ^---------------^ | ||||||
|  | 				// ObjId1    Extension | ||||||
|  | 				// ^ marker | ||||||
|  | 				buffer := uint16(buf[marker]) | ||||||
|  | 				buffer = buffer << 8 | ||||||
|  | 				buffer |= uint16(buf[marker+1]) | ||||||
|  | 				var extension uint8 = uint8(buffer & 0xff0 >> 4) | ||||||
|  | 				objId += int(extension) | ||||||
|  | 			} else { | ||||||
|  | 				// ObjHeader | ||||||
|  | 				// 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 | ||||||
|  | 				//         ^-----^ ^-------------------^ | ||||||
|  | 				//          ObjId1    extension | ||||||
|  | 				//                 ^ marker | ||||||
|  | 				var extension uint8 = uint8(buf[marker]) | ||||||
|  | 				objId += int(extension) | ||||||
|  | 			} | ||||||
|  | 			bitsToBeIncremented += 8 | ||||||
|  | 		} | ||||||
|  |  | ||||||
|  | 		marker, err = incrementMarker(&bitsToBeIncremented, &byteIncrementCnt, framingElen, curObjIdx) | ||||||
|  | 		if err != nil { | ||||||
|  | 			return | ||||||
|  | 		} | ||||||
|  | 		oneFrameObj.ObjId = FrameObjType(objId) | ||||||
|  |  | ||||||
|  | 		// Then get the obj length | ||||||
|  | 		// ------------------------- | ||||||
|  | 		var objLenRaw uint8 | ||||||
|  | 		var objLen int | ||||||
|  | 		if bitsToBeIncremented > 0 { | ||||||
|  | 			// ObjHeader | ||||||
|  | 			// 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 | ||||||
|  | 			//                 ^         ^---------^ | ||||||
|  | 			//                 marker       objLen | ||||||
|  | 			objLenRaw = uint8(buf[marker]) & 0x0f | ||||||
|  | 		} else { | ||||||
|  | 			// ObjHeader | ||||||
|  | 			// 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 | ||||||
|  | 			//                                        ^--------^ | ||||||
|  | 			//                                          objLen | ||||||
|  | 			//                                        ^ marker | ||||||
|  | 			objLenRaw = uint8(buf[marker]) & 0xf0 >> 4 | ||||||
|  | 		} | ||||||
|  | 		bitsToBeIncremented += 4 | ||||||
|  |  | ||||||
|  | 		marker, err = incrementMarker(&bitsToBeIncremented, &byteIncrementCnt, framingElen, curObjIdx) | ||||||
|  | 		if err != nil { | ||||||
|  | 			return | ||||||
|  | 		} | ||||||
|  |  | ||||||
|  | 		// Length is 0-14 | ||||||
|  | 		objLen = int(objLenRaw & 0xe) | ||||||
|  | 		// If bit 15 is set, lenghth is 15 + value of next byte | ||||||
|  | 		if objLenRaw&0x1 > 0 { | ||||||
|  | 			if bitsToBeIncremented == 0 { | ||||||
|  | 				// ObjHeader | ||||||
|  | 				// 12 13 14 15 16 17 18 19 20 21 22 23 | ||||||
|  | 				// ^---------^ ^--------------------^ | ||||||
|  | 				//   objLen        extension | ||||||
|  | 				//             ^ marker | ||||||
|  | 				var extension uint8 = uint8(buf[marker]) | ||||||
|  | 				objLen += int(extension) | ||||||
|  | 			} else { | ||||||
|  | 				// ObjHeader | ||||||
|  | 				// 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 | ||||||
|  | 				// ^--------^  ^---------------------^ | ||||||
|  | 				//  objLen          extension | ||||||
|  | 				// ^ marker				var buffer uint16 | ||||||
|  | 				buffer := uint16(buf[marker]) | ||||||
|  | 				buffer = buffer << 8 | ||||||
|  | 				buffer |= uint16(buf[marker+1]) | ||||||
|  | 				var extension uint8 = uint8(buffer & 0xff0 >> 4) | ||||||
|  | 				objLen += int(extension) | ||||||
|  | 			} | ||||||
|  | 			bitsToBeIncremented += 8 | ||||||
|  | 		} | ||||||
|  |  | ||||||
|  | 		marker, err = incrementMarker(&bitsToBeIncremented, &byteIncrementCnt, framingElen, curObjIdx) | ||||||
|  | 		if err != nil { | ||||||
|  | 			return | ||||||
|  | 		} | ||||||
|  | 		oneFrameObj.ObjLen = objLen | ||||||
|  |  | ||||||
|  | 		// The rest is N-bytes of data based on the length | ||||||
|  | 		if bitsToBeIncremented == 0 { | ||||||
|  | 			// No weird alignment needed | ||||||
|  | 			oneFrameObj.ObjData = buf[marker : marker+objLen] | ||||||
|  | 		} else { | ||||||
|  | 			// 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 | ||||||
|  | 			// ^--------^  ^---------------------^ ^---------> | ||||||
|  | 			//  objLen          extension            data | ||||||
|  | 			//                          ^ marker | ||||||
|  | 			oneFrameObj.ObjData = ShiftByteSliceLeft4Bits(buf[marker : marker+objLen+1]) | ||||||
|  | 		} | ||||||
|  | 		err = oneFrameObj.Validate() | ||||||
|  | 		if err != nil { | ||||||
|  | 			return | ||||||
|  | 		} | ||||||
|  | 		objs = append(objs, oneFrameObj) | ||||||
|  |  | ||||||
|  | 		bitsToBeIncremented += 8 * objLen | ||||||
|  | 		marker, err = incrementMarker(&bitsToBeIncremented, &byteIncrementCnt, framingElen, curObjIdx) | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	if bitsToBeIncremented > 0 { | ||||||
|  | 		halfByteRemaining = true | ||||||
|  | 	} | ||||||
|  | 	return | ||||||
|  | } | ||||||
|  |  | ||||||
|  | func ShiftByteSliceLeft4Bits(slice []byte) (replacement []byte) { | ||||||
|  | 	var buffer uint16 | ||||||
|  | 	var i int | ||||||
|  | 	sliceLen := len(slice) | ||||||
|  |  | ||||||
|  | 	if sliceLen < 2 { | ||||||
|  | 		// Let's not shift less than 16 bits | ||||||
|  | 		return | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	replacement = make([]byte, sliceLen, cap(slice)) | ||||||
|  |  | ||||||
|  | 	for i = 0; i < sliceLen-1; i++ { | ||||||
|  | 		// 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 | ||||||
|  | 		// ^-----^ ^---------------^ ^----------- | ||||||
|  | 		// garbage   data byte 0       data byte 1 | ||||||
|  | 		buffer = uint16(slice[i]) | ||||||
|  | 		buffer = buffer << 8 | ||||||
|  | 		buffer |= uint16(slice[i+1]) | ||||||
|  | 		replacement[i] = uint8(buffer & 0xff0 >> 4) | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	if i < sliceLen { | ||||||
|  | 		lastByte := slice[sliceLen-1] | ||||||
|  | 		lastByte = lastByte << 4 | ||||||
|  | 		replacement[i] = lastByte | ||||||
|  | 	} | ||||||
|  | 	return | ||||||
|  | } | ||||||
|  |  | ||||||
|  | // The following is used to theoretically support frameInfo ObjID extensions | ||||||
|  | // for completeness, but they are not very efficient though | ||||||
|  | func ShiftByteSliceRight4Bits(slice []byte) (replacement []byte) { | ||||||
|  | 	var buffer uint16 | ||||||
|  | 	var i int | ||||||
|  | 	var leftovers uint8 // 4 bits only | ||||||
|  | 	var replacementUnit uint16 | ||||||
|  | 	var first bool = true | ||||||
|  | 	var firstLeftovers uint8 | ||||||
|  | 	var lastLeftovers uint8 | ||||||
|  | 	sliceLen := len(slice) | ||||||
|  |  | ||||||
|  | 	if sliceLen < 2 { | ||||||
|  | 		// Let's not shift less than 16 bits | ||||||
|  | 		return | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	if slice[sliceLen-1]&0xf == 0 { | ||||||
|  | 		replacement = make([]byte, sliceLen, cap(slice)) | ||||||
|  | 	} else { | ||||||
|  | 		replacement = make([]byte, sliceLen+1, cap(slice)+1) | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	for i = 0; i < sliceLen-1; i++ { | ||||||
|  | 		buffer = binary.BigEndian.Uint16(slice[i : i+2]) | ||||||
|  | 		// (buffer) | ||||||
|  | 		// 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 | ||||||
|  | 		// ^-------------^ ^-------------------^ | ||||||
|  | 		//     data byte 0        data byte 1 | ||||||
|  | 		// | ||||||
|  | 		// into | ||||||
|  | 		// | ||||||
|  | 		// 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 | ||||||
|  | 		// ^-----^ ^---------------^ ^--------------------^ ^----------^ | ||||||
|  | 		// zeroes   data byte 0      data byte 1              zeroes | ||||||
|  |  | ||||||
|  | 		if first { | ||||||
|  | 			// The leftover OR'ing will overwrite the first 4 bits of data byte 0. Save them | ||||||
|  | 			firstLeftovers = uint8(buffer & 0xf000 >> 12) | ||||||
|  | 			first = false | ||||||
|  | 		} | ||||||
|  | 		replacementUnit = 0 | ||||||
|  | 		replacementUnit |= uint16(leftovers) << 12 | ||||||
|  | 		replacementUnit |= (buffer & 0xff00) >> 4 // data byte 0 | ||||||
|  | 		replacementUnit |= buffer & 0xff >> 4     // data byte 1 first 4 bits | ||||||
|  | 		lastLeftovers = uint8(buffer&0xf) << 4 | ||||||
|  |  | ||||||
|  | 		replacement[i+1] = byte(replacementUnit) | ||||||
|  |  | ||||||
|  | 		leftovers = uint8((buffer & 0x000f) << 4) | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	replacement[0] = byte(uint8(replacement[0]) | firstLeftovers) | ||||||
|  | 	if lastLeftovers > 0 { | ||||||
|  | 		replacement[sliceLen] = byte(lastLeftovers) | ||||||
|  | 	} | ||||||
|  | 	return | ||||||
|  | } | ||||||
|  |  | ||||||
|  | func Merge2HalfByteSlices(src1, src2 []byte) (output []byte) { | ||||||
|  | 	src1Len := len(src1) | ||||||
|  | 	src2Len := len(src2) | ||||||
|  | 	output = make([]byte, src1Len+src2Len-1) | ||||||
|  |  | ||||||
|  | 	var mergeByte uint8 = src1[src1Len-1] | ||||||
|  | 	mergeByte |= uint8(src2[0]) | ||||||
|  |  | ||||||
|  | 	copy(output, src1) | ||||||
|  | 	copy(output[src1Len:], src2[1:]) | ||||||
|  |  | ||||||
|  | 	output[src1Len-1] = byte(mergeByte) | ||||||
|  |  | ||||||
|  | 	return | ||||||
|  | } | ||||||
							
								
								
									
										8
									
								
								vendor/github.com/couchbase/gomemcached/mc_constants.go
									
									
									
										generated
									
									
										vendored
									
									
								
							
							
						
						
									
										8
									
								
								vendor/github.com/couchbase/gomemcached/mc_constants.go
									
									
									
										generated
									
									
										vendored
									
									
								
							| @@ -6,8 +6,10 @@ import ( | |||||||
| ) | ) | ||||||
|  |  | ||||||
| const ( | const ( | ||||||
| 	REQ_MAGIC = 0x80 | 	REQ_MAGIC      = 0x80 | ||||||
| 	RES_MAGIC = 0x81 | 	RES_MAGIC      = 0x81 | ||||||
|  | 	FLEX_MAGIC     = 0x08 | ||||||
|  | 	FLEX_RES_MAGIC = 0x18 | ||||||
| ) | ) | ||||||
|  |  | ||||||
| // CommandCode for memcached packets. | // CommandCode for memcached packets. | ||||||
| @@ -99,6 +101,8 @@ const ( | |||||||
| 	SUBDOC_GET               = CommandCode(0xc5) // Get subdoc. Returns with xattrs | 	SUBDOC_GET               = CommandCode(0xc5) // Get subdoc. Returns with xattrs | ||||||
| 	SUBDOC_MULTI_LOOKUP      = CommandCode(0xd0) // Multi lookup. Doc xattrs and meta. | 	SUBDOC_MULTI_LOOKUP      = CommandCode(0xd0) // Multi lookup. Doc xattrs and meta. | ||||||
|  |  | ||||||
|  | 	DCP_SYSTEM_EVENT = CommandCode(0x5f) // A system event has occurred | ||||||
|  |  | ||||||
| ) | ) | ||||||
|  |  | ||||||
| // command codes that are counted toward DCP control buffer | // command codes that are counted toward DCP control buffer | ||||||
|   | |||||||
							
								
								
									
										359
									
								
								vendor/github.com/couchbase/gomemcached/mc_req.go
									
									
									
										generated
									
									
										vendored
									
									
								
							
							
						
						
									
										359
									
								
								vendor/github.com/couchbase/gomemcached/mc_req.go
									
									
									
										generated
									
									
										vendored
									
									
								
							| @@ -25,11 +25,17 @@ type MCRequest struct { | |||||||
| 	Extras, Key, Body, ExtMeta []byte | 	Extras, Key, Body, ExtMeta []byte | ||||||
| 	// Datatype identifier | 	// Datatype identifier | ||||||
| 	DataType uint8 | 	DataType uint8 | ||||||
|  | 	// len() calls are expensive - cache this in case for collection | ||||||
|  | 	Keylen int | ||||||
|  | 	// Flexible Framing Extras | ||||||
|  | 	FramingExtras []FrameInfo | ||||||
|  | 	// Stored length of incoming framing extras | ||||||
|  | 	FramingElen int | ||||||
| } | } | ||||||
|  |  | ||||||
| // Size gives the number of bytes this request requires. | // Size gives the number of bytes this request requires. | ||||||
| func (req *MCRequest) Size() int { | func (req *MCRequest) Size() int { | ||||||
| 	return HDR_LEN + len(req.Extras) + len(req.Key) + len(req.Body) + len(req.ExtMeta) | 	return HDR_LEN + len(req.Extras) + len(req.Key) + len(req.Body) + len(req.ExtMeta) + req.FramingElen | ||||||
| } | } | ||||||
|  |  | ||||||
| // A debugging string representation of this request | // A debugging string representation of this request | ||||||
| @@ -38,7 +44,23 @@ func (req MCRequest) String() string { | |||||||
| 		req.Opcode, len(req.Body), req.Key) | 		req.Opcode, len(req.Body), req.Key) | ||||||
| } | } | ||||||
|  |  | ||||||
| func (req *MCRequest) fillHeaderBytes(data []byte) int { | func (req *MCRequest) fillRegularHeaderBytes(data []byte) int { | ||||||
|  | 	//  Byte/     0       |       1       |       2       |       3       | | ||||||
|  | 	//     /              |               |               |               | | ||||||
|  | 	//    |0 1 2 3 4 5 6 7|0 1 2 3 4 5 6 7|0 1 2 3 4 5 6 7|0 1 2 3 4 5 6 7| | ||||||
|  | 	//    +---------------+---------------+---------------+---------------+ | ||||||
|  | 	//   0| Magic         | Opcode        | Key length                    | | ||||||
|  | 	//    +---------------+---------------+---------------+---------------+ | ||||||
|  | 	//   4| Extras length | Data type     | vbucket id                    | | ||||||
|  | 	//    +---------------+---------------+---------------+---------------+ | ||||||
|  | 	//   8| Total body length                                             | | ||||||
|  | 	//    +---------------+---------------+---------------+---------------+ | ||||||
|  | 	//  12| Opaque                                                        | | ||||||
|  | 	//    +---------------+---------------+---------------+---------------+ | ||||||
|  | 	//  16| CAS                                                           | | ||||||
|  | 	//    |                                                               | | ||||||
|  | 	//    +---------------+---------------+---------------+---------------+ | ||||||
|  | 	//    Total 24 bytes | ||||||
|  |  | ||||||
| 	pos := 0 | 	pos := 0 | ||||||
| 	data[pos] = REQ_MAGIC | 	data[pos] = REQ_MAGIC | ||||||
| @@ -84,16 +106,130 @@ func (req *MCRequest) fillHeaderBytes(data []byte) int { | |||||||
| 		copy(data[pos:pos+len(req.Key)], req.Key) | 		copy(data[pos:pos+len(req.Key)], req.Key) | ||||||
| 		pos += len(req.Key) | 		pos += len(req.Key) | ||||||
| 	} | 	} | ||||||
|  |  | ||||||
| 	return pos | 	return pos | ||||||
| } | } | ||||||
|  |  | ||||||
|  | // Returns pos and if trailing by half byte | ||||||
|  | func (req *MCRequest) fillFlexHeaderBytes(data []byte) (int, bool) { | ||||||
|  |  | ||||||
|  | 	//  Byte/     0       |       1       |       2       |       3       | | ||||||
|  | 	//     /              |               |               |               | | ||||||
|  | 	//    |0 1 2 3 4 5 6 7|0 1 2 3 4 5 6 7|0 1 2 3 4 5 6 7|0 1 2 3 4 5 6 7| | ||||||
|  | 	//    +---------------+---------------+---------------+---------------+ | ||||||
|  | 	//   0| Magic (0x08)  | Opcode        | Framing extras| Key Length    | | ||||||
|  | 	//    +---------------+---------------+---------------+---------------+ | ||||||
|  | 	//   4| Extras length | Data type     | vbucket id                    | | ||||||
|  | 	//    +---------------+---------------+---------------+---------------+ | ||||||
|  | 	//   8| Total body length                                             | | ||||||
|  | 	//    +---------------+---------------+---------------+---------------+ | ||||||
|  | 	//  12| Opaque                                                        | | ||||||
|  | 	//    +---------------+---------------+---------------+---------------+ | ||||||
|  | 	//  16| CAS                                                           | | ||||||
|  | 	//    |                                                               | | ||||||
|  | 	//    +---------------+---------------+---------------+---------------+ | ||||||
|  | 	//    Total 24 bytes | ||||||
|  |  | ||||||
|  | 	data[0] = FLEX_MAGIC | ||||||
|  | 	data[1] = byte(req.Opcode) | ||||||
|  | 	data[2] = byte(req.FramingElen) | ||||||
|  | 	data[3] = byte(req.Keylen) | ||||||
|  | 	elen := len(req.Extras) | ||||||
|  | 	data[4] = byte(elen) | ||||||
|  | 	if req.DataType != 0 { | ||||||
|  | 		data[5] = byte(req.DataType) | ||||||
|  | 	} | ||||||
|  | 	binary.BigEndian.PutUint16(data[6:8], req.VBucket) | ||||||
|  | 	binary.BigEndian.PutUint32(data[8:12], | ||||||
|  | 		uint32(len(req.Body)+req.Keylen+elen+len(req.ExtMeta)+req.FramingElen)) | ||||||
|  | 	binary.BigEndian.PutUint32(data[12:16], req.Opaque) | ||||||
|  | 	if req.Cas != 0 { | ||||||
|  | 		binary.BigEndian.PutUint64(data[16:24], req.Cas) | ||||||
|  | 	} | ||||||
|  | 	pos := HDR_LEN | ||||||
|  |  | ||||||
|  | 	// Add framing infos | ||||||
|  | 	var framingExtras []byte | ||||||
|  | 	var outputBytes []byte | ||||||
|  | 	var mergeModeSrc []byte | ||||||
|  | 	var frameBytes int | ||||||
|  | 	var halfByteMode bool | ||||||
|  | 	var mergeMode bool | ||||||
|  | 	for _, frameInfo := range req.FramingExtras { | ||||||
|  | 		if !mergeMode { | ||||||
|  | 			outputBytes, halfByteMode = frameInfo.Bytes() | ||||||
|  | 			if !halfByteMode { | ||||||
|  | 				framingExtras = append(framingExtras, outputBytes...) | ||||||
|  | 				frameBytes += len(outputBytes) | ||||||
|  | 			} else { | ||||||
|  | 				mergeMode = true | ||||||
|  | 				mergeModeSrc = outputBytes | ||||||
|  | 			} | ||||||
|  | 		} else { | ||||||
|  | 			outputBytes, halfByteMode = frameInfo.Bytes() | ||||||
|  | 			outputBytes := ShiftByteSliceRight4Bits(outputBytes) | ||||||
|  | 			if halfByteMode { | ||||||
|  | 				// Previous halfbyte merge with this halfbyte will result in a complete byte | ||||||
|  | 				mergeMode = false | ||||||
|  | 				outputBytes = Merge2HalfByteSlices(mergeModeSrc, outputBytes) | ||||||
|  | 				framingExtras = append(framingExtras, outputBytes...) | ||||||
|  | 				frameBytes += len(outputBytes) | ||||||
|  | 			} else { | ||||||
|  | 				// Merge half byte with a non-half byte will result in a combined half-byte that will | ||||||
|  | 				// become the source for the next iteration | ||||||
|  | 				mergeModeSrc = Merge2HalfByteSlices(mergeModeSrc, outputBytes) | ||||||
|  | 			} | ||||||
|  | 		} | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	if mergeMode { | ||||||
|  | 		// Commit the temporary merge area into framingExtras | ||||||
|  | 		framingExtras = append(framingExtras, mergeModeSrc...) | ||||||
|  | 		frameBytes += len(mergeModeSrc) | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	copy(data[pos:pos+frameBytes], framingExtras) | ||||||
|  |  | ||||||
|  | 	pos += frameBytes | ||||||
|  |  | ||||||
|  | 	// Add Extras | ||||||
|  | 	if len(req.Extras) > 0 { | ||||||
|  | 		if mergeMode { | ||||||
|  | 			outputBytes = ShiftByteSliceRight4Bits(req.Extras) | ||||||
|  | 			data = Merge2HalfByteSlices(data, outputBytes) | ||||||
|  | 		} else { | ||||||
|  | 			copy(data[pos:pos+elen], req.Extras) | ||||||
|  | 		} | ||||||
|  | 		pos += elen | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	// Add keys | ||||||
|  | 	if req.Keylen > 0 { | ||||||
|  | 		if mergeMode { | ||||||
|  | 			outputBytes = ShiftByteSliceRight4Bits(req.Key) | ||||||
|  | 			data = Merge2HalfByteSlices(data, outputBytes) | ||||||
|  | 		} else { | ||||||
|  | 			copy(data[pos:pos+req.Keylen], req.Key) | ||||||
|  | 		} | ||||||
|  | 		pos += req.Keylen | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	return pos, mergeMode | ||||||
|  | } | ||||||
|  |  | ||||||
|  | func (req *MCRequest) FillHeaderBytes(data []byte) (int, bool) { | ||||||
|  | 	if req.FramingElen == 0 { | ||||||
|  | 		return req.fillRegularHeaderBytes(data), false | ||||||
|  | 	} else { | ||||||
|  | 		return req.fillFlexHeaderBytes(data) | ||||||
|  | 	} | ||||||
|  | } | ||||||
|  |  | ||||||
| // HeaderBytes will return the wire representation of the request header | // HeaderBytes will return the wire representation of the request header | ||||||
| // (with the extras and key). | // (with the extras and key). | ||||||
| func (req *MCRequest) HeaderBytes() []byte { | func (req *MCRequest) HeaderBytes() []byte { | ||||||
| 	data := make([]byte, HDR_LEN+len(req.Extras)+len(req.Key)) | 	data := make([]byte, HDR_LEN+len(req.Extras)+len(req.Key)+req.FramingElen) | ||||||
|  |  | ||||||
| 	req.fillHeaderBytes(data) | 	req.FillHeaderBytes(data) | ||||||
|  |  | ||||||
| 	return data | 	return data | ||||||
| } | } | ||||||
| @@ -102,16 +238,27 @@ func (req *MCRequest) HeaderBytes() []byte { | |||||||
| func (req *MCRequest) Bytes() []byte { | func (req *MCRequest) Bytes() []byte { | ||||||
| 	data := make([]byte, req.Size()) | 	data := make([]byte, req.Size()) | ||||||
|  |  | ||||||
| 	pos := req.fillHeaderBytes(data) | 	pos, halfByteMode := req.FillHeaderBytes(data) | ||||||
|  | 	// TODO - the halfByteMode should be revisited for a more efficient | ||||||
|  | 	// way of doing things | ||||||
|  |  | ||||||
| 	if len(req.Body) > 0 { | 	if len(req.Body) > 0 { | ||||||
| 		copy(data[pos:pos+len(req.Body)], req.Body) | 		if halfByteMode { | ||||||
|  | 			shifted := ShiftByteSliceRight4Bits(req.Body) | ||||||
|  | 			data = Merge2HalfByteSlices(data, shifted) | ||||||
|  | 		} else { | ||||||
|  | 			copy(data[pos:pos+len(req.Body)], req.Body) | ||||||
|  | 		} | ||||||
| 	} | 	} | ||||||
|  |  | ||||||
| 	if len(req.ExtMeta) > 0 { | 	if len(req.ExtMeta) > 0 { | ||||||
| 		copy(data[pos+len(req.Body):pos+len(req.Body)+len(req.ExtMeta)], req.ExtMeta) | 		if halfByteMode { | ||||||
|  | 			shifted := ShiftByteSliceRight4Bits(req.ExtMeta) | ||||||
|  | 			data = Merge2HalfByteSlices(data, shifted) | ||||||
|  | 		} else { | ||||||
|  | 			copy(data[pos+len(req.Body):pos+len(req.Body)+len(req.ExtMeta)], req.ExtMeta) | ||||||
|  | 		} | ||||||
| 	} | 	} | ||||||
|  |  | ||||||
| 	return data | 	return data | ||||||
| } | } | ||||||
|  |  | ||||||
| @@ -130,40 +277,44 @@ func (req *MCRequest) Transmit(w io.Writer) (n int, err error) { | |||||||
| 	return | 	return | ||||||
| } | } | ||||||
|  |  | ||||||
| // Receive will fill this MCRequest with the data from a reader. | func (req *MCRequest) receiveHeaderCommon(hdrBytes []byte) (elen, totalBodyLen int) { | ||||||
| func (req *MCRequest) Receive(r io.Reader, hdrBytes []byte) (int, error) { | 	elen = int(hdrBytes[4]) | ||||||
| 	if len(hdrBytes) < HDR_LEN { |  | ||||||
| 		hdrBytes = []byte{ |  | ||||||
| 			0, 0, 0, 0, 0, 0, 0, 0, |  | ||||||
| 			0, 0, 0, 0, 0, 0, 0, 0, |  | ||||||
| 			0, 0, 0, 0, 0, 0, 0, 0} |  | ||||||
| 	} |  | ||||||
| 	n, err := io.ReadFull(r, hdrBytes) |  | ||||||
| 	if err != nil { |  | ||||||
| 		return n, err |  | ||||||
| 	} |  | ||||||
|  |  | ||||||
| 	if hdrBytes[0] != RES_MAGIC && hdrBytes[0] != REQ_MAGIC { |  | ||||||
| 		return n, fmt.Errorf("bad magic: 0x%02x", hdrBytes[0]) |  | ||||||
| 	} |  | ||||||
|  |  | ||||||
| 	klen := int(binary.BigEndian.Uint16(hdrBytes[2:])) |  | ||||||
| 	elen := int(hdrBytes[4]) |  | ||||||
| 	// Data type at 5 | 	// Data type at 5 | ||||||
| 	req.DataType = uint8(hdrBytes[5]) | 	req.DataType = uint8(hdrBytes[5]) | ||||||
|  |  | ||||||
| 	req.Opcode = CommandCode(hdrBytes[1]) | 	req.Opcode = CommandCode(hdrBytes[1]) | ||||||
| 	// Vbucket at 6:7 | 	// Vbucket at 6:7 | ||||||
| 	req.VBucket = binary.BigEndian.Uint16(hdrBytes[6:]) | 	req.VBucket = binary.BigEndian.Uint16(hdrBytes[6:]) | ||||||
| 	totalBodyLen := int(binary.BigEndian.Uint32(hdrBytes[8:])) | 	totalBodyLen = int(binary.BigEndian.Uint32(hdrBytes[8:])) | ||||||
|  |  | ||||||
| 	req.Opaque = binary.BigEndian.Uint32(hdrBytes[12:]) | 	req.Opaque = binary.BigEndian.Uint32(hdrBytes[12:]) | ||||||
| 	req.Cas = binary.BigEndian.Uint64(hdrBytes[16:]) | 	req.Cas = binary.BigEndian.Uint64(hdrBytes[16:]) | ||||||
|  | 	return | ||||||
|  | } | ||||||
|  |  | ||||||
|  | func (req *MCRequest) receiveRegHeader(hdrBytes []byte) (elen, totalBodyLen int) { | ||||||
|  | 	elen, totalBodyLen = req.receiveHeaderCommon(hdrBytes) | ||||||
|  | 	req.Keylen = int(binary.BigEndian.Uint16(hdrBytes[2:])) | ||||||
|  | 	return | ||||||
|  | } | ||||||
|  |  | ||||||
|  | func (req *MCRequest) receiveFlexibleFramingHeader(hdrBytes []byte) (elen, totalBodyLen, framingElen int) { | ||||||
|  | 	elen, totalBodyLen = req.receiveHeaderCommon(hdrBytes) | ||||||
|  |  | ||||||
|  | 	//	 For flexible framing header, key length is a single byte at byte index 3 | ||||||
|  | 	req.Keylen = int(binary.BigEndian.Uint16(hdrBytes[2:]) & 0x0ff) | ||||||
|  | 	// Flexible framing lengh is a single byte at index 2 | ||||||
|  | 	framingElen = int(binary.BigEndian.Uint16(hdrBytes[2:]) >> 8) | ||||||
|  | 	req.FramingElen = framingElen | ||||||
|  | 	return | ||||||
|  | } | ||||||
|  |  | ||||||
|  | func (req *MCRequest) populateRegularBody(r io.Reader, totalBodyLen, elen int) (int, error) { | ||||||
|  | 	var m int | ||||||
|  | 	var err error | ||||||
| 	if totalBodyLen > 0 { | 	if totalBodyLen > 0 { | ||||||
| 		buf := make([]byte, totalBodyLen) | 		buf := make([]byte, totalBodyLen) | ||||||
| 		m, err := io.ReadFull(r, buf) | 		m, err = io.ReadFull(r, buf) | ||||||
| 		n += m |  | ||||||
| 		if err == nil { | 		if err == nil { | ||||||
| 			if req.Opcode >= TAP_MUTATION && | 			if req.Opcode >= TAP_MUTATION && | ||||||
| 				req.Opcode <= TAP_CHECKPOINT_END && | 				req.Opcode <= TAP_CHECKPOINT_END && | ||||||
| @@ -175,7 +326,7 @@ func (req *MCRequest) Receive(r io.Reader, hdrBytes []byte) (int, error) { | |||||||
| 			} | 			} | ||||||
|  |  | ||||||
| 			req.Extras = buf[0:elen] | 			req.Extras = buf[0:elen] | ||||||
| 			req.Key = buf[elen : klen+elen] | 			req.Key = buf[elen : req.Keylen+elen] | ||||||
|  |  | ||||||
| 			// get the length of extended metadata | 			// get the length of extended metadata | ||||||
| 			extMetaLen := 0 | 			extMetaLen := 0 | ||||||
| @@ -183,15 +334,149 @@ func (req *MCRequest) Receive(r io.Reader, hdrBytes []byte) (int, error) { | |||||||
| 				extMetaLen = int(binary.BigEndian.Uint16(req.Extras[28:30])) | 				extMetaLen = int(binary.BigEndian.Uint16(req.Extras[28:30])) | ||||||
| 			} | 			} | ||||||
|  |  | ||||||
| 			bodyLen := totalBodyLen - klen - elen - extMetaLen | 			bodyLen := totalBodyLen - req.Keylen - elen - extMetaLen | ||||||
| 			if bodyLen > MaxBodyLen { | 			if bodyLen > MaxBodyLen { | ||||||
| 				return n, fmt.Errorf("%d is too big (max %d)", | 				return m, fmt.Errorf("%d is too big (max %d)", | ||||||
| 					bodyLen, MaxBodyLen) | 					bodyLen, MaxBodyLen) | ||||||
| 			} | 			} | ||||||
|  |  | ||||||
| 			req.Body = buf[klen+elen : klen+elen+bodyLen] | 			req.Body = buf[req.Keylen+elen : req.Keylen+elen+bodyLen] | ||||||
| 			req.ExtMeta = buf[klen+elen+bodyLen:] | 			req.ExtMeta = buf[req.Keylen+elen+bodyLen:] | ||||||
| 		} | 		} | ||||||
| 	} | 	} | ||||||
| 	return n, err | 	return m, err | ||||||
|  | } | ||||||
|  |  | ||||||
|  | func (req *MCRequest) populateFlexBody(r io.Reader, totalBodyLen, elen, framingElen int) (int, error) { | ||||||
|  | 	var m int | ||||||
|  | 	var err error | ||||||
|  | 	if totalBodyLen > 0 { | ||||||
|  | 		buf := make([]byte, totalBodyLen) | ||||||
|  | 		m, err = io.ReadFull(r, buf) | ||||||
|  | 		if err != nil { | ||||||
|  | 			return m, err | ||||||
|  | 		} | ||||||
|  | 		err = req.populateFlexBodyInternal(buf, totalBodyLen, elen, framingElen) | ||||||
|  | 	} | ||||||
|  | 	return m, err | ||||||
|  | } | ||||||
|  |  | ||||||
|  | func (req *MCRequest) populateFlexBodyInternal(buf []byte, totalBodyLen, elen, framingElen int) error { | ||||||
|  | 	var halfByteOffset bool | ||||||
|  | 	var err error | ||||||
|  | 	if framingElen > 0 { | ||||||
|  | 		var objs []FrameInfo | ||||||
|  | 		objs, err, halfByteOffset = parseFrameInfoObjects(buf, framingElen) | ||||||
|  | 		if err != nil { | ||||||
|  | 			return err | ||||||
|  | 		} | ||||||
|  | 		req.FramingExtras = objs | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	err = req.populateFlexBodyAfterFrames(buf, totalBodyLen, elen, framingElen, halfByteOffset) | ||||||
|  | 	if err != nil { | ||||||
|  | 		return err | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	return nil | ||||||
|  | } | ||||||
|  |  | ||||||
|  | func (req *MCRequest) populateFlexBodyAfterFrames(buf []byte, totalBodyLen, elen, framingElen int, halfByteOffset bool) error { | ||||||
|  | 	var idxCursor int = framingElen | ||||||
|  | 	if req.Opcode >= TAP_MUTATION && req.Opcode <= TAP_CHECKPOINT_END && len(buf[idxCursor:]) > 1 { | ||||||
|  | 		// In these commands there is "engine private" | ||||||
|  | 		// data at the end of the extras.  The first 2 | ||||||
|  | 		// bytes of extra data give its length. | ||||||
|  | 		if !halfByteOffset { | ||||||
|  | 			elen += int(binary.BigEndian.Uint16(buf[idxCursor:])) | ||||||
|  | 		} else { | ||||||
|  | 			// 0 1 2 3 4 .... 19 20 21 22 ... 32 | ||||||
|  | 			// ^-----^ ^-------^  ^------------^ | ||||||
|  | 			//  offset   data       do not care | ||||||
|  | 			var buffer uint32 = binary.BigEndian.Uint32(buf[idxCursor:]) | ||||||
|  | 			buffer &= 0xffff000 | ||||||
|  | 			elen += int(buffer >> 12) | ||||||
|  | 		} | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	// Get the extras | ||||||
|  | 	if !halfByteOffset { | ||||||
|  | 		req.Extras = buf[idxCursor : idxCursor+elen] | ||||||
|  | 	} else { | ||||||
|  | 		preShift := buf[idxCursor : idxCursor+elen+1] | ||||||
|  | 		req.Extras = ShiftByteSliceLeft4Bits(preShift) | ||||||
|  | 	} | ||||||
|  | 	idxCursor += elen | ||||||
|  |  | ||||||
|  | 	// Get the Key | ||||||
|  | 	if !halfByteOffset { | ||||||
|  | 		req.Key = buf[idxCursor : idxCursor+req.Keylen] | ||||||
|  | 	} else { | ||||||
|  | 		preShift := buf[idxCursor : idxCursor+req.Keylen+1] | ||||||
|  | 		req.Key = ShiftByteSliceLeft4Bits(preShift) | ||||||
|  | 	} | ||||||
|  | 	idxCursor += req.Keylen | ||||||
|  |  | ||||||
|  | 	// get the length of extended metadata | ||||||
|  | 	extMetaLen := 0 | ||||||
|  | 	if elen > 29 { | ||||||
|  | 		extMetaLen = int(binary.BigEndian.Uint16(req.Extras[28:30])) | ||||||
|  | 	} | ||||||
|  | 	idxCursor += extMetaLen | ||||||
|  |  | ||||||
|  | 	bodyLen := totalBodyLen - req.Keylen - elen - extMetaLen - framingElen | ||||||
|  | 	if bodyLen > MaxBodyLen { | ||||||
|  | 		return fmt.Errorf("%d is too big (max %d)", | ||||||
|  | 			bodyLen, MaxBodyLen) | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	if !halfByteOffset { | ||||||
|  | 		req.Body = buf[idxCursor : idxCursor+bodyLen] | ||||||
|  | 		idxCursor += bodyLen | ||||||
|  | 	} else { | ||||||
|  | 		preShift := buf[idxCursor : idxCursor+bodyLen+1] | ||||||
|  | 		req.Body = ShiftByteSliceLeft4Bits(preShift) | ||||||
|  | 		idxCursor += bodyLen | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	if extMetaLen > 0 { | ||||||
|  | 		if !halfByteOffset { | ||||||
|  | 			req.ExtMeta = buf[idxCursor:] | ||||||
|  | 		} else { | ||||||
|  | 			preShift := buf[idxCursor:] | ||||||
|  | 			req.ExtMeta = ShiftByteSliceLeft4Bits(preShift) | ||||||
|  | 		} | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	return nil | ||||||
|  | } | ||||||
|  |  | ||||||
|  | // Receive will fill this MCRequest with the data from a reader. | ||||||
|  | func (req *MCRequest) Receive(r io.Reader, hdrBytes []byte) (int, error) { | ||||||
|  | 	if len(hdrBytes) < HDR_LEN { | ||||||
|  | 		hdrBytes = []byte{ | ||||||
|  | 			0, 0, 0, 0, 0, 0, 0, 0, | ||||||
|  | 			0, 0, 0, 0, 0, 0, 0, 0, | ||||||
|  | 			0, 0, 0, 0, 0, 0, 0, 0} | ||||||
|  | 	} | ||||||
|  | 	n, err := io.ReadFull(r, hdrBytes) | ||||||
|  | 	if err != nil { | ||||||
|  | 		fmt.Printf("Err %v\n", err) | ||||||
|  | 		return n, err | ||||||
|  | 	} | ||||||
|  |  | ||||||
|  | 	switch hdrBytes[0] { | ||||||
|  | 	case RES_MAGIC: | ||||||
|  | 		fallthrough | ||||||
|  | 	case REQ_MAGIC: | ||||||
|  | 		elen, totalBodyLen := req.receiveRegHeader(hdrBytes) | ||||||
|  | 		bodyRead, err := req.populateRegularBody(r, totalBodyLen, elen) | ||||||
|  | 		return n + bodyRead, err | ||||||
|  | 	case FLEX_MAGIC: | ||||||
|  | 		elen, totalBodyLen, framingElen := req.receiveFlexibleFramingHeader(hdrBytes) | ||||||
|  | 		bodyRead, err := req.populateFlexBody(r, totalBodyLen, elen, framingElen) | ||||||
|  | 		return n + bodyRead, err | ||||||
|  | 	default: | ||||||
|  | 		return n, fmt.Errorf("bad magic: 0x%02x", hdrBytes[0]) | ||||||
|  | 	} | ||||||
| } | } | ||||||
|   | |||||||
							
								
								
									
										15
									
								
								vendor/github.com/couchbase/gomemcached/mc_res.go
									
									
									
										generated
									
									
										vendored
									
									
								
							
							
						
						
									
										15
									
								
								vendor/github.com/couchbase/gomemcached/mc_res.go
									
									
									
										generated
									
									
										vendored
									
									
								
							| @@ -153,6 +153,13 @@ func (res *MCResponse) Transmit(w io.Writer) (n int, err error) { | |||||||
|  |  | ||||||
| // Receive will fill this MCResponse with the data from this reader. | // Receive will fill this MCResponse with the data from this reader. | ||||||
| func (res *MCResponse) Receive(r io.Reader, hdrBytes []byte) (n int, err error) { | func (res *MCResponse) Receive(r io.Reader, hdrBytes []byte) (n int, err error) { | ||||||
|  | 	return res.ReceiveWithBuf(r, hdrBytes, nil) | ||||||
|  | } | ||||||
|  |  | ||||||
|  | // ReceiveWithBuf takes an optional pre-allocated []byte buf which | ||||||
|  | // will be used if its capacity is large enough, otherwise a new | ||||||
|  | // []byte slice is allocated. | ||||||
|  | func (res *MCResponse) ReceiveWithBuf(r io.Reader, hdrBytes, buf []byte) (n int, err error) { | ||||||
| 	if len(hdrBytes) < HDR_LEN { | 	if len(hdrBytes) < HDR_LEN { | ||||||
| 		hdrBytes = []byte{ | 		hdrBytes = []byte{ | ||||||
| 			0, 0, 0, 0, 0, 0, 0, 0, | 			0, 0, 0, 0, 0, 0, 0, 0, | ||||||
| @@ -187,7 +194,13 @@ func (res *MCResponse) Receive(r io.Reader, hdrBytes []byte) (n int, err error) | |||||||
| 		} | 		} | ||||||
| 	}() | 	}() | ||||||
|  |  | ||||||
| 	buf := make([]byte, klen+elen+bodyLen) | 	bufNeed := klen + elen + bodyLen | ||||||
|  | 	if buf != nil && cap(buf) >= bufNeed { | ||||||
|  | 		buf = buf[0:bufNeed] | ||||||
|  | 	} else { | ||||||
|  | 		buf = make([]byte, bufNeed) | ||||||
|  | 	} | ||||||
|  |  | ||||||
| 	m, err := io.ReadFull(r, buf) | 	m, err := io.ReadFull(r, buf) | ||||||
| 	if err == nil { | 	if err == nil { | ||||||
| 		res.Extras = buf[0:elen] | 		res.Extras = buf[0:elen] | ||||||
|   | |||||||
							
								
								
									
										235
									
								
								vendor/github.com/couchbase/goutils/LICENSE.md
									
									
									
										generated
									
									
										vendored
									
									
								
							
							
						
						
									
										235
									
								
								vendor/github.com/couchbase/goutils/LICENSE.md
									
									
									
										generated
									
									
										vendored
									
									
								
							| @@ -1,47 +1,202 @@ | |||||||
| COUCHBASE INC. COMMUNITY EDITION LICENSE AGREEMENT | Apache License | ||||||
|  |                            Version 2.0, January 2004 | ||||||
|  |                         http://www.apache.org/licenses/ | ||||||
|  |  | ||||||
| IMPORTANT-READ CAREFULLY: BY CLICKING THE "I ACCEPT" BOX OR INSTALLING, |    TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION | ||||||
| DOWNLOADING OR OTHERWISE USING THIS SOFTWARE AND ANY ASSOCIATED |  | ||||||
| DOCUMENTATION, YOU, ON BEHALF OF YOURSELF OR AS AN AUTHORIZED |  | ||||||
| REPRESENTATIVE ON BEHALF OF AN ENTITY ("LICENSEE") AGREE TO ALL THE |  | ||||||
| TERMS OF THIS COMMUNITY EDITION LICENSE AGREEMENT (THE "AGREEMENT") |  | ||||||
| REGARDING YOUR USE OF THE SOFTWARE.  YOU REPRESENT AND WARRANT THAT YOU |  | ||||||
| HAVE FULL LEGAL AUTHORITY TO BIND THE LICENSEE TO THIS AGREEMENT. IF YOU |  | ||||||
| DO NOT AGREE WITH ALL OF THESE TERMS, DO NOT SELECT THE "I ACCEPT" BOX |  | ||||||
| AND DO NOT INSTALL, DOWNLOAD OR OTHERWISE USE THE SOFTWARE. THE |  | ||||||
| EFFECTIVE DATE OF THIS AGREEMENT IS THE DATE ON WHICH YOU CLICK "I |  | ||||||
| ACCEPT" OR OTHERWISE INSTALL, DOWNLOAD OR USE THE SOFTWARE. |  | ||||||
|  |  | ||||||
| 1. License Grant. Couchbase Inc. hereby grants Licensee, free of charge, |    1. Definitions. | ||||||
| the non-exclusive right to use, copy, merge, publish, distribute, |  | ||||||
| sublicense, and/or sell copies of the Software, and to permit persons to |  | ||||||
| whom the Software is furnished to do so, subject to Licensee including |  | ||||||
| the following copyright notice in all copies or substantial portions of |  | ||||||
| the Software: |  | ||||||
|  |  | ||||||
| Couchbase (r) http://www.Couchbase.com Copyright 2016 Couchbase, Inc. |       "License" shall mean the terms and conditions for use, reproduction, | ||||||
|  |       and distribution as defined by Sections 1 through 9 of this document. | ||||||
|  |  | ||||||
| As used in this Agreement, "Software" means the object code version of |       "Licensor" shall mean the copyright owner or entity authorized by | ||||||
| the applicable elastic data management server software provided by |       the copyright owner that is granting the License. | ||||||
| Couchbase Inc. |  | ||||||
|  |  | ||||||
| 2. Restrictions. Licensee will not reverse engineer, disassemble, or |       "Legal Entity" shall mean the union of the acting entity and all | ||||||
| decompile the Software (except to the extent such restrictions are |       other entities that control, are controlled by, or are under common | ||||||
| prohibited by law). |       control with that entity. For the purposes of this definition, | ||||||
|  |       "control" means (i) the power, direct or indirect, to cause the | ||||||
|  |       direction or management of such entity, whether by contract or | ||||||
|  |       otherwise, or (ii) ownership of fifty percent (50%) or more of the | ||||||
|  |       outstanding shares, or (iii) beneficial ownership of such entity. | ||||||
|  |  | ||||||
| 3. Support. Couchbase, Inc. will provide Licensee with access to, and |       "You" (or "Your") shall mean an individual or Legal Entity | ||||||
| use of, the Couchbase, Inc. support forum available at the following |       exercising permissions granted by this License. | ||||||
| URL: http://www.couchbase.org/forums/. Couchbase, Inc. may, at its |  | ||||||
| discretion, modify, suspend or terminate support at any time upon notice |       "Source" form shall mean the preferred form for making modifications, | ||||||
| to Licensee. |       including but not limited to software source code, documentation | ||||||
|  |       source, and configuration files. | ||||||
|  |  | ||||||
|  |       "Object" form shall mean any form resulting from mechanical | ||||||
|  |       transformation or translation of a Source form, including but | ||||||
|  |       not limited to compiled object code, generated documentation, | ||||||
|  |       and conversions to other media types. | ||||||
|  |  | ||||||
|  |       "Work" shall mean the work of authorship, whether in Source or | ||||||
|  |       Object form, made available under the License, as indicated by a | ||||||
|  |       copyright notice that is included in or attached to the work | ||||||
|  |       (an example is provided in the Appendix below). | ||||||
|  |  | ||||||
|  |       "Derivative Works" shall mean any work, whether in Source or Object | ||||||
|  |       form, that is based on (or derived from) the Work and for which the | ||||||
|  |       editorial revisions, annotations, elaborations, or other modifications | ||||||
|  |       represent, as a whole, an original work of authorship. For the purposes | ||||||
|  |       of this License, Derivative Works shall not include works that remain | ||||||
|  |       separable from, or merely link (or bind by name) to the interfaces of, | ||||||
|  |       the Work and Derivative Works thereof. | ||||||
|  |  | ||||||
|  |       "Contribution" shall mean any work of authorship, including | ||||||
|  |       the original version of the Work and any modifications or additions | ||||||
|  |       to that Work or Derivative Works thereof, that is intentionally | ||||||
|  |       submitted to Licensor for inclusion in the Work by the copyright owner | ||||||
|  |       or by an individual or Legal Entity authorized to submit on behalf of | ||||||
|  |       the copyright owner. For the purposes of this definition, "submitted" | ||||||
|  |       means any form of electronic, verbal, or written communication sent | ||||||
|  |       to the Licensor or its representatives, including but not limited to | ||||||
|  |       communication on electronic mailing lists, source code control systems, | ||||||
|  |       and issue tracking systems that are managed by, or on behalf of, the | ||||||
|  |       Licensor for the purpose of discussing and improving the Work, but | ||||||
|  |       excluding communication that is conspicuously marked or otherwise | ||||||
|  |       designated in writing by the copyright owner as "Not a Contribution." | ||||||
|  |  | ||||||
|  |       "Contributor" shall mean Licensor and any individual or Legal Entity | ||||||
|  |       on behalf of whom a Contribution has been received by Licensor and | ||||||
|  |       subsequently incorporated within the Work. | ||||||
|  |  | ||||||
|  |    2. Grant of Copyright License. Subject to the terms and conditions of | ||||||
|  |       this License, each Contributor hereby grants to You a perpetual, | ||||||
|  |       worldwide, non-exclusive, no-charge, royalty-free, irrevocable | ||||||
|  |       copyright license to reproduce, prepare Derivative Works of, | ||||||
|  |       publicly display, publicly perform, sublicense, and distribute the | ||||||
|  |       Work and such Derivative Works in Source or Object form. | ||||||
|  |  | ||||||
|  |    3. Grant of Patent License. Subject to the terms and conditions of | ||||||
|  |       this License, each Contributor hereby grants to You a perpetual, | ||||||
|  |       worldwide, non-exclusive, no-charge, royalty-free, irrevocable | ||||||
|  |       (except as stated in this section) patent license to make, have made, | ||||||
|  |       use, offer to sell, sell, import, and otherwise transfer the Work, | ||||||
|  |       where such license applies only to those patent claims licensable | ||||||
|  |       by such Contributor that are necessarily infringed by their | ||||||
|  |       Contribution(s) alone or by combination of their Contribution(s) | ||||||
|  |       with the Work to which such Contribution(s) was submitted. If You | ||||||
|  |       institute patent litigation against any entity (including a | ||||||
|  |       cross-claim or counterclaim in a lawsuit) alleging that the Work | ||||||
|  |       or a Contribution incorporated within the Work constitutes direct | ||||||
|  |       or contributory patent infringement, then any patent licenses | ||||||
|  |       granted to You under this License for that Work shall terminate | ||||||
|  |       as of the date such litigation is filed. | ||||||
|  |  | ||||||
|  |    4. Redistribution. You may reproduce and distribute copies of the | ||||||
|  |       Work or Derivative Works thereof in any medium, with or without | ||||||
|  |       modifications, and in Source or Object form, provided that You | ||||||
|  |       meet the following conditions: | ||||||
|  |  | ||||||
|  |       (a) You must give any other recipients of the Work or | ||||||
|  |           Derivative Works a copy of this License; and | ||||||
|  |  | ||||||
|  |       (b) You must cause any modified files to carry prominent notices | ||||||
|  |           stating that You changed the files; and | ||||||
|  |  | ||||||
|  |       (c) You must retain, in the Source form of any Derivative Works | ||||||
|  |           that You distribute, all copyright, patent, trademark, and | ||||||
|  |           attribution notices from the Source form of the Work, | ||||||
|  |           excluding those notices that do not pertain to any part of | ||||||
|  |           the Derivative Works; and | ||||||
|  |  | ||||||
|  |       (d) If the Work includes a "NOTICE" text file as part of its | ||||||
|  |           distribution, then any Derivative Works that You distribute must | ||||||
|  |           include a readable copy of the attribution notices contained | ||||||
|  |           within such NOTICE file, excluding those notices that do not | ||||||
|  |           pertain to any part of the Derivative Works, in at least one | ||||||
|  |           of the following places: within a NOTICE text file distributed | ||||||
|  |           as part of the Derivative Works; within the Source form or | ||||||
|  |           documentation, if provided along with the Derivative Works; or, | ||||||
|  |           within a display generated by the Derivative Works, if and | ||||||
|  |           wherever such third-party notices normally appear. The contents | ||||||
|  |           of the NOTICE file are for informational purposes only and | ||||||
|  |           do not modify the License. You may add Your own attribution | ||||||
|  |           notices within Derivative Works that You distribute, alongside | ||||||
|  |           or as an addendum to the NOTICE text from the Work, provided | ||||||
|  |           that such additional attribution notices cannot be construed | ||||||
|  |           as modifying the License. | ||||||
|  |  | ||||||
|  |       You may add Your own copyright statement to Your modifications and | ||||||
|  |       may provide additional or different license terms and conditions | ||||||
|  |       for use, reproduction, or distribution of Your modifications, or | ||||||
|  |       for any such Derivative Works as a whole, provided Your use, | ||||||
|  |       reproduction, and distribution of the Work otherwise complies with | ||||||
|  |       the conditions stated in this License. | ||||||
|  |  | ||||||
|  |    5. Submission of Contributions. Unless You explicitly state otherwise, | ||||||
|  |       any Contribution intentionally submitted for inclusion in the Work | ||||||
|  |       by You to the Licensor shall be under the terms and conditions of | ||||||
|  |       this License, without any additional terms or conditions. | ||||||
|  |       Notwithstanding the above, nothing herein shall supersede or modify | ||||||
|  |       the terms of any separate license agreement you may have executed | ||||||
|  |       with Licensor regarding such Contributions. | ||||||
|  |  | ||||||
|  |    6. Trademarks. This License does not grant permission to use the trade | ||||||
|  |       names, trademarks, service marks, or product names of the Licensor, | ||||||
|  |       except as required for reasonable and customary use in describing the | ||||||
|  |       origin of the Work and reproducing the content of the NOTICE file. | ||||||
|  |  | ||||||
|  |    7. Disclaimer of Warranty. Unless required by applicable law or | ||||||
|  |       agreed to in writing, Licensor provides the Work (and each | ||||||
|  |       Contributor provides its Contributions) on an "AS IS" BASIS, | ||||||
|  |       WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or | ||||||
|  |       implied, including, without limitation, any warranties or conditions | ||||||
|  |       of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A | ||||||
|  |       PARTICULAR PURPOSE. You are solely responsible for determining the | ||||||
|  |       appropriateness of using or redistributing the Work and assume any | ||||||
|  |       risks associated with Your exercise of permissions under this License. | ||||||
|  |  | ||||||
|  |    8. Limitation of Liability. In no event and under no legal theory, | ||||||
|  |       whether in tort (including negligence), contract, or otherwise, | ||||||
|  |       unless required by applicable law (such as deliberate and grossly | ||||||
|  |       negligent acts) or agreed to in writing, shall any Contributor be | ||||||
|  |       liable to You for damages, including any direct, indirect, special, | ||||||
|  |       incidental, or consequential damages of any character arising as a | ||||||
|  |       result of this License or out of the use or inability to use the | ||||||
|  |       Work (including but not limited to damages for loss of goodwill, | ||||||
|  |       work stoppage, computer failure or malfunction, or any and all | ||||||
|  |       other commercial damages or losses), even if such Contributor | ||||||
|  |       has been advised of the possibility of such damages. | ||||||
|  |  | ||||||
|  |    9. Accepting Warranty or Additional Liability. While redistributing | ||||||
|  |       the Work or Derivative Works thereof, You may choose to offer, | ||||||
|  |       and charge a fee for, acceptance of support, warranty, indemnity, | ||||||
|  |       or other liability obligations and/or rights consistent with this | ||||||
|  |       License. However, in accepting such obligations, You may act only | ||||||
|  |       on Your own behalf and on Your sole responsibility, not on behalf | ||||||
|  |       of any other Contributor, and only if You agree to indemnify, | ||||||
|  |       defend, and hold each Contributor harmless for any liability | ||||||
|  |       incurred by, or claims asserted against, such Contributor by reason | ||||||
|  |       of your accepting any such warranty or additional liability. | ||||||
|  |  | ||||||
|  |    END OF TERMS AND CONDITIONS | ||||||
|  |  | ||||||
|  |    APPENDIX: How to apply the Apache License to your work. | ||||||
|  |  | ||||||
|  |       To apply the Apache License to your work, attach the following | ||||||
|  |       boilerplate notice, with the fields enclosed by brackets "{}" | ||||||
|  |       replaced with your own identifying information. (Don't include | ||||||
|  |       the brackets!)  The text should be enclosed in the appropriate | ||||||
|  |       comment syntax for the file format. We also recommend that a | ||||||
|  |       file or class name and description of purpose be included on the | ||||||
|  |       same "printed page" as the copyright notice for easier | ||||||
|  |       identification within third-party archives. | ||||||
|  |  | ||||||
|  |    Copyright {yyyy} {name of copyright owner} | ||||||
|  |  | ||||||
|  |    Licensed under the Apache License, Version 2.0 (the "License"); | ||||||
|  |    you may not use this file except in compliance with the License. | ||||||
|  |    You may obtain a copy of the License at | ||||||
|  |  | ||||||
|  |        http://www.apache.org/licenses/LICENSE-2.0 | ||||||
|  |  | ||||||
|  |    Unless required by applicable law or agreed to in writing, software | ||||||
|  |    distributed under the License is distributed on an "AS IS" BASIS, | ||||||
|  |    WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||||||
|  |    See the License for the specific language governing permissions and | ||||||
|  |    limitations under the License. | ||||||
|  |  | ||||||
| 4. Warranty Disclaimer and Limitation of Liability. THE SOFTWARE IS |  | ||||||
| PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, |  | ||||||
| INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS |  | ||||||
| FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL |  | ||||||
| COUCHBASE INC. OR THE AUTHORS OR COPYRIGHT HOLDERS IN THE SOFTWARE BE |  | ||||||
| LIABLE FOR ANY CLAIM, DAMAGES (IINCLUDING, WITHOUT LIMITATION, DIRECT, |  | ||||||
| INDIRECT OR CONSEQUENTIAL DAMAGES) OR OTHER LIABILITY, WHETHER IN AN |  | ||||||
| ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN |  | ||||||
| CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE |  | ||||||
| SOFTWARE. |  | ||||||
|   | |||||||
							
								
								
									
										4
									
								
								vendor/modules.txt
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										4
									
								
								vendor/modules.txt
									
									
									
									
										vendored
									
									
								
							| @@ -99,10 +99,10 @@ github.com/boombuler/barcode/qr | |||||||
| github.com/boombuler/barcode/utils | github.com/boombuler/barcode/utils | ||||||
| # github.com/bradfitz/gomemcache v0.0.0-20190329173943-551aad21a668 | # github.com/bradfitz/gomemcache v0.0.0-20190329173943-551aad21a668 | ||||||
| github.com/bradfitz/gomemcache/memcache | github.com/bradfitz/gomemcache/memcache | ||||||
| # github.com/couchbase/gomemcached v0.0.0-20190515232915-c4b4ca0eb21d | # github.com/couchbase/gomemcached v0.0.0-20191004160342-7b5da2ec40b2 | ||||||
| github.com/couchbase/gomemcached | github.com/couchbase/gomemcached | ||||||
| github.com/couchbase/gomemcached/client | github.com/couchbase/gomemcached/client | ||||||
| # github.com/couchbase/goutils v0.0.0-20190315194238-f9d42b11473b | # github.com/couchbase/goutils v0.0.0-20191018232750-b49639060d85 | ||||||
| github.com/couchbase/goutils/logging | github.com/couchbase/goutils/logging | ||||||
| github.com/couchbase/goutils/scramsha | github.com/couchbase/goutils/scramsha | ||||||
| # github.com/couchbase/vellum v0.0.0-20190829182332-ef2e028c01fd | # github.com/couchbase/vellum v0.0.0-20190829182332-ef2e028c01fd | ||||||
|   | |||||||
		Reference in New Issue
	
	Block a user