mirror of
				https://github.com/go-gitea/gitea.git
				synced 2025-10-27 00:23:41 +09:00 
			
		
		
		
	Replace all contexts in tests with go1.24 t.Context() --------- Co-authored-by: Giteabot <teabot@gitea.io> Co-authored-by: wxiaoguang <wxiaoguang@gmail.com>
		
			
				
	
	
		
			182 lines
		
	
	
		
			4.2 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			182 lines
		
	
	
		
			4.2 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| // Copyright 2024 The Gitea Authors. All rights reserved.
 | |
| // SPDX-License-Identifier: MIT
 | |
| 
 | |
| package globallock
 | |
| 
 | |
| import (
 | |
| 	"context"
 | |
| 	"os"
 | |
| 	"sync"
 | |
| 	"testing"
 | |
| 	"time"
 | |
| 
 | |
| 	"github.com/go-redsync/redsync/v4"
 | |
| 	"github.com/stretchr/testify/assert"
 | |
| 	"github.com/stretchr/testify/require"
 | |
| )
 | |
| 
 | |
| func TestLocker(t *testing.T) {
 | |
| 	t.Run("redis", func(t *testing.T) {
 | |
| 		url := "redis://127.0.0.1:6379/0"
 | |
| 		if os.Getenv("CI") == "" {
 | |
| 			// Make it possible to run tests against a local redis instance
 | |
| 			url = os.Getenv("TEST_REDIS_URL")
 | |
| 			if url == "" {
 | |
| 				t.Skip("TEST_REDIS_URL not set and not running in CI")
 | |
| 				return
 | |
| 			}
 | |
| 		}
 | |
| 		oldExpiry := redisLockExpiry
 | |
| 		redisLockExpiry = 5 * time.Second // make it shorter for testing
 | |
| 		defer func() {
 | |
| 			redisLockExpiry = oldExpiry
 | |
| 		}()
 | |
| 
 | |
| 		locker := NewRedisLocker(url)
 | |
| 		testLocker(t, locker)
 | |
| 		testRedisLocker(t, locker.(*redisLocker))
 | |
| 		require.NoError(t, locker.(*redisLocker).Close())
 | |
| 	})
 | |
| 	t.Run("memory", func(t *testing.T) {
 | |
| 		locker := NewMemoryLocker()
 | |
| 		testLocker(t, locker)
 | |
| 		testMemoryLocker(t, locker.(*memoryLocker))
 | |
| 	})
 | |
| }
 | |
| 
 | |
| func testLocker(t *testing.T, locker Locker) {
 | |
| 	t.Run("lock", func(t *testing.T) {
 | |
| 		parentCtx := t.Context()
 | |
| 		release, err := locker.Lock(parentCtx, "test")
 | |
| 		defer release()
 | |
| 
 | |
| 		assert.NoError(t, err)
 | |
| 
 | |
| 		func() {
 | |
| 			ctx, cancel := context.WithTimeout(t.Context(), time.Second)
 | |
| 			defer cancel()
 | |
| 			release, err := locker.Lock(ctx, "test")
 | |
| 			defer release()
 | |
| 
 | |
| 			assert.Error(t, err)
 | |
| 		}()
 | |
| 
 | |
| 		release()
 | |
| 
 | |
| 		func() {
 | |
| 			release, err := locker.Lock(t.Context(), "test")
 | |
| 			defer release()
 | |
| 
 | |
| 			assert.NoError(t, err)
 | |
| 		}()
 | |
| 	})
 | |
| 
 | |
| 	t.Run("try lock", func(t *testing.T) {
 | |
| 		parentCtx := t.Context()
 | |
| 		ok, release, err := locker.TryLock(parentCtx, "test")
 | |
| 		defer release()
 | |
| 
 | |
| 		assert.True(t, ok)
 | |
| 		assert.NoError(t, err)
 | |
| 
 | |
| 		func() {
 | |
| 			ctx, cancel := context.WithTimeout(t.Context(), time.Second)
 | |
| 			defer cancel()
 | |
| 			ok, release, err := locker.TryLock(ctx, "test")
 | |
| 			defer release()
 | |
| 
 | |
| 			assert.False(t, ok)
 | |
| 			assert.NoError(t, err)
 | |
| 		}()
 | |
| 
 | |
| 		release()
 | |
| 
 | |
| 		func() {
 | |
| 			ok, release, _ := locker.TryLock(t.Context(), "test")
 | |
| 			defer release()
 | |
| 
 | |
| 			assert.True(t, ok)
 | |
| 		}()
 | |
| 	})
 | |
| 
 | |
| 	t.Run("wait and acquired", func(t *testing.T) {
 | |
| 		ctx := t.Context()
 | |
| 		release, err := locker.Lock(ctx, "test")
 | |
| 		require.NoError(t, err)
 | |
| 
 | |
| 		wg := &sync.WaitGroup{}
 | |
| 		wg.Add(1)
 | |
| 		go func() {
 | |
| 			defer wg.Done()
 | |
| 			started := time.Now()
 | |
| 			release, err := locker.Lock(t.Context(), "test") // should be blocked for seconds
 | |
| 			defer release()
 | |
| 			assert.Greater(t, time.Since(started), time.Second)
 | |
| 			assert.NoError(t, err)
 | |
| 		}()
 | |
| 
 | |
| 		time.Sleep(2 * time.Second)
 | |
| 		release()
 | |
| 
 | |
| 		wg.Wait()
 | |
| 	})
 | |
| 
 | |
| 	t.Run("multiple release", func(t *testing.T) {
 | |
| 		ctx := t.Context()
 | |
| 
 | |
| 		release1, err := locker.Lock(ctx, "test")
 | |
| 		require.NoError(t, err)
 | |
| 
 | |
| 		release1()
 | |
| 
 | |
| 		release2, err := locker.Lock(ctx, "test")
 | |
| 		defer release2()
 | |
| 		require.NoError(t, err)
 | |
| 
 | |
| 		// Call release1 again,
 | |
| 		// it should not panic or block,
 | |
| 		// and it shouldn't affect the other lock
 | |
| 		release1()
 | |
| 
 | |
| 		ok, release3, err := locker.TryLock(ctx, "test")
 | |
| 		defer release3()
 | |
| 		require.NoError(t, err)
 | |
| 		// It should be able to acquire the lock;
 | |
| 		// otherwise, it means the lock has been released by release1
 | |
| 		assert.False(t, ok)
 | |
| 	})
 | |
| }
 | |
| 
 | |
| // testMemoryLocker does specific tests for memoryLocker
 | |
| func testMemoryLocker(t *testing.T, locker *memoryLocker) {
 | |
| 	// nothing to do
 | |
| }
 | |
| 
 | |
| // testRedisLocker does specific tests for redisLocker
 | |
| func testRedisLocker(t *testing.T, locker *redisLocker) {
 | |
| 	defer func() {
 | |
| 		// This case should be tested at the end.
 | |
| 		// Otherwise, it will affect other tests.
 | |
| 		t.Run("close", func(t *testing.T) {
 | |
| 			assert.NoError(t, locker.Close())
 | |
| 			_, err := locker.Lock(t.Context(), "test")
 | |
| 			assert.Error(t, err)
 | |
| 		})
 | |
| 	}()
 | |
| 
 | |
| 	t.Run("failed extend", func(t *testing.T) {
 | |
| 		release, err := locker.Lock(t.Context(), "test")
 | |
| 		defer release()
 | |
| 		require.NoError(t, err)
 | |
| 
 | |
| 		// It simulates that there are some problems with extending like network issues or redis server down.
 | |
| 		v, ok := locker.mutexM.Load("test")
 | |
| 		require.True(t, ok)
 | |
| 		m := v.(*redsync.Mutex)
 | |
| 		_, _ = m.Unlock() // release it to make it impossible to extend
 | |
| 
 | |
| 		// In current design, callers can't know the lock can't be extended.
 | |
| 		// Just keep this case to improve the test coverage.
 | |
| 	})
 | |
| }
 |