mirror of
https://github.com/zeromicro/go-zero.git
synced 2026-05-13 01:40:00 +08:00
Compare commits
112 Commits
v1.2.3
...
tools/goct
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
0e0abc3a95 | ||
|
|
696fda1db4 | ||
|
|
c1d2634427 | ||
|
|
4b7a680ac5 | ||
|
|
b3e7d2901f | ||
|
|
cdf7ec213c | ||
|
|
f1102fb262 | ||
|
|
09d1fad6e0 | ||
|
|
379c65a3ef | ||
|
|
fdc7f64d6f | ||
|
|
df0f8ed59e | ||
|
|
c903966fc7 | ||
|
|
e57fa8ff53 | ||
|
|
bf2feee5b7 | ||
|
|
ce05c429fc | ||
|
|
272a3f347d | ||
|
|
13db7a1931 | ||
|
|
468c237189 | ||
|
|
b9b80c068b | ||
|
|
9b592b3dee | ||
|
|
2203809e5e | ||
|
|
8d6d37f71e | ||
|
|
ea4f2af67f | ||
|
|
53af194ef9 | ||
|
|
5e0e2d2b14 | ||
|
|
74c99184c5 | ||
|
|
eb4b86137a | ||
|
|
9c4f4f3b4e | ||
|
|
240132e7c7 | ||
|
|
9d67fc4cfb | ||
|
|
892f93a716 | ||
|
|
ba6a7c9dc8 | ||
|
|
a91c3907a8 | ||
|
|
e267d94ee1 | ||
|
|
89ce5e492b | ||
|
|
290de6aa96 | ||
|
|
a7aeb8ac0e | ||
|
|
a8e7fafebf | ||
|
|
7cc64070b1 | ||
|
|
c19d2637ea | ||
|
|
fe1da14332 | ||
|
|
8e9110cedf | ||
|
|
d6ff30a570 | ||
|
|
b98d46bfd6 | ||
|
|
768936b256 | ||
|
|
c6eb1a9670 | ||
|
|
e4ab518576 | ||
|
|
dfc67b5fac | ||
|
|
62266d8f91 | ||
|
|
b8ea16a88e | ||
|
|
23deaf50e6 | ||
|
|
38a36ed8d3 | ||
|
|
49bab23c54 | ||
|
|
78ba00d3a7 | ||
|
|
787b046a70 | ||
|
|
f827a7b985 | ||
|
|
f5f2097d14 | ||
|
|
cfcfb87fd4 | ||
|
|
1d223fc114 | ||
|
|
c0647f0719 | ||
|
|
8745ed9c61 | ||
|
|
836726e710 | ||
|
|
a67c118dcf | ||
|
|
cd289465fd | ||
|
|
263e426ae1 | ||
|
|
d5e493383a | ||
|
|
6f1d27354a | ||
|
|
26101732d2 | ||
|
|
71d40e0c08 | ||
|
|
4ba2ff7cdd | ||
|
|
2cdf5e7395 | ||
|
|
8315a55b3f | ||
|
|
d1c2a31af7 | ||
|
|
3e6c217408 | ||
|
|
b299f350be | ||
|
|
8fd16c17dc | ||
|
|
5979b2aa0f | ||
|
|
0b17e0e5d9 | ||
|
|
3d8ad5e4f6 | ||
|
|
ff1752dd39 | ||
|
|
1becaeb7be | ||
|
|
171afaadb9 | ||
|
|
776e6e647d | ||
|
|
4ccdf4ec72 | ||
|
|
a7bd993c0c | ||
|
|
a290ff4486 | ||
|
|
490ef13822 | ||
|
|
1b14de2ff9 | ||
|
|
914692cc82 | ||
|
|
07191dc430 | ||
|
|
af3fb2b04d | ||
|
|
0240fa131a | ||
|
|
e96577dd38 | ||
|
|
403dd7367a | ||
|
|
8086ad120b | ||
|
|
87a445689c | ||
|
|
b6bda54870 | ||
|
|
9d528dddd6 | ||
|
|
543d590710 | ||
|
|
f1d70eb6b2 | ||
|
|
d828c3f37e | ||
|
|
038491b7bc | ||
|
|
cf683411ee | ||
|
|
de5ed6a677 | ||
|
|
3dda557410 | ||
|
|
c800f6f723 | ||
|
|
0395ba1816 | ||
|
|
86f9f63b46 | ||
|
|
a7a6753118 | ||
|
|
2e80d12d6a | ||
|
|
417a96cbf2 | ||
|
|
2d4c29ea7c |
2
.github/FUNDING.yml
vendored
2
.github/FUNDING.yml
vendored
@@ -9,4 +9,4 @@ community_bridge: # Replace with a single Community Bridge project-name e.g., cl
|
|||||||
liberapay: # Replace with a single Liberapay username
|
liberapay: # Replace with a single Liberapay username
|
||||||
issuehunt: # Replace with a single IssueHunt username
|
issuehunt: # Replace with a single IssueHunt username
|
||||||
otechie: # Replace with a single Otechie username
|
otechie: # Replace with a single Otechie username
|
||||||
custom: # https://gitee.com/kevwan/static/raw/master/images/sponsor.jpg
|
custom: https://gitee.com/kevwan/static/raw/master/images/sponsor.jpg
|
||||||
|
|||||||
2
.github/workflows/go.yml
vendored
2
.github/workflows/go.yml
vendored
@@ -15,7 +15,7 @@ jobs:
|
|||||||
- name: Set up Go 1.x
|
- name: Set up Go 1.x
|
||||||
uses: actions/setup-go@v2
|
uses: actions/setup-go@v2
|
||||||
with:
|
with:
|
||||||
go-version: ^1.14
|
go-version: ^1.15
|
||||||
id: go
|
id: go
|
||||||
|
|
||||||
- name: Check out code into the Go module directory
|
- name: Check out code into the Go module directory
|
||||||
|
|||||||
18
.github/workflows/issue-translator.yml
vendored
Normal file
18
.github/workflows/issue-translator.yml
vendored
Normal file
@@ -0,0 +1,18 @@
|
|||||||
|
name: 'issue-translator'
|
||||||
|
on:
|
||||||
|
issue_comment:
|
||||||
|
types: [created]
|
||||||
|
issues:
|
||||||
|
types: [opened]
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
build:
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
steps:
|
||||||
|
- uses: tomsun28/issues-translate-action@v2.6
|
||||||
|
with:
|
||||||
|
IS_MODIFY_TITLE: true
|
||||||
|
# not require, default false, . Decide whether to modify the issue title
|
||||||
|
# if true, the robot account @Issues-translate-bot must have modification permissions, invite @Issues-translate-bot to your project or use your custom bot.
|
||||||
|
CUSTOM_BOT_NOTE: Bot detected the issue body's language is not English, translate it automatically. 👯👭🏻🧑🤝🧑👫🧑🏿🤝🧑🏻👩🏾🤝👨🏿👬🏿
|
||||||
|
# not require. Customize the translation robot prefix message.
|
||||||
@@ -40,7 +40,7 @@ We will help you to contribute in different areas like filing issues, developing
|
|||||||
getting your work reviewed and merged.
|
getting your work reviewed and merged.
|
||||||
|
|
||||||
If you have questions about the development process,
|
If you have questions about the development process,
|
||||||
feel free to [file an issue](https://github.com/tal-tech/go-zero/issues/new/choose).
|
feel free to [file an issue](https://github.com/zeromicro/go-zero/issues/new/choose).
|
||||||
|
|
||||||
## Find something to work on
|
## Find something to work on
|
||||||
|
|
||||||
@@ -50,10 +50,10 @@ Here is how you get started.
|
|||||||
|
|
||||||
### Find a good first topic
|
### Find a good first topic
|
||||||
|
|
||||||
[go-zero](https://github.com/tal-tech/go-zero) has beginner-friendly issues that provide a good first issue.
|
[go-zero](https://github.com/zeromicro/go-zero) has beginner-friendly issues that provide a good first issue.
|
||||||
For example, [go-zero](https://github.com/tal-tech/go-zero) has
|
For example, [go-zero](https://github.com/zeromicro/go-zero) has
|
||||||
[help wanted](https://github.com/tal-tech/go-zero/issues?q=is%3Aopen+is%3Aissue+label%3A%22help+wanted%22) and
|
[help wanted](https://github.com/zeromicro/go-zero/issues?q=is%3Aopen+is%3Aissue+label%3A%22help+wanted%22) and
|
||||||
[good first issue](https://github.com/tal-tech/go-zero/issues?q=is%3Aopen+is%3Aissue+label%3A%22good+first+issue%22)
|
[good first issue](https://github.com/zeromicro/go-zero/issues?q=is%3Aopen+is%3Aissue+label%3A%22good+first+issue%22)
|
||||||
labels for issues that should not need deep knowledge of the system.
|
labels for issues that should not need deep knowledge of the system.
|
||||||
We can help new contributors who wish to work on such issues.
|
We can help new contributors who wish to work on such issues.
|
||||||
|
|
||||||
@@ -79,7 +79,7 @@ This is a rough outline of what a contributor's workflow looks like:
|
|||||||
- Create a topic branch from where to base the contribution. This is usually master.
|
- Create a topic branch from where to base the contribution. This is usually master.
|
||||||
- Make commits of logical units.
|
- Make commits of logical units.
|
||||||
- Push changes in a topic branch to a personal fork of the repository.
|
- Push changes in a topic branch to a personal fork of the repository.
|
||||||
- Submit a pull request to [go-zero](https://github.com/tal-tech/go-zero).
|
- Submit a pull request to [go-zero](https://github.com/zeromicro/go-zero).
|
||||||
|
|
||||||
## Creating Pull Requests
|
## Creating Pull Requests
|
||||||
|
|
||||||
|
|||||||
@@ -14,9 +14,10 @@ We hope that the items listed below will inspire further engagement from the com
|
|||||||
|
|
||||||
## 2021 Q4
|
## 2021 Q4
|
||||||
- [x] Support `username/password` authentication in ETCD
|
- [x] Support `username/password` authentication in ETCD
|
||||||
|
- [x] Support `SSL/TLS` in ETCD
|
||||||
- [x] Support `SSL/TLS` in `zRPC`
|
- [x] Support `SSL/TLS` in `zRPC`
|
||||||
- [x] Support `TLS` in redis connections
|
- [x] Support `TLS` in redis connections
|
||||||
- [ ] Support `retry strategies` in `zRPC`
|
- [x] Support `goctl bug` to report bugs conveniently
|
||||||
|
|
||||||
## 2022
|
## 2022
|
||||||
- [ ] Support `goctl mock` command to start a mocking server with given `.api` file
|
- [ ] Support `goctl mock` command to start a mocking server with given `.api` file
|
||||||
|
|||||||
@@ -4,8 +4,8 @@ import (
|
|||||||
"errors"
|
"errors"
|
||||||
"strconv"
|
"strconv"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/hash"
|
"github.com/zeromicro/go-zero/core/hash"
|
||||||
"github.com/tal-tech/go-zero/core/stores/redis"
|
"github.com/zeromicro/go-zero/core/stores/redis"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
|
|||||||
@@ -4,7 +4,7 @@ import (
|
|||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/stores/redis/redistest"
|
"github.com/zeromicro/go-zero/core/stores/redis/redistest"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestRedisBitSet_New_Set_Test(t *testing.T) {
|
func TestRedisBitSet_New_Set_Test(t *testing.T) {
|
||||||
|
|||||||
@@ -6,11 +6,11 @@ import (
|
|||||||
"strings"
|
"strings"
|
||||||
"sync"
|
"sync"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/mathx"
|
"github.com/zeromicro/go-zero/core/mathx"
|
||||||
"github.com/tal-tech/go-zero/core/proc"
|
"github.com/zeromicro/go-zero/core/proc"
|
||||||
"github.com/tal-tech/go-zero/core/stat"
|
"github.com/zeromicro/go-zero/core/stat"
|
||||||
"github.com/tal-tech/go-zero/core/stringx"
|
"github.com/zeromicro/go-zero/core/stringx"
|
||||||
"github.com/tal-tech/go-zero/core/timex"
|
"github.com/zeromicro/go-zero/core/timex"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
|
|||||||
@@ -8,7 +8,7 @@ import (
|
|||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/stat"
|
"github.com/zeromicro/go-zero/core/stat"
|
||||||
)
|
)
|
||||||
|
|
||||||
func init() {
|
func init() {
|
||||||
|
|||||||
@@ -6,7 +6,7 @@ import (
|
|||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/stat"
|
"github.com/zeromicro/go-zero/core/stat"
|
||||||
)
|
)
|
||||||
|
|
||||||
func init() {
|
func init() {
|
||||||
|
|||||||
@@ -4,8 +4,8 @@ import (
|
|||||||
"math"
|
"math"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/collection"
|
"github.com/zeromicro/go-zero/core/collection"
|
||||||
"github.com/tal-tech/go-zero/core/mathx"
|
"github.com/zeromicro/go-zero/core/mathx"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
|
|||||||
@@ -7,9 +7,9 @@ import (
|
|||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/collection"
|
"github.com/zeromicro/go-zero/core/collection"
|
||||||
"github.com/tal-tech/go-zero/core/mathx"
|
"github.com/zeromicro/go-zero/core/mathx"
|
||||||
"github.com/tal-tech/go-zero/core/stat"
|
"github.com/zeromicro/go-zero/core/stat"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
|
|||||||
@@ -8,8 +8,8 @@ import (
|
|||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/iox"
|
"github.com/zeromicro/go-zero/core/iox"
|
||||||
"github.com/tal-tech/go-zero/core/lang"
|
"github.com/zeromicro/go-zero/core/lang"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestEnterToContinue(t *testing.T) {
|
func TestEnterToContinue(t *testing.T) {
|
||||||
|
|||||||
@@ -7,7 +7,7 @@ import (
|
|||||||
"encoding/base64"
|
"encoding/base64"
|
||||||
"errors"
|
"errors"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/logx"
|
"github.com/zeromicro/go-zero/core/logx"
|
||||||
)
|
)
|
||||||
|
|
||||||
// ErrPaddingSize indicates bad padding size.
|
// ErrPaddingSize indicates bad padding size.
|
||||||
|
|||||||
@@ -5,7 +5,7 @@ import (
|
|||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/fs"
|
"github.com/zeromicro/go-zero/core/fs"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
|
|||||||
@@ -6,9 +6,9 @@ import (
|
|||||||
"sync/atomic"
|
"sync/atomic"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/logx"
|
"github.com/zeromicro/go-zero/core/logx"
|
||||||
"github.com/tal-tech/go-zero/core/mathx"
|
"github.com/zeromicro/go-zero/core/mathx"
|
||||||
"github.com/tal-tech/go-zero/core/syncx"
|
"github.com/zeromicro/go-zero/core/syncx"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
|
|||||||
@@ -4,7 +4,7 @@ import (
|
|||||||
"sync"
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/timex"
|
"github.com/zeromicro/go-zero/core/timex"
|
||||||
)
|
)
|
||||||
|
|
||||||
type (
|
type (
|
||||||
|
|||||||
@@ -6,7 +6,7 @@ import (
|
|||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/stringx"
|
"github.com/zeromicro/go-zero/core/stringx"
|
||||||
)
|
)
|
||||||
|
|
||||||
const duration = time.Millisecond * 50
|
const duration = time.Millisecond * 50
|
||||||
|
|||||||
@@ -4,7 +4,7 @@ import (
|
|||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/stringx"
|
"github.com/zeromicro/go-zero/core/stringx"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestSafeMap(t *testing.T) {
|
func TestSafeMap(t *testing.T) {
|
||||||
|
|||||||
@@ -1,8 +1,8 @@
|
|||||||
package collection
|
package collection
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"github.com/tal-tech/go-zero/core/lang"
|
"github.com/zeromicro/go-zero/core/lang"
|
||||||
"github.com/tal-tech/go-zero/core/logx"
|
"github.com/zeromicro/go-zero/core/logx"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
|
|||||||
@@ -5,7 +5,7 @@ import (
|
|||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/logx"
|
"github.com/zeromicro/go-zero/core/logx"
|
||||||
)
|
)
|
||||||
|
|
||||||
func init() {
|
func init() {
|
||||||
|
|||||||
@@ -5,9 +5,9 @@ import (
|
|||||||
"fmt"
|
"fmt"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/lang"
|
"github.com/zeromicro/go-zero/core/lang"
|
||||||
"github.com/tal-tech/go-zero/core/threading"
|
"github.com/zeromicro/go-zero/core/threading"
|
||||||
"github.com/tal-tech/go-zero/core/timex"
|
"github.com/zeromicro/go-zero/core/timex"
|
||||||
)
|
)
|
||||||
|
|
||||||
const drainWorkers = 8
|
const drainWorkers = 8
|
||||||
|
|||||||
@@ -8,10 +8,10 @@ import (
|
|||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/lang"
|
"github.com/zeromicro/go-zero/core/lang"
|
||||||
"github.com/tal-tech/go-zero/core/stringx"
|
"github.com/zeromicro/go-zero/core/stringx"
|
||||||
"github.com/tal-tech/go-zero/core/syncx"
|
"github.com/zeromicro/go-zero/core/syncx"
|
||||||
"github.com/tal-tech/go-zero/core/timex"
|
"github.com/zeromicro/go-zero/core/timex"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
|
|||||||
@@ -7,7 +7,7 @@ import (
|
|||||||
"os"
|
"os"
|
||||||
"path"
|
"path"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/mapping"
|
"github.com/zeromicro/go-zero/core/mapping"
|
||||||
)
|
)
|
||||||
|
|
||||||
var loaders = map[string]func([]byte, interface{}) error{
|
var loaders = map[string]func([]byte, interface{}) error{
|
||||||
|
|||||||
@@ -6,8 +6,8 @@ import (
|
|||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/fs"
|
"github.com/zeromicro/go-zero/core/fs"
|
||||||
"github.com/tal-tech/go-zero/core/hash"
|
"github.com/zeromicro/go-zero/core/hash"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestLoadConfig_notExists(t *testing.T) {
|
func TestLoadConfig_notExists(t *testing.T) {
|
||||||
|
|||||||
@@ -7,7 +7,7 @@ import (
|
|||||||
"strings"
|
"strings"
|
||||||
"sync"
|
"sync"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/iox"
|
"github.com/zeromicro/go-zero/core/iox"
|
||||||
)
|
)
|
||||||
|
|
||||||
// PropertyError represents a configuration error message.
|
// PropertyError represents a configuration error message.
|
||||||
|
|||||||
@@ -5,7 +5,7 @@ import (
|
|||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/fs"
|
"github.com/zeromicro/go-zero/core/fs"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestProperties(t *testing.T) {
|
func TestProperties(t *testing.T) {
|
||||||
|
|||||||
@@ -3,7 +3,7 @@ package contextx
|
|||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/mapping"
|
"github.com/zeromicro/go-zero/core/mapping"
|
||||||
)
|
)
|
||||||
|
|
||||||
const contextTagKey = "ctx"
|
const contextTagKey = "ctx"
|
||||||
|
|||||||
@@ -1,7 +1,14 @@
|
|||||||
package discov
|
package discov
|
||||||
|
|
||||||
import "github.com/tal-tech/go-zero/core/discov/internal"
|
import "github.com/zeromicro/go-zero/core/discov/internal"
|
||||||
|
|
||||||
|
// RegisterAccount registers the username/password to the given etcd cluster.
|
||||||
func RegisterAccount(endpoints []string, user, pass string) {
|
func RegisterAccount(endpoints []string, user, pass string) {
|
||||||
internal.AddAccount(endpoints, user, pass)
|
internal.AddAccount(endpoints, user, pass)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// RegisterTLS registers the CertFile/CertKeyFile/CACertFile to the given etcd.
|
||||||
|
func RegisterTLS(endpoints []string, certFile, certKeyFile, caFile string,
|
||||||
|
insecureSkipVerify bool) error {
|
||||||
|
return internal.AddTLS(endpoints, certFile, certKeyFile, caFile, insecureSkipVerify)
|
||||||
|
}
|
||||||
|
|||||||
@@ -4,8 +4,8 @@ import (
|
|||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/discov/internal"
|
"github.com/zeromicro/go-zero/core/discov/internal"
|
||||||
"github.com/tal-tech/go-zero/core/stringx"
|
"github.com/zeromicro/go-zero/core/stringx"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestRegisterAccount(t *testing.T) {
|
func TestRegisterAccount(t *testing.T) {
|
||||||
|
|||||||
@@ -4,7 +4,7 @@ import (
|
|||||||
"fmt"
|
"fmt"
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/discov/internal"
|
"github.com/zeromicro/go-zero/core/discov/internal"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
|
|||||||
@@ -5,7 +5,7 @@ import (
|
|||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/discov/internal"
|
"github.com/zeromicro/go-zero/core/discov/internal"
|
||||||
)
|
)
|
||||||
|
|
||||||
var mockLock sync.Mutex
|
var mockLock sync.Mutex
|
||||||
|
|||||||
@@ -4,10 +4,14 @@ import "errors"
|
|||||||
|
|
||||||
// EtcdConf is the config item with the given key on etcd.
|
// EtcdConf is the config item with the given key on etcd.
|
||||||
type EtcdConf struct {
|
type EtcdConf struct {
|
||||||
Hosts []string
|
Hosts []string
|
||||||
Key string
|
Key string
|
||||||
User string `json:",optional"`
|
User string `json:",optional"`
|
||||||
Pass string `json:",optional"`
|
Pass string `json:",optional"`
|
||||||
|
CertFile string `json:",optional"`
|
||||||
|
CertKeyFile string `json:",optional=CertFile"`
|
||||||
|
CACertFile string `json:",optional=CertFile"`
|
||||||
|
InsecureSkipVerify bool `json:",optional"`
|
||||||
}
|
}
|
||||||
|
|
||||||
// HasAccount returns if account provided.
|
// HasAccount returns if account provided.
|
||||||
@@ -15,6 +19,11 @@ func (c EtcdConf) HasAccount() bool {
|
|||||||
return len(c.User) > 0 && len(c.Pass) > 0
|
return len(c.User) > 0 && len(c.Pass) > 0
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// HasTLS returns if TLS CertFile/CertKeyFile/CACertFile are provided.
|
||||||
|
func (c EtcdConf) HasTLS() bool {
|
||||||
|
return len(c.CertFile) > 0 && len(c.CertKeyFile) > 0 && len(c.CACertFile) > 0
|
||||||
|
}
|
||||||
|
|
||||||
// Validate validates c.
|
// Validate validates c.
|
||||||
func (c EtcdConf) Validate() error {
|
func (c EtcdConf) Validate() error {
|
||||||
if len(c.Hosts) == 0 {
|
if len(c.Hosts) == 0 {
|
||||||
|
|||||||
@@ -1,17 +1,25 @@
|
|||||||
package internal
|
package internal
|
||||||
|
|
||||||
import "sync"
|
import (
|
||||||
|
"crypto/tls"
|
||||||
|
"crypto/x509"
|
||||||
|
"io/ioutil"
|
||||||
|
"sync"
|
||||||
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
accounts = make(map[string]Account)
|
||||||
|
tlsConfigs = make(map[string]*tls.Config)
|
||||||
|
lock sync.RWMutex
|
||||||
|
)
|
||||||
|
|
||||||
|
// Account holds the username/password for an etcd cluster.
|
||||||
type Account struct {
|
type Account struct {
|
||||||
User string
|
User string
|
||||||
Pass string
|
Pass string
|
||||||
}
|
}
|
||||||
|
|
||||||
var (
|
// AddAccount adds the username/password for the given etcd cluster.
|
||||||
accounts = make(map[string]Account)
|
|
||||||
lock sync.RWMutex
|
|
||||||
)
|
|
||||||
|
|
||||||
func AddAccount(endpoints []string, user, pass string) {
|
func AddAccount(endpoints []string, user, pass string) {
|
||||||
lock.Lock()
|
lock.Lock()
|
||||||
defer lock.Unlock()
|
defer lock.Unlock()
|
||||||
@@ -22,6 +30,33 @@ func AddAccount(endpoints []string, user, pass string) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// AddTLS adds the tls cert files for the given etcd cluster.
|
||||||
|
func AddTLS(endpoints []string, certFile, certKeyFile, caFile string, insecureSkipVerify bool) error {
|
||||||
|
cert, err := tls.LoadX509KeyPair(certFile, certKeyFile)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
caData, err := ioutil.ReadFile(caFile)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
pool := x509.NewCertPool()
|
||||||
|
pool.AppendCertsFromPEM(caData)
|
||||||
|
|
||||||
|
lock.Lock()
|
||||||
|
defer lock.Unlock()
|
||||||
|
tlsConfigs[getClusterKey(endpoints)] = &tls.Config{
|
||||||
|
Certificates: []tls.Certificate{cert},
|
||||||
|
RootCAs: pool,
|
||||||
|
InsecureSkipVerify: insecureSkipVerify,
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// GetAccount gets the username/password for the given etcd cluster.
|
||||||
func GetAccount(endpoints []string) (Account, bool) {
|
func GetAccount(endpoints []string) (Account, bool) {
|
||||||
lock.RLock()
|
lock.RLock()
|
||||||
defer lock.RUnlock()
|
defer lock.RUnlock()
|
||||||
@@ -29,3 +64,12 @@ func GetAccount(endpoints []string) (Account, bool) {
|
|||||||
account, ok := accounts[getClusterKey(endpoints)]
|
account, ok := accounts[getClusterKey(endpoints)]
|
||||||
return account, ok
|
return account, ok
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// GetTLS gets the tls config for the given etcd cluster.
|
||||||
|
func GetTLS(endpoints []string) (*tls.Config, bool) {
|
||||||
|
lock.RLock()
|
||||||
|
defer lock.RUnlock()
|
||||||
|
|
||||||
|
cfg, ok := tlsConfigs[getClusterKey(endpoints)]
|
||||||
|
return cfg, ok
|
||||||
|
}
|
||||||
|
|||||||
@@ -4,7 +4,7 @@ import (
|
|||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/stringx"
|
"github.com/zeromicro/go-zero/core/stringx"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestAccount(t *testing.T) {
|
func TestAccount(t *testing.T) {
|
||||||
|
|||||||
@@ -9,11 +9,11 @@ import (
|
|||||||
"sync"
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/contextx"
|
"github.com/zeromicro/go-zero/core/contextx"
|
||||||
"github.com/tal-tech/go-zero/core/lang"
|
"github.com/zeromicro/go-zero/core/lang"
|
||||||
"github.com/tal-tech/go-zero/core/logx"
|
"github.com/zeromicro/go-zero/core/logx"
|
||||||
"github.com/tal-tech/go-zero/core/syncx"
|
"github.com/zeromicro/go-zero/core/syncx"
|
||||||
"github.com/tal-tech/go-zero/core/threading"
|
"github.com/zeromicro/go-zero/core/threading"
|
||||||
clientv3 "go.etcd.io/etcd/client/v3"
|
clientv3 "go.etcd.io/etcd/client/v3"
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -37,25 +37,35 @@ func GetRegistry() *Registry {
|
|||||||
|
|
||||||
// GetConn returns an etcd client connection associated with given endpoints.
|
// GetConn returns an etcd client connection associated with given endpoints.
|
||||||
func (r *Registry) GetConn(endpoints []string) (EtcdClient, error) {
|
func (r *Registry) GetConn(endpoints []string) (EtcdClient, error) {
|
||||||
return r.getCluster(endpoints).getClient()
|
c, _ := r.getCluster(endpoints)
|
||||||
|
return c.getClient()
|
||||||
}
|
}
|
||||||
|
|
||||||
// Monitor monitors the key on given etcd endpoints, notify with the given UpdateListener.
|
// Monitor monitors the key on given etcd endpoints, notify with the given UpdateListener.
|
||||||
func (r *Registry) Monitor(endpoints []string, key string, l UpdateListener) error {
|
func (r *Registry) Monitor(endpoints []string, key string, l UpdateListener) error {
|
||||||
return r.getCluster(endpoints).monitor(key, l)
|
c, exists := r.getCluster(endpoints)
|
||||||
|
// if exists, the existing values should be updated to the listener.
|
||||||
|
if exists {
|
||||||
|
kvs := c.getCurrent(key)
|
||||||
|
for _, kv := range kvs {
|
||||||
|
l.OnAdd(kv)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return c.monitor(key, l)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (r *Registry) getCluster(endpoints []string) *cluster {
|
func (r *Registry) getCluster(endpoints []string) (c *cluster, exists bool) {
|
||||||
clusterKey := getClusterKey(endpoints)
|
clusterKey := getClusterKey(endpoints)
|
||||||
r.lock.Lock()
|
r.lock.Lock()
|
||||||
defer r.lock.Unlock()
|
defer r.lock.Unlock()
|
||||||
c, ok := r.clusters[clusterKey]
|
c, exists = r.clusters[clusterKey]
|
||||||
if !ok {
|
if !exists {
|
||||||
c = newCluster(endpoints)
|
c = newCluster(endpoints)
|
||||||
r.clusters[clusterKey] = c
|
r.clusters[clusterKey] = c
|
||||||
}
|
}
|
||||||
|
|
||||||
return c
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
type cluster struct {
|
type cluster struct {
|
||||||
@@ -94,6 +104,21 @@ func (c *cluster) getClient() (EtcdClient, error) {
|
|||||||
return val.(EtcdClient), nil
|
return val.(EtcdClient), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (c *cluster) getCurrent(key string) []KV {
|
||||||
|
c.lock.Lock()
|
||||||
|
defer c.lock.Unlock()
|
||||||
|
|
||||||
|
var kvs []KV
|
||||||
|
for k, v := range c.values[key] {
|
||||||
|
kvs = append(kvs, KV{
|
||||||
|
Key: k,
|
||||||
|
Val: v,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
return kvs
|
||||||
|
}
|
||||||
|
|
||||||
func (c *cluster) handleChanges(key string, kvs []KV) {
|
func (c *cluster) handleChanges(key string, kvs []KV) {
|
||||||
var add []KV
|
var add []KV
|
||||||
var remove []KV
|
var remove []KV
|
||||||
@@ -197,14 +222,12 @@ func (c *cluster) load(cli EtcdClient, key string) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
var kvs []KV
|
var kvs []KV
|
||||||
c.lock.Lock()
|
|
||||||
for _, ev := range resp.Kvs {
|
for _, ev := range resp.Kvs {
|
||||||
kvs = append(kvs, KV{
|
kvs = append(kvs, KV{
|
||||||
Key: string(ev.Key),
|
Key: string(ev.Key),
|
||||||
Val: string(ev.Value),
|
Val: string(ev.Value),
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
c.lock.Unlock()
|
|
||||||
|
|
||||||
c.handleChanges(key, kvs)
|
c.handleChanges(key, kvs)
|
||||||
}
|
}
|
||||||
@@ -314,6 +337,9 @@ func DialClient(endpoints []string) (EtcdClient, error) {
|
|||||||
cfg.Username = account.User
|
cfg.Username = account.User
|
||||||
cfg.Password = account.Pass
|
cfg.Password = account.Pass
|
||||||
}
|
}
|
||||||
|
if tlsCfg, ok := GetTLS(endpoints); ok {
|
||||||
|
cfg.TLS = tlsCfg
|
||||||
|
}
|
||||||
|
|
||||||
return clientv3.New(cfg)
|
return clientv3.New(cfg)
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -7,10 +7,10 @@ import (
|
|||||||
|
|
||||||
"github.com/golang/mock/gomock"
|
"github.com/golang/mock/gomock"
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/contextx"
|
"github.com/zeromicro/go-zero/core/contextx"
|
||||||
"github.com/tal-tech/go-zero/core/lang"
|
"github.com/zeromicro/go-zero/core/lang"
|
||||||
"github.com/tal-tech/go-zero/core/logx"
|
"github.com/zeromicro/go-zero/core/logx"
|
||||||
"github.com/tal-tech/go-zero/core/stringx"
|
"github.com/zeromicro/go-zero/core/stringx"
|
||||||
"go.etcd.io/etcd/api/v3/mvccpb"
|
"go.etcd.io/etcd/api/v3/mvccpb"
|
||||||
clientv3 "go.etcd.io/etcd/client/v3"
|
clientv3 "go.etcd.io/etcd/client/v3"
|
||||||
)
|
)
|
||||||
@@ -34,9 +34,9 @@ func setMockClient(cli EtcdClient) func() {
|
|||||||
|
|
||||||
func TestGetCluster(t *testing.T) {
|
func TestGetCluster(t *testing.T) {
|
||||||
AddAccount([]string{"first"}, "foo", "bar")
|
AddAccount([]string{"first"}, "foo", "bar")
|
||||||
c1 := GetRegistry().getCluster([]string{"first"})
|
c1, _ := GetRegistry().getCluster([]string{"first"})
|
||||||
c2 := GetRegistry().getCluster([]string{"second"})
|
c2, _ := GetRegistry().getCluster([]string{"second"})
|
||||||
c3 := GetRegistry().getCluster([]string{"first"})
|
c3, _ := GetRegistry().getCluster([]string{"first"})
|
||||||
assert.Equal(t, c1, c3)
|
assert.Equal(t, c1, c3)
|
||||||
assert.NotEqual(t, c1, c2)
|
assert.NotEqual(t, c1, c2)
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,12 +1,12 @@
|
|||||||
package discov
|
package discov
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"github.com/tal-tech/go-zero/core/discov/internal"
|
"github.com/zeromicro/go-zero/core/discov/internal"
|
||||||
"github.com/tal-tech/go-zero/core/lang"
|
"github.com/zeromicro/go-zero/core/lang"
|
||||||
"github.com/tal-tech/go-zero/core/logx"
|
"github.com/zeromicro/go-zero/core/logx"
|
||||||
"github.com/tal-tech/go-zero/core/proc"
|
"github.com/zeromicro/go-zero/core/proc"
|
||||||
"github.com/tal-tech/go-zero/core/syncx"
|
"github.com/zeromicro/go-zero/core/syncx"
|
||||||
"github.com/tal-tech/go-zero/core/threading"
|
"github.com/zeromicro/go-zero/core/threading"
|
||||||
clientv3 "go.etcd.io/etcd/client/v3"
|
clientv3 "go.etcd.io/etcd/client/v3"
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -145,16 +145,23 @@ func (p *Publisher) revoke(cli internal.EtcdClient) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// WithPubEtcdAccount provides the etcd username/password.
|
|
||||||
func WithPubEtcdAccount(user, pass string) PubOption {
|
|
||||||
return func(pub *Publisher) {
|
|
||||||
internal.AddAccount(pub.endpoints, user, pass)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// WithId customizes a Publisher with the id.
|
// WithId customizes a Publisher with the id.
|
||||||
func WithId(id int64) PubOption {
|
func WithId(id int64) PubOption {
|
||||||
return func(publisher *Publisher) {
|
return func(publisher *Publisher) {
|
||||||
publisher.id = id
|
publisher.id = id
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// WithPubEtcdAccount provides the etcd username/password.
|
||||||
|
func WithPubEtcdAccount(user, pass string) PubOption {
|
||||||
|
return func(pub *Publisher) {
|
||||||
|
RegisterAccount(pub.endpoints, user, pass)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// WithPubEtcdTLS provides the etcd CertFile/CertKeyFile/CACertFile.
|
||||||
|
func WithPubEtcdTLS(certFile, certKeyFile, caFile string, insecureSkipVerify bool) PubOption {
|
||||||
|
return func(pub *Publisher) {
|
||||||
|
logx.Must(RegisterTLS(pub.endpoints, certFile, certKeyFile, caFile, insecureSkipVerify))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|||||||
@@ -8,10 +8,10 @@ import (
|
|||||||
|
|
||||||
"github.com/golang/mock/gomock"
|
"github.com/golang/mock/gomock"
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/discov/internal"
|
"github.com/zeromicro/go-zero/core/discov/internal"
|
||||||
"github.com/tal-tech/go-zero/core/lang"
|
"github.com/zeromicro/go-zero/core/lang"
|
||||||
"github.com/tal-tech/go-zero/core/logx"
|
"github.com/zeromicro/go-zero/core/logx"
|
||||||
"github.com/tal-tech/go-zero/core/stringx"
|
"github.com/zeromicro/go-zero/core/stringx"
|
||||||
clientv3 "go.etcd.io/etcd/client/v3"
|
clientv3 "go.etcd.io/etcd/client/v3"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|||||||
@@ -4,8 +4,9 @@ import (
|
|||||||
"sync"
|
"sync"
|
||||||
"sync/atomic"
|
"sync/atomic"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/discov/internal"
|
"github.com/zeromicro/go-zero/core/discov/internal"
|
||||||
"github.com/tal-tech/go-zero/core/syncx"
|
"github.com/zeromicro/go-zero/core/logx"
|
||||||
|
"github.com/zeromicro/go-zero/core/syncx"
|
||||||
)
|
)
|
||||||
|
|
||||||
type (
|
type (
|
||||||
@@ -58,9 +59,17 @@ func Exclusive() SubOption {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// WithSubEtcdAccount provides the etcd username/password.
|
||||||
func WithSubEtcdAccount(user, pass string) SubOption {
|
func WithSubEtcdAccount(user, pass string) SubOption {
|
||||||
return func(sub *Subscriber) {
|
return func(sub *Subscriber) {
|
||||||
internal.AddAccount(sub.endpoints, user, pass)
|
RegisterAccount(sub.endpoints, user, pass)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// WithSubEtcdTLS provides the etcd CertFile/CertKeyFile/CACertFile.
|
||||||
|
func WithSubEtcdTLS(certFile, certKeyFile, caFile string, insecureSkipVerify bool) SubOption {
|
||||||
|
return func(sub *Subscriber) {
|
||||||
|
logx.Must(RegisterTLS(sub.endpoints, certFile, certKeyFile, caFile, insecureSkipVerify))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -5,8 +5,8 @@ import (
|
|||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/discov/internal"
|
"github.com/zeromicro/go-zero/core/discov/internal"
|
||||||
"github.com/tal-tech/go-zero/core/stringx"
|
"github.com/zeromicro/go-zero/core/stringx"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
|
|||||||
@@ -4,7 +4,7 @@ import (
|
|||||||
"sync"
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/threading"
|
"github.com/zeromicro/go-zero/core/threading"
|
||||||
)
|
)
|
||||||
|
|
||||||
// A DelayExecutor delays a tasks on given delay interval.
|
// A DelayExecutor delays a tasks on given delay interval.
|
||||||
|
|||||||
@@ -3,8 +3,8 @@ package executors
|
|||||||
import (
|
import (
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/syncx"
|
"github.com/zeromicro/go-zero/core/syncx"
|
||||||
"github.com/tal-tech/go-zero/core/timex"
|
"github.com/zeromicro/go-zero/core/timex"
|
||||||
)
|
)
|
||||||
|
|
||||||
// A LessExecutor is an executor to limit execution once within given time interval.
|
// A LessExecutor is an executor to limit execution once within given time interval.
|
||||||
|
|||||||
@@ -5,7 +5,7 @@ import (
|
|||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/timex"
|
"github.com/zeromicro/go-zero/core/timex"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestLessExecutor_DoOrDiscard(t *testing.T) {
|
func TestLessExecutor_DoOrDiscard(t *testing.T) {
|
||||||
|
|||||||
@@ -6,11 +6,11 @@ import (
|
|||||||
"sync/atomic"
|
"sync/atomic"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/lang"
|
"github.com/zeromicro/go-zero/core/lang"
|
||||||
"github.com/tal-tech/go-zero/core/proc"
|
"github.com/zeromicro/go-zero/core/proc"
|
||||||
"github.com/tal-tech/go-zero/core/syncx"
|
"github.com/zeromicro/go-zero/core/syncx"
|
||||||
"github.com/tal-tech/go-zero/core/threading"
|
"github.com/zeromicro/go-zero/core/threading"
|
||||||
"github.com/tal-tech/go-zero/core/timex"
|
"github.com/zeromicro/go-zero/core/timex"
|
||||||
)
|
)
|
||||||
|
|
||||||
const idleRound = 10
|
const idleRound = 10
|
||||||
|
|||||||
@@ -8,7 +8,7 @@ import (
|
|||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/timex"
|
"github.com/zeromicro/go-zero/core/timex"
|
||||||
)
|
)
|
||||||
|
|
||||||
const threshold = 10
|
const threshold = 10
|
||||||
|
|||||||
@@ -5,7 +5,7 @@ import (
|
|||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/fs"
|
"github.com/zeromicro/go-zero/core/fs"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
|
|||||||
@@ -5,7 +5,7 @@ import (
|
|||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/fs"
|
"github.com/zeromicro/go-zero/core/fs"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestSplitLineChunks(t *testing.T) {
|
func TestSplitLineChunks(t *testing.T) {
|
||||||
|
|||||||
@@ -5,7 +5,7 @@ import (
|
|||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/fs"
|
"github.com/zeromicro/go-zero/core/fs"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestRangeReader(t *testing.T) {
|
func TestRangeReader(t *testing.T) {
|
||||||
|
|||||||
@@ -4,7 +4,7 @@ import (
|
|||||||
"io/ioutil"
|
"io/ioutil"
|
||||||
"os"
|
"os"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/hash"
|
"github.com/zeromicro/go-zero/core/hash"
|
||||||
)
|
)
|
||||||
|
|
||||||
// TempFileWithText creates the temporary file with the given content,
|
// TempFileWithText creates the temporary file with the given content,
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
package fx
|
package fx
|
||||||
|
|
||||||
import "github.com/tal-tech/go-zero/core/threading"
|
import "github.com/zeromicro/go-zero/core/threading"
|
||||||
|
|
||||||
// Parallel runs fns parallelly and waits for done.
|
// Parallel runs fns parallelly and waits for done.
|
||||||
func Parallel(fns ...func()) {
|
func Parallel(fns ...func()) {
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
package fx
|
package fx
|
||||||
|
|
||||||
import "github.com/tal-tech/go-zero/core/errorx"
|
import "github.com/zeromicro/go-zero/core/errorx"
|
||||||
|
|
||||||
const defaultRetryTimes = 3
|
const defaultRetryTimes = 3
|
||||||
|
|
||||||
|
|||||||
@@ -4,9 +4,9 @@ import (
|
|||||||
"sort"
|
"sort"
|
||||||
"sync"
|
"sync"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/collection"
|
"github.com/zeromicro/go-zero/core/collection"
|
||||||
"github.com/tal-tech/go-zero/core/lang"
|
"github.com/zeromicro/go-zero/core/lang"
|
||||||
"github.com/tal-tech/go-zero/core/threading"
|
"github.com/zeromicro/go-zero/core/threading"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
@@ -90,6 +90,8 @@ func Range(source <-chan interface{}) Stream {
|
|||||||
func (s Stream) AllMach(predicate func(item interface{}) bool) bool {
|
func (s Stream) AllMach(predicate func(item interface{}) bool) bool {
|
||||||
for item := range s.source {
|
for item := range s.source {
|
||||||
if !predicate(item) {
|
if !predicate(item) {
|
||||||
|
// make sure the former goroutine not block, and current func returns fast.
|
||||||
|
go drain(s.source)
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -103,6 +105,8 @@ func (s Stream) AllMach(predicate func(item interface{}) bool) bool {
|
|||||||
func (s Stream) AnyMach(predicate func(item interface{}) bool) bool {
|
func (s Stream) AnyMach(predicate func(item interface{}) bool) bool {
|
||||||
for item := range s.source {
|
for item := range s.source {
|
||||||
if predicate(item) {
|
if predicate(item) {
|
||||||
|
// make sure the former goroutine not block, and current func returns fast.
|
||||||
|
go drain(s.source)
|
||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -186,8 +190,7 @@ func (s Stream) Distinct(fn KeyFunc) Stream {
|
|||||||
|
|
||||||
// Done waits all upstreaming operations to be done.
|
// Done waits all upstreaming operations to be done.
|
||||||
func (s Stream) Done() {
|
func (s Stream) Done() {
|
||||||
for range s.source {
|
drain(s.source)
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Filter filters the items by the given FilterFunc.
|
// Filter filters the items by the given FilterFunc.
|
||||||
@@ -199,9 +202,22 @@ func (s Stream) Filter(fn FilterFunc, opts ...Option) Stream {
|
|||||||
}, opts...)
|
}, opts...)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// First returns the first item, nil if no items.
|
||||||
|
func (s Stream) First() interface{} {
|
||||||
|
for item := range s.source {
|
||||||
|
// make sure the former goroutine not block, and current func returns fast.
|
||||||
|
go drain(s.source)
|
||||||
|
return item
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
// ForAll handles the streaming elements from the source and no later streams.
|
// ForAll handles the streaming elements from the source and no later streams.
|
||||||
func (s Stream) ForAll(fn ForAllFunc) {
|
func (s Stream) ForAll(fn ForAllFunc) {
|
||||||
fn(s.source)
|
fn(s.source)
|
||||||
|
// avoid goroutine leak on fn not consuming all items.
|
||||||
|
go drain(s.source)
|
||||||
}
|
}
|
||||||
|
|
||||||
// ForEach seals the Stream with the ForEachFunc on each item, no successive operations.
|
// ForEach seals the Stream with the ForEachFunc on each item, no successive operations.
|
||||||
@@ -246,11 +262,14 @@ func (s Stream) Head(n int64) Stream {
|
|||||||
}
|
}
|
||||||
if n == 0 {
|
if n == 0 {
|
||||||
// let successive method go ASAP even we have more items to skip
|
// let successive method go ASAP even we have more items to skip
|
||||||
// why we don't just break the loop, because if break,
|
|
||||||
// this former goroutine will block forever, which will cause goroutine leak.
|
|
||||||
close(source)
|
close(source)
|
||||||
|
// why we don't just break the loop, and drain to consume all items.
|
||||||
|
// because if breaks, this former goroutine will block forever,
|
||||||
|
// which will cause goroutine leak.
|
||||||
|
drain(s.source)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
// not enough items in s.source, but we need to let successive method to go ASAP.
|
||||||
if n > 0 {
|
if n > 0 {
|
||||||
close(source)
|
close(source)
|
||||||
}
|
}
|
||||||
@@ -259,6 +278,13 @@ func (s Stream) Head(n int64) Stream {
|
|||||||
return Range(source)
|
return Range(source)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Last returns the last item, or nil if no items.
|
||||||
|
func (s Stream) Last() (item interface{}) {
|
||||||
|
for item = range s.source {
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
// Map converts each item to another corresponding item, which means it's a 1:1 model.
|
// Map converts each item to another corresponding item, which means it's a 1:1 model.
|
||||||
func (s Stream) Map(fn MapFunc, opts ...Option) Stream {
|
func (s Stream) Map(fn MapFunc, opts ...Option) Stream {
|
||||||
return s.Walk(func(item interface{}, pipe chan<- interface{}) {
|
return s.Walk(func(item interface{}, pipe chan<- interface{}) {
|
||||||
@@ -280,6 +306,21 @@ func (s Stream) Merge() Stream {
|
|||||||
return Range(source)
|
return Range(source)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// NoneMatch returns whether all elements of this stream don't match the provided predicate.
|
||||||
|
// May not evaluate the predicate on all elements if not necessary for determining the result.
|
||||||
|
// If the stream is empty then true is returned and the predicate is not evaluated.
|
||||||
|
func (s Stream) NoneMatch(predicate func(item interface{}) bool) bool {
|
||||||
|
for item := range s.source {
|
||||||
|
if predicate(item) {
|
||||||
|
// make sure the former goroutine not block, and current func returns fast.
|
||||||
|
go drain(s.source)
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
// Parallel applies the given ParallelFunc to each item concurrently with given number of workers.
|
// Parallel applies the given ParallelFunc to each item concurrently with given number of workers.
|
||||||
func (s Stream) Parallel(fn ParallelFunc, opts ...Option) {
|
func (s Stream) Parallel(fn ParallelFunc, opts ...Option) {
|
||||||
s.Walk(func(item interface{}, pipe chan<- interface{}) {
|
s.Walk(func(item interface{}, pipe chan<- interface{}) {
|
||||||
@@ -411,15 +452,12 @@ func (s Stream) walkLimited(fn WalkFunc, option *rxOptions) Stream {
|
|||||||
var wg sync.WaitGroup
|
var wg sync.WaitGroup
|
||||||
pool := make(chan lang.PlaceholderType, option.workers)
|
pool := make(chan lang.PlaceholderType, option.workers)
|
||||||
|
|
||||||
for {
|
for item := range s.source {
|
||||||
|
// important, used in another goroutine
|
||||||
|
val := item
|
||||||
pool <- lang.Placeholder
|
pool <- lang.Placeholder
|
||||||
item, ok := <-s.source
|
|
||||||
if !ok {
|
|
||||||
<-pool
|
|
||||||
break
|
|
||||||
}
|
|
||||||
|
|
||||||
wg.Add(1)
|
wg.Add(1)
|
||||||
|
|
||||||
// better to safely run caller defined method
|
// better to safely run caller defined method
|
||||||
threading.GoSafe(func() {
|
threading.GoSafe(func() {
|
||||||
defer func() {
|
defer func() {
|
||||||
@@ -427,7 +465,7 @@ func (s Stream) walkLimited(fn WalkFunc, option *rxOptions) Stream {
|
|||||||
<-pool
|
<-pool
|
||||||
}()
|
}()
|
||||||
|
|
||||||
fn(item, pipe)
|
fn(val, pipe)
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -439,22 +477,19 @@ func (s Stream) walkLimited(fn WalkFunc, option *rxOptions) Stream {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (s Stream) walkUnlimited(fn WalkFunc, option *rxOptions) Stream {
|
func (s Stream) walkUnlimited(fn WalkFunc, option *rxOptions) Stream {
|
||||||
pipe := make(chan interface{}, defaultWorkers)
|
pipe := make(chan interface{}, option.workers)
|
||||||
|
|
||||||
go func() {
|
go func() {
|
||||||
var wg sync.WaitGroup
|
var wg sync.WaitGroup
|
||||||
|
|
||||||
for {
|
for item := range s.source {
|
||||||
item, ok := <-s.source
|
// important, used in another goroutine
|
||||||
if !ok {
|
val := item
|
||||||
break
|
|
||||||
}
|
|
||||||
|
|
||||||
wg.Add(1)
|
wg.Add(1)
|
||||||
// better to safely run caller defined method
|
// better to safely run caller defined method
|
||||||
threading.GoSafe(func() {
|
threading.GoSafe(func() {
|
||||||
defer wg.Done()
|
defer wg.Done()
|
||||||
fn(item, pipe)
|
fn(val, pipe)
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -465,14 +500,14 @@ func (s Stream) walkUnlimited(fn WalkFunc, option *rxOptions) Stream {
|
|||||||
return Range(pipe)
|
return Range(pipe)
|
||||||
}
|
}
|
||||||
|
|
||||||
// UnlimitedWorkers lets the caller to use as many workers as the tasks.
|
// UnlimitedWorkers lets the caller use as many workers as the tasks.
|
||||||
func UnlimitedWorkers() Option {
|
func UnlimitedWorkers() Option {
|
||||||
return func(opts *rxOptions) {
|
return func(opts *rxOptions) {
|
||||||
opts.unlimitedWorkers = true
|
opts.unlimitedWorkers = true
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// WithWorkers lets the caller to customize the concurrent workers.
|
// WithWorkers lets the caller customize the concurrent workers.
|
||||||
func WithWorkers(workers int) Option {
|
func WithWorkers(workers int) Option {
|
||||||
return func(opts *rxOptions) {
|
return func(opts *rxOptions) {
|
||||||
if workers < minWorkers {
|
if workers < minWorkers {
|
||||||
@@ -483,6 +518,7 @@ func WithWorkers(workers int) Option {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// buildOptions returns a rxOptions with given customizations.
|
||||||
func buildOptions(opts ...Option) *rxOptions {
|
func buildOptions(opts ...Option) *rxOptions {
|
||||||
options := newOptions()
|
options := newOptions()
|
||||||
for _, opt := range opts {
|
for _, opt := range opts {
|
||||||
@@ -492,6 +528,13 @@ func buildOptions(opts ...Option) *rxOptions {
|
|||||||
return options
|
return options
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// drain drains the given channel.
|
||||||
|
func drain(channel <-chan interface{}) {
|
||||||
|
for range channel {
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// newOptions returns a default rxOptions.
|
||||||
func newOptions() *rxOptions {
|
func newOptions() *rxOptions {
|
||||||
return &rxOptions{
|
return &rxOptions{
|
||||||
workers: defaultWorkers,
|
workers: defaultWorkers,
|
||||||
|
|||||||
@@ -13,324 +13,494 @@ import (
|
|||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/stringx"
|
"github.com/zeromicro/go-zero/core/stringx"
|
||||||
|
"go.uber.org/goleak"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestBuffer(t *testing.T) {
|
func TestBuffer(t *testing.T) {
|
||||||
const N = 5
|
runCheckedTest(t, func(t *testing.T) {
|
||||||
var count int32
|
const N = 5
|
||||||
var wait sync.WaitGroup
|
var count int32
|
||||||
wait.Add(1)
|
var wait sync.WaitGroup
|
||||||
From(func(source chan<- interface{}) {
|
wait.Add(1)
|
||||||
ticker := time.NewTicker(10 * time.Millisecond)
|
From(func(source chan<- interface{}) {
|
||||||
defer ticker.Stop()
|
ticker := time.NewTicker(10 * time.Millisecond)
|
||||||
|
defer ticker.Stop()
|
||||||
|
|
||||||
for i := 0; i < 2*N; i++ {
|
for i := 0; i < 2*N; i++ {
|
||||||
select {
|
select {
|
||||||
case source <- i:
|
case source <- i:
|
||||||
atomic.AddInt32(&count, 1)
|
atomic.AddInt32(&count, 1)
|
||||||
case <-ticker.C:
|
case <-ticker.C:
|
||||||
wait.Done()
|
wait.Done()
|
||||||
return
|
return
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}).Buffer(N).ForAll(func(pipe <-chan interface{}) {
|
||||||
}).Buffer(N).ForAll(func(pipe <-chan interface{}) {
|
wait.Wait()
|
||||||
wait.Wait()
|
// why N+1, because take one more to wait for sending into the channel
|
||||||
// why N+1, because take one more to wait for sending into the channel
|
assert.Equal(t, int32(N+1), atomic.LoadInt32(&count))
|
||||||
assert.Equal(t, int32(N+1), atomic.LoadInt32(&count))
|
})
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestBufferNegative(t *testing.T) {
|
func TestBufferNegative(t *testing.T) {
|
||||||
var result int
|
runCheckedTest(t, func(t *testing.T) {
|
||||||
Just(1, 2, 3, 4).Buffer(-1).Reduce(func(pipe <-chan interface{}) (interface{}, error) {
|
var result int
|
||||||
for item := range pipe {
|
Just(1, 2, 3, 4).Buffer(-1).Reduce(func(pipe <-chan interface{}) (interface{}, error) {
|
||||||
result += item.(int)
|
for item := range pipe {
|
||||||
}
|
result += item.(int)
|
||||||
return result, nil
|
}
|
||||||
|
return result, nil
|
||||||
|
})
|
||||||
|
assert.Equal(t, 10, result)
|
||||||
})
|
})
|
||||||
assert.Equal(t, 10, result)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestCount(t *testing.T) {
|
func TestCount(t *testing.T) {
|
||||||
tests := []struct {
|
runCheckedTest(t, func(t *testing.T) {
|
||||||
name string
|
tests := []struct {
|
||||||
elements []interface{}
|
name string
|
||||||
}{
|
elements []interface{}
|
||||||
{
|
}{
|
||||||
name: "no elements with nil",
|
{
|
||||||
},
|
name: "no elements with nil",
|
||||||
{
|
},
|
||||||
name: "no elements",
|
{
|
||||||
elements: []interface{}{},
|
name: "no elements",
|
||||||
},
|
elements: []interface{}{},
|
||||||
{
|
},
|
||||||
name: "1 element",
|
{
|
||||||
elements: []interface{}{1},
|
name: "1 element",
|
||||||
},
|
elements: []interface{}{1},
|
||||||
{
|
},
|
||||||
name: "multiple elements",
|
{
|
||||||
elements: []interface{}{1, 2, 3},
|
name: "multiple elements",
|
||||||
},
|
elements: []interface{}{1, 2, 3},
|
||||||
}
|
},
|
||||||
|
}
|
||||||
|
|
||||||
for _, test := range tests {
|
for _, test := range tests {
|
||||||
t.Run(test.name, func(t *testing.T) {
|
t.Run(test.name, func(t *testing.T) {
|
||||||
val := Just(test.elements...).Count()
|
val := Just(test.elements...).Count()
|
||||||
assert.Equal(t, len(test.elements), val)
|
assert.Equal(t, len(test.elements), val)
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestDone(t *testing.T) {
|
func TestDone(t *testing.T) {
|
||||||
var count int32
|
runCheckedTest(t, func(t *testing.T) {
|
||||||
Just(1, 2, 3).Walk(func(item interface{}, pipe chan<- interface{}) {
|
var count int32
|
||||||
time.Sleep(time.Millisecond * 100)
|
Just(1, 2, 3).Walk(func(item interface{}, pipe chan<- interface{}) {
|
||||||
atomic.AddInt32(&count, int32(item.(int)))
|
time.Sleep(time.Millisecond * 100)
|
||||||
}).Done()
|
atomic.AddInt32(&count, int32(item.(int)))
|
||||||
assert.Equal(t, int32(6), count)
|
}).Done()
|
||||||
|
assert.Equal(t, int32(6), count)
|
||||||
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestJust(t *testing.T) {
|
func TestJust(t *testing.T) {
|
||||||
var result int
|
runCheckedTest(t, func(t *testing.T) {
|
||||||
Just(1, 2, 3, 4).Reduce(func(pipe <-chan interface{}) (interface{}, error) {
|
var result int
|
||||||
for item := range pipe {
|
Just(1, 2, 3, 4).Reduce(func(pipe <-chan interface{}) (interface{}, error) {
|
||||||
result += item.(int)
|
for item := range pipe {
|
||||||
}
|
result += item.(int)
|
||||||
return result, nil
|
}
|
||||||
|
return result, nil
|
||||||
|
})
|
||||||
|
assert.Equal(t, 10, result)
|
||||||
})
|
})
|
||||||
assert.Equal(t, 10, result)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestDistinct(t *testing.T) {
|
func TestDistinct(t *testing.T) {
|
||||||
var result int
|
runCheckedTest(t, func(t *testing.T) {
|
||||||
Just(4, 1, 3, 2, 3, 4).Distinct(func(item interface{}) interface{} {
|
var result int
|
||||||
return item
|
Just(4, 1, 3, 2, 3, 4).Distinct(func(item interface{}) interface{} {
|
||||||
}).Reduce(func(pipe <-chan interface{}) (interface{}, error) {
|
return item
|
||||||
for item := range pipe {
|
}).Reduce(func(pipe <-chan interface{}) (interface{}, error) {
|
||||||
result += item.(int)
|
for item := range pipe {
|
||||||
}
|
result += item.(int)
|
||||||
return result, nil
|
}
|
||||||
|
return result, nil
|
||||||
|
})
|
||||||
|
assert.Equal(t, 10, result)
|
||||||
})
|
})
|
||||||
assert.Equal(t, 10, result)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestFilter(t *testing.T) {
|
func TestFilter(t *testing.T) {
|
||||||
var result int
|
runCheckedTest(t, func(t *testing.T) {
|
||||||
Just(1, 2, 3, 4).Filter(func(item interface{}) bool {
|
var result int
|
||||||
return item.(int)%2 == 0
|
Just(1, 2, 3, 4).Filter(func(item interface{}) bool {
|
||||||
}).Reduce(func(pipe <-chan interface{}) (interface{}, error) {
|
return item.(int)%2 == 0
|
||||||
for item := range pipe {
|
}).Reduce(func(pipe <-chan interface{}) (interface{}, error) {
|
||||||
result += item.(int)
|
for item := range pipe {
|
||||||
}
|
result += item.(int)
|
||||||
return result, nil
|
}
|
||||||
|
return result, nil
|
||||||
|
})
|
||||||
|
assert.Equal(t, 6, result)
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestFirst(t *testing.T) {
|
||||||
|
runCheckedTest(t, func(t *testing.T) {
|
||||||
|
assert.Nil(t, Just().First())
|
||||||
|
assert.Equal(t, "foo", Just("foo").First())
|
||||||
|
assert.Equal(t, "foo", Just("foo", "bar").First())
|
||||||
})
|
})
|
||||||
assert.Equal(t, 6, result)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestForAll(t *testing.T) {
|
func TestForAll(t *testing.T) {
|
||||||
var result int
|
runCheckedTest(t, func(t *testing.T) {
|
||||||
Just(1, 2, 3, 4).Filter(func(item interface{}) bool {
|
var result int
|
||||||
return item.(int)%2 == 0
|
Just(1, 2, 3, 4).Filter(func(item interface{}) bool {
|
||||||
}).ForAll(func(pipe <-chan interface{}) {
|
return item.(int)%2 == 0
|
||||||
for item := range pipe {
|
}).ForAll(func(pipe <-chan interface{}) {
|
||||||
result += item.(int)
|
for item := range pipe {
|
||||||
}
|
result += item.(int)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
assert.Equal(t, 6, result)
|
||||||
})
|
})
|
||||||
assert.Equal(t, 6, result)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestGroup(t *testing.T) {
|
func TestGroup(t *testing.T) {
|
||||||
var groups [][]int
|
runCheckedTest(t, func(t *testing.T) {
|
||||||
Just(10, 11, 20, 21).Group(func(item interface{}) interface{} {
|
var groups [][]int
|
||||||
v := item.(int)
|
Just(10, 11, 20, 21).Group(func(item interface{}) interface{} {
|
||||||
return v / 10
|
v := item.(int)
|
||||||
}).ForEach(func(item interface{}) {
|
return v / 10
|
||||||
v := item.([]interface{})
|
}).ForEach(func(item interface{}) {
|
||||||
var group []int
|
v := item.([]interface{})
|
||||||
for _, each := range v {
|
var group []int
|
||||||
group = append(group, each.(int))
|
for _, each := range v {
|
||||||
}
|
group = append(group, each.(int))
|
||||||
groups = append(groups, group)
|
}
|
||||||
})
|
groups = append(groups, group)
|
||||||
|
})
|
||||||
|
|
||||||
assert.Equal(t, 2, len(groups))
|
assert.Equal(t, 2, len(groups))
|
||||||
for _, group := range groups {
|
for _, group := range groups {
|
||||||
assert.Equal(t, 2, len(group))
|
assert.Equal(t, 2, len(group))
|
||||||
assert.True(t, group[0]/10 == group[1]/10)
|
assert.True(t, group[0]/10 == group[1]/10)
|
||||||
}
|
}
|
||||||
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestHead(t *testing.T) {
|
func TestHead(t *testing.T) {
|
||||||
var result int
|
runCheckedTest(t, func(t *testing.T) {
|
||||||
Just(1, 2, 3, 4).Head(2).Reduce(func(pipe <-chan interface{}) (interface{}, error) {
|
var result int
|
||||||
for item := range pipe {
|
Just(1, 2, 3, 4).Head(2).Reduce(func(pipe <-chan interface{}) (interface{}, error) {
|
||||||
result += item.(int)
|
for item := range pipe {
|
||||||
}
|
result += item.(int)
|
||||||
return result, nil
|
}
|
||||||
|
return result, nil
|
||||||
|
})
|
||||||
|
assert.Equal(t, 3, result)
|
||||||
})
|
})
|
||||||
assert.Equal(t, 3, result)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestHeadZero(t *testing.T) {
|
func TestHeadZero(t *testing.T) {
|
||||||
assert.Panics(t, func() {
|
runCheckedTest(t, func(t *testing.T) {
|
||||||
Just(1, 2, 3, 4).Head(0).Reduce(func(pipe <-chan interface{}) (interface{}, error) {
|
assert.Panics(t, func() {
|
||||||
return nil, nil
|
Just(1, 2, 3, 4).Head(0).Reduce(func(pipe <-chan interface{}) (interface{}, error) {
|
||||||
|
return nil, nil
|
||||||
|
})
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestHeadMore(t *testing.T) {
|
func TestHeadMore(t *testing.T) {
|
||||||
var result int
|
runCheckedTest(t, func(t *testing.T) {
|
||||||
Just(1, 2, 3, 4).Head(6).Reduce(func(pipe <-chan interface{}) (interface{}, error) {
|
var result int
|
||||||
for item := range pipe {
|
Just(1, 2, 3, 4).Head(6).Reduce(func(pipe <-chan interface{}) (interface{}, error) {
|
||||||
result += item.(int)
|
for item := range pipe {
|
||||||
}
|
result += item.(int)
|
||||||
return result, nil
|
}
|
||||||
|
return result, nil
|
||||||
|
})
|
||||||
|
assert.Equal(t, 10, result)
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestLast(t *testing.T) {
|
||||||
|
runCheckedTest(t, func(t *testing.T) {
|
||||||
|
goroutines := runtime.NumGoroutine()
|
||||||
|
assert.Nil(t, Just().Last())
|
||||||
|
assert.Equal(t, "foo", Just("foo").Last())
|
||||||
|
assert.Equal(t, "bar", Just("foo", "bar").Last())
|
||||||
|
// let scheduler schedule first
|
||||||
|
runtime.Gosched()
|
||||||
|
assert.Equal(t, goroutines, runtime.NumGoroutine())
|
||||||
})
|
})
|
||||||
assert.Equal(t, 10, result)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestMap(t *testing.T) {
|
func TestMap(t *testing.T) {
|
||||||
log.SetOutput(ioutil.Discard)
|
runCheckedTest(t, func(t *testing.T) {
|
||||||
|
log.SetOutput(ioutil.Discard)
|
||||||
|
|
||||||
tests := []struct {
|
tests := []struct {
|
||||||
mapper MapFunc
|
mapper MapFunc
|
||||||
expect int
|
expect int
|
||||||
}{
|
}{
|
||||||
{
|
{
|
||||||
mapper: func(item interface{}) interface{} {
|
mapper: func(item interface{}) interface{} {
|
||||||
v := item.(int)
|
v := item.(int)
|
||||||
return v * v
|
return v * v
|
||||||
|
},
|
||||||
|
expect: 30,
|
||||||
},
|
},
|
||||||
expect: 30,
|
{
|
||||||
},
|
mapper: func(item interface{}) interface{} {
|
||||||
{
|
v := item.(int)
|
||||||
mapper: func(item interface{}) interface{} {
|
if v%2 == 0 {
|
||||||
v := item.(int)
|
return 0
|
||||||
if v%2 == 0 {
|
|
||||||
return 0
|
|
||||||
}
|
|
||||||
return v * v
|
|
||||||
},
|
|
||||||
expect: 10,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
mapper: func(item interface{}) interface{} {
|
|
||||||
v := item.(int)
|
|
||||||
if v%2 == 0 {
|
|
||||||
panic(v)
|
|
||||||
}
|
|
||||||
return v * v
|
|
||||||
},
|
|
||||||
expect: 10,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
// Map(...) works even WithWorkers(0)
|
|
||||||
for i, test := range tests {
|
|
||||||
t.Run(stringx.Rand(), func(t *testing.T) {
|
|
||||||
var result int
|
|
||||||
var workers int
|
|
||||||
if i%2 == 0 {
|
|
||||||
workers = 0
|
|
||||||
} else {
|
|
||||||
workers = runtime.NumCPU()
|
|
||||||
}
|
|
||||||
From(func(source chan<- interface{}) {
|
|
||||||
for i := 1; i < 5; i++ {
|
|
||||||
source <- i
|
|
||||||
}
|
|
||||||
}).Map(test.mapper, WithWorkers(workers)).Reduce(
|
|
||||||
func(pipe <-chan interface{}) (interface{}, error) {
|
|
||||||
for item := range pipe {
|
|
||||||
result += item.(int)
|
|
||||||
}
|
}
|
||||||
return result, nil
|
return v * v
|
||||||
})
|
},
|
||||||
|
expect: 10,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
mapper: func(item interface{}) interface{} {
|
||||||
|
v := item.(int)
|
||||||
|
if v%2 == 0 {
|
||||||
|
panic(v)
|
||||||
|
}
|
||||||
|
return v * v
|
||||||
|
},
|
||||||
|
expect: 10,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
assert.Equal(t, test.expect, result)
|
// Map(...) works even WithWorkers(0)
|
||||||
})
|
for i, test := range tests {
|
||||||
}
|
t.Run(stringx.Rand(), func(t *testing.T) {
|
||||||
|
var result int
|
||||||
|
var workers int
|
||||||
|
if i%2 == 0 {
|
||||||
|
workers = 0
|
||||||
|
} else {
|
||||||
|
workers = runtime.NumCPU()
|
||||||
|
}
|
||||||
|
From(func(source chan<- interface{}) {
|
||||||
|
for i := 1; i < 5; i++ {
|
||||||
|
source <- i
|
||||||
|
}
|
||||||
|
}).Map(test.mapper, WithWorkers(workers)).Reduce(
|
||||||
|
func(pipe <-chan interface{}) (interface{}, error) {
|
||||||
|
for item := range pipe {
|
||||||
|
result += item.(int)
|
||||||
|
}
|
||||||
|
return result, nil
|
||||||
|
})
|
||||||
|
|
||||||
|
assert.Equal(t, test.expect, result)
|
||||||
|
})
|
||||||
|
}
|
||||||
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestMerge(t *testing.T) {
|
func TestMerge(t *testing.T) {
|
||||||
Just(1, 2, 3, 4).Merge().ForEach(func(item interface{}) {
|
runCheckedTest(t, func(t *testing.T) {
|
||||||
assert.ElementsMatch(t, []interface{}{1, 2, 3, 4}, item.([]interface{}))
|
Just(1, 2, 3, 4).Merge().ForEach(func(item interface{}) {
|
||||||
|
assert.ElementsMatch(t, []interface{}{1, 2, 3, 4}, item.([]interface{}))
|
||||||
|
})
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestParallelJust(t *testing.T) {
|
func TestParallelJust(t *testing.T) {
|
||||||
var count int32
|
runCheckedTest(t, func(t *testing.T) {
|
||||||
Just(1, 2, 3).Parallel(func(item interface{}) {
|
var count int32
|
||||||
time.Sleep(time.Millisecond * 100)
|
Just(1, 2, 3).Parallel(func(item interface{}) {
|
||||||
atomic.AddInt32(&count, int32(item.(int)))
|
time.Sleep(time.Millisecond * 100)
|
||||||
}, UnlimitedWorkers())
|
atomic.AddInt32(&count, int32(item.(int)))
|
||||||
assert.Equal(t, int32(6), count)
|
}, UnlimitedWorkers())
|
||||||
|
assert.Equal(t, int32(6), count)
|
||||||
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestReverse(t *testing.T) {
|
func TestReverse(t *testing.T) {
|
||||||
Just(1, 2, 3, 4).Reverse().Merge().ForEach(func(item interface{}) {
|
runCheckedTest(t, func(t *testing.T) {
|
||||||
assert.ElementsMatch(t, []interface{}{4, 3, 2, 1}, item.([]interface{}))
|
Just(1, 2, 3, 4).Reverse().Merge().ForEach(func(item interface{}) {
|
||||||
|
assert.ElementsMatch(t, []interface{}{4, 3, 2, 1}, item.([]interface{}))
|
||||||
|
})
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestSort(t *testing.T) {
|
func TestSort(t *testing.T) {
|
||||||
var prev int
|
runCheckedTest(t, func(t *testing.T) {
|
||||||
Just(5, 3, 7, 1, 9, 6, 4, 8, 2).Sort(func(a, b interface{}) bool {
|
var prev int
|
||||||
return a.(int) < b.(int)
|
Just(5, 3, 7, 1, 9, 6, 4, 8, 2).Sort(func(a, b interface{}) bool {
|
||||||
}).ForEach(func(item interface{}) {
|
return a.(int) < b.(int)
|
||||||
next := item.(int)
|
}).ForEach(func(item interface{}) {
|
||||||
assert.True(t, prev < next)
|
next := item.(int)
|
||||||
prev = next
|
assert.True(t, prev < next)
|
||||||
|
prev = next
|
||||||
|
})
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestSplit(t *testing.T) {
|
func TestSplit(t *testing.T) {
|
||||||
assert.Panics(t, func() {
|
runCheckedTest(t, func(t *testing.T) {
|
||||||
Just(1, 2, 3, 4, 5, 6, 7, 8, 9, 10).Split(0).Done()
|
assert.Panics(t, func() {
|
||||||
|
Just(1, 2, 3, 4, 5, 6, 7, 8, 9, 10).Split(0).Done()
|
||||||
|
})
|
||||||
|
var chunks [][]interface{}
|
||||||
|
Just(1, 2, 3, 4, 5, 6, 7, 8, 9, 10).Split(4).ForEach(func(item interface{}) {
|
||||||
|
chunk := item.([]interface{})
|
||||||
|
chunks = append(chunks, chunk)
|
||||||
|
})
|
||||||
|
assert.EqualValues(t, [][]interface{}{
|
||||||
|
{1, 2, 3, 4},
|
||||||
|
{5, 6, 7, 8},
|
||||||
|
{9, 10},
|
||||||
|
}, chunks)
|
||||||
})
|
})
|
||||||
var chunks [][]interface{}
|
|
||||||
Just(1, 2, 3, 4, 5, 6, 7, 8, 9, 10).Split(4).ForEach(func(item interface{}) {
|
|
||||||
chunk := item.([]interface{})
|
|
||||||
chunks = append(chunks, chunk)
|
|
||||||
})
|
|
||||||
assert.EqualValues(t, [][]interface{}{
|
|
||||||
{1, 2, 3, 4},
|
|
||||||
{5, 6, 7, 8},
|
|
||||||
{9, 10},
|
|
||||||
}, chunks)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestTail(t *testing.T) {
|
func TestTail(t *testing.T) {
|
||||||
var result int
|
runCheckedTest(t, func(t *testing.T) {
|
||||||
Just(1, 2, 3, 4).Tail(2).Reduce(func(pipe <-chan interface{}) (interface{}, error) {
|
var result int
|
||||||
for item := range pipe {
|
Just(1, 2, 3, 4).Tail(2).Reduce(func(pipe <-chan interface{}) (interface{}, error) {
|
||||||
result += item.(int)
|
for item := range pipe {
|
||||||
}
|
result += item.(int)
|
||||||
return result, nil
|
}
|
||||||
|
return result, nil
|
||||||
|
})
|
||||||
|
assert.Equal(t, 7, result)
|
||||||
})
|
})
|
||||||
assert.Equal(t, 7, result)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestTailZero(t *testing.T) {
|
func TestTailZero(t *testing.T) {
|
||||||
assert.Panics(t, func() {
|
runCheckedTest(t, func(t *testing.T) {
|
||||||
Just(1, 2, 3, 4).Tail(0).Reduce(func(pipe <-chan interface{}) (interface{}, error) {
|
assert.Panics(t, func() {
|
||||||
return nil, nil
|
Just(1, 2, 3, 4).Tail(0).Reduce(func(pipe <-chan interface{}) (interface{}, error) {
|
||||||
|
return nil, nil
|
||||||
|
})
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestWalk(t *testing.T) {
|
func TestWalk(t *testing.T) {
|
||||||
var result int
|
runCheckedTest(t, func(t *testing.T) {
|
||||||
Just(1, 2, 3, 4, 5).Walk(func(item interface{}, pipe chan<- interface{}) {
|
var result int
|
||||||
if item.(int)%2 != 0 {
|
Just(1, 2, 3, 4, 5).Walk(func(item interface{}, pipe chan<- interface{}) {
|
||||||
pipe <- item
|
if item.(int)%2 != 0 {
|
||||||
}
|
pipe <- item
|
||||||
}, UnlimitedWorkers()).ForEach(func(item interface{}) {
|
}
|
||||||
result += item.(int)
|
}, UnlimitedWorkers()).ForEach(func(item interface{}) {
|
||||||
|
result += item.(int)
|
||||||
|
})
|
||||||
|
assert.Equal(t, 9, result)
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestStream_AnyMach(t *testing.T) {
|
||||||
|
runCheckedTest(t, func(t *testing.T) {
|
||||||
|
assetEqual(t, false, Just(1, 2, 3).AnyMach(func(item interface{}) bool {
|
||||||
|
return item.(int) == 4
|
||||||
|
}))
|
||||||
|
assetEqual(t, false, Just(1, 2, 3).AnyMach(func(item interface{}) bool {
|
||||||
|
return item.(int) == 0
|
||||||
|
}))
|
||||||
|
assetEqual(t, true, Just(1, 2, 3).AnyMach(func(item interface{}) bool {
|
||||||
|
return item.(int) == 2
|
||||||
|
}))
|
||||||
|
assetEqual(t, true, Just(1, 2, 3).AnyMach(func(item interface{}) bool {
|
||||||
|
return item.(int) == 2
|
||||||
|
}))
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestStream_AllMach(t *testing.T) {
|
||||||
|
runCheckedTest(t, func(t *testing.T) {
|
||||||
|
assetEqual(
|
||||||
|
t, true, Just(1, 2, 3).AllMach(func(item interface{}) bool {
|
||||||
|
return true
|
||||||
|
}),
|
||||||
|
)
|
||||||
|
assetEqual(
|
||||||
|
t, false, Just(1, 2, 3).AllMach(func(item interface{}) bool {
|
||||||
|
return false
|
||||||
|
}),
|
||||||
|
)
|
||||||
|
assetEqual(
|
||||||
|
t, false, Just(1, 2, 3).AllMach(func(item interface{}) bool {
|
||||||
|
return item.(int) == 1
|
||||||
|
}),
|
||||||
|
)
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestStream_NoneMatch(t *testing.T) {
|
||||||
|
runCheckedTest(t, func(t *testing.T) {
|
||||||
|
assetEqual(
|
||||||
|
t, true, Just(1, 2, 3).NoneMatch(func(item interface{}) bool {
|
||||||
|
return false
|
||||||
|
}),
|
||||||
|
)
|
||||||
|
assetEqual(
|
||||||
|
t, false, Just(1, 2, 3).NoneMatch(func(item interface{}) bool {
|
||||||
|
return true
|
||||||
|
}),
|
||||||
|
)
|
||||||
|
assetEqual(
|
||||||
|
t, true, Just(1, 2, 3).NoneMatch(func(item interface{}) bool {
|
||||||
|
return item.(int) == 4
|
||||||
|
}),
|
||||||
|
)
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestConcat(t *testing.T) {
|
||||||
|
runCheckedTest(t, func(t *testing.T) {
|
||||||
|
a1 := []interface{}{1, 2, 3}
|
||||||
|
a2 := []interface{}{4, 5, 6}
|
||||||
|
s1 := Just(a1...)
|
||||||
|
s2 := Just(a2...)
|
||||||
|
stream := Concat(s1, s2)
|
||||||
|
var items []interface{}
|
||||||
|
for item := range stream.source {
|
||||||
|
items = append(items, item)
|
||||||
|
}
|
||||||
|
sort.Slice(items, func(i, j int) bool {
|
||||||
|
return items[i].(int) < items[j].(int)
|
||||||
|
})
|
||||||
|
ints := make([]interface{}, 0)
|
||||||
|
ints = append(ints, a1...)
|
||||||
|
ints = append(ints, a2...)
|
||||||
|
assetEqual(t, ints, items)
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestStream_Skip(t *testing.T) {
|
||||||
|
runCheckedTest(t, func(t *testing.T) {
|
||||||
|
assetEqual(t, 3, Just(1, 2, 3, 4).Skip(1).Count())
|
||||||
|
assetEqual(t, 1, Just(1, 2, 3, 4).Skip(3).Count())
|
||||||
|
assetEqual(t, 4, Just(1, 2, 3, 4).Skip(0).Count())
|
||||||
|
equal(t, Just(1, 2, 3, 4).Skip(3), []interface{}{4})
|
||||||
|
assert.Panics(t, func() {
|
||||||
|
Just(1, 2, 3, 4).Skip(-1)
|
||||||
|
})
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestStream_Concat(t *testing.T) {
|
||||||
|
runCheckedTest(t, func(t *testing.T) {
|
||||||
|
stream := Just(1).Concat(Just(2), Just(3))
|
||||||
|
var items []interface{}
|
||||||
|
for item := range stream.source {
|
||||||
|
items = append(items, item)
|
||||||
|
}
|
||||||
|
sort.Slice(items, func(i, j int) bool {
|
||||||
|
return items[i].(int) < items[j].(int)
|
||||||
|
})
|
||||||
|
assetEqual(t, []interface{}{1, 2, 3}, items)
|
||||||
|
|
||||||
|
just := Just(1)
|
||||||
|
equal(t, just.Concat(just), []interface{}{1})
|
||||||
})
|
})
|
||||||
assert.Equal(t, 9, result)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func BenchmarkParallelMapReduce(b *testing.B) {
|
func BenchmarkParallelMapReduce(b *testing.B) {
|
||||||
@@ -377,6 +547,12 @@ func BenchmarkMapReduce(b *testing.B) {
|
|||||||
}).Map(mapper).Reduce(reducer)
|
}).Map(mapper).Reduce(reducer)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func assetEqual(t *testing.T, except, data interface{}) {
|
||||||
|
if !reflect.DeepEqual(except, data) {
|
||||||
|
t.Errorf(" %v, want %v", data, except)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
func equal(t *testing.T, stream Stream, data []interface{}) {
|
func equal(t *testing.T, stream Stream, data []interface{}) {
|
||||||
items := make([]interface{}, 0)
|
items := make([]interface{}, 0)
|
||||||
for item := range stream.source {
|
for item := range stream.source {
|
||||||
@@ -387,85 +563,7 @@ func equal(t *testing.T, stream Stream, data []interface{}) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func assetEqual(t *testing.T, except, data interface{}) {
|
func runCheckedTest(t *testing.T, fn func(t *testing.T)) {
|
||||||
if !reflect.DeepEqual(except, data) {
|
defer goleak.VerifyNone(t)
|
||||||
t.Errorf(" %v, want %v", data, except)
|
fn(t)
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestStream_AnyMach(t *testing.T) {
|
|
||||||
assetEqual(t, false, Just(1, 2, 3).AnyMach(func(item interface{}) bool {
|
|
||||||
return item.(int) == 4
|
|
||||||
}))
|
|
||||||
assetEqual(t, false, Just(1, 2, 3).AnyMach(func(item interface{}) bool {
|
|
||||||
return item.(int) == 0
|
|
||||||
}))
|
|
||||||
assetEqual(t, true, Just(1, 2, 3).AnyMach(func(item interface{}) bool {
|
|
||||||
return item.(int) == 2
|
|
||||||
}))
|
|
||||||
assetEqual(t, true, Just(1, 2, 3).AnyMach(func(item interface{}) bool {
|
|
||||||
return item.(int) == 2
|
|
||||||
}))
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestStream_AllMach(t *testing.T) {
|
|
||||||
assetEqual(
|
|
||||||
t, true, Just(1, 2, 3).AllMach(func(item interface{}) bool {
|
|
||||||
return true
|
|
||||||
}),
|
|
||||||
)
|
|
||||||
assetEqual(
|
|
||||||
t, false, Just(1, 2, 3).AllMach(func(item interface{}) bool {
|
|
||||||
return false
|
|
||||||
}),
|
|
||||||
)
|
|
||||||
assetEqual(
|
|
||||||
t, false, Just(1, 2, 3).AllMach(func(item interface{}) bool {
|
|
||||||
return item.(int) == 1
|
|
||||||
}),
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestConcat(t *testing.T) {
|
|
||||||
a1 := []interface{}{1, 2, 3}
|
|
||||||
a2 := []interface{}{4, 5, 6}
|
|
||||||
s1 := Just(a1...)
|
|
||||||
s2 := Just(a2...)
|
|
||||||
stream := Concat(s1, s2)
|
|
||||||
var items []interface{}
|
|
||||||
for item := range stream.source {
|
|
||||||
items = append(items, item)
|
|
||||||
}
|
|
||||||
sort.Slice(items, func(i, j int) bool {
|
|
||||||
return items[i].(int) < items[j].(int)
|
|
||||||
})
|
|
||||||
ints := make([]interface{}, 0)
|
|
||||||
ints = append(ints, a1...)
|
|
||||||
ints = append(ints, a2...)
|
|
||||||
assetEqual(t, ints, items)
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestStream_Skip(t *testing.T) {
|
|
||||||
assetEqual(t, 3, Just(1, 2, 3, 4).Skip(1).Count())
|
|
||||||
assetEqual(t, 1, Just(1, 2, 3, 4).Skip(3).Count())
|
|
||||||
assetEqual(t, 4, Just(1, 2, 3, 4).Skip(0).Count())
|
|
||||||
equal(t, Just(1, 2, 3, 4).Skip(3), []interface{}{4})
|
|
||||||
assert.Panics(t, func() {
|
|
||||||
Just(1, 2, 3, 4).Skip(-1)
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestStream_Concat(t *testing.T) {
|
|
||||||
stream := Just(1).Concat(Just(2), Just(3))
|
|
||||||
var items []interface{}
|
|
||||||
for item := range stream.source {
|
|
||||||
items = append(items, item)
|
|
||||||
}
|
|
||||||
sort.Slice(items, func(i, j int) bool {
|
|
||||||
return items[i].(int) < items[j].(int)
|
|
||||||
})
|
|
||||||
assetEqual(t, []interface{}{1, 2, 3}, items)
|
|
||||||
|
|
||||||
just := Just(1)
|
|
||||||
equal(t, just.Concat(just), []interface{}{1})
|
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -6,8 +6,8 @@ import (
|
|||||||
"strconv"
|
"strconv"
|
||||||
"sync"
|
"sync"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/lang"
|
"github.com/zeromicro/go-zero/core/lang"
|
||||||
"github.com/tal-tech/go-zero/core/mapping"
|
"github.com/zeromicro/go-zero/core/mapping"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
|
|||||||
@@ -6,7 +6,7 @@ import (
|
|||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/mathx"
|
"github.com/zeromicro/go-zero/core/mathx"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
|
|||||||
@@ -9,8 +9,8 @@ import (
|
|||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/fs"
|
"github.com/zeromicro/go-zero/core/fs"
|
||||||
"github.com/tal-tech/go-zero/core/stringx"
|
"github.com/zeromicro/go-zero/core/stringx"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestReadText(t *testing.T) {
|
func TestReadText(t *testing.T) {
|
||||||
|
|||||||
@@ -4,8 +4,8 @@ package lang
|
|||||||
var Placeholder PlaceholderType
|
var Placeholder PlaceholderType
|
||||||
|
|
||||||
type (
|
type (
|
||||||
// GenericType can be used to hold any type.
|
// AnyType can be used to hold any type.
|
||||||
GenericType = interface{}
|
AnyType = interface{}
|
||||||
// PlaceholderType represents a placeholder type.
|
// PlaceholderType represents a placeholder type.
|
||||||
PlaceholderType = struct{}
|
PlaceholderType = struct{}
|
||||||
)
|
)
|
||||||
|
|||||||
@@ -5,12 +5,11 @@ import (
|
|||||||
"strconv"
|
"strconv"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/stores/redis"
|
"github.com/zeromicro/go-zero/core/stores/redis"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
// to be compatible with aliyun redis, we cannot use `local key = KEYS[1]` to reuse the key
|
||||||
// to be compatible with aliyun redis, we cannot use `local key = KEYS[1]` to reuse the key
|
const periodScript = `local limit = tonumber(ARGV[1])
|
||||||
periodScript = `local limit = tonumber(ARGV[1])
|
|
||||||
local window = tonumber(ARGV[2])
|
local window = tonumber(ARGV[2])
|
||||||
local current = redis.call("INCRBY", KEYS[1], 1)
|
local current = redis.call("INCRBY", KEYS[1], 1)
|
||||||
if current == 1 then
|
if current == 1 then
|
||||||
@@ -23,8 +22,6 @@ elseif current == limit then
|
|||||||
else
|
else
|
||||||
return 0
|
return 0
|
||||||
end`
|
end`
|
||||||
zoneDiff = 3600 * 8 // GMT+8 for our services
|
|
||||||
)
|
|
||||||
|
|
||||||
const (
|
const (
|
||||||
// Unknown means not initialized state.
|
// Unknown means not initialized state.
|
||||||
@@ -104,7 +101,9 @@ func (h *PeriodLimit) Take(key string) (int, error) {
|
|||||||
|
|
||||||
func (h *PeriodLimit) calcExpireSeconds() int {
|
func (h *PeriodLimit) calcExpireSeconds() int {
|
||||||
if h.align {
|
if h.align {
|
||||||
unix := time.Now().Unix() + zoneDiff
|
now := time.Now()
|
||||||
|
_, offset := now.Zone()
|
||||||
|
unix := now.Unix() + int64(offset)
|
||||||
return h.period - int(unix%int64(h.period))
|
return h.period - int(unix%int64(h.period))
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -112,6 +111,8 @@ func (h *PeriodLimit) calcExpireSeconds() int {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Align returns a func to customize a PeriodLimit with alignment.
|
// Align returns a func to customize a PeriodLimit with alignment.
|
||||||
|
// For example, if we want to limit end users with 5 sms verification messages every day,
|
||||||
|
// we need to align with the local timezone and the start of the day.
|
||||||
func Align() PeriodOption {
|
func Align() PeriodOption {
|
||||||
return func(l *PeriodLimit) {
|
return func(l *PeriodLimit) {
|
||||||
l.align = true
|
l.align = true
|
||||||
|
|||||||
@@ -5,8 +5,8 @@ import (
|
|||||||
|
|
||||||
"github.com/alicebob/miniredis/v2"
|
"github.com/alicebob/miniredis/v2"
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/stores/redis"
|
"github.com/zeromicro/go-zero/core/stores/redis"
|
||||||
"github.com/tal-tech/go-zero/core/stores/redis/redistest"
|
"github.com/zeromicro/go-zero/core/stores/redis/redistest"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestPeriodLimit_Take(t *testing.T) {
|
func TestPeriodLimit_Take(t *testing.T) {
|
||||||
|
|||||||
@@ -7,8 +7,8 @@ import (
|
|||||||
"sync/atomic"
|
"sync/atomic"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/logx"
|
"github.com/zeromicro/go-zero/core/logx"
|
||||||
"github.com/tal-tech/go-zero/core/stores/redis"
|
"github.com/zeromicro/go-zero/core/stores/redis"
|
||||||
xrate "golang.org/x/time/rate"
|
xrate "golang.org/x/time/rate"
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -85,8 +85,8 @@ func (lim *TokenLimiter) Allow() bool {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// AllowN reports whether n events may happen at time now.
|
// AllowN reports whether n events may happen at time now.
|
||||||
// Use this method if you intend to drop / skip events that exceed the rate rate.
|
// Use this method if you intend to drop / skip events that exceed the rate.
|
||||||
// Otherwise use Reserve or Wait.
|
// Otherwise, use Reserve or Wait.
|
||||||
func (lim *TokenLimiter) AllowN(now time.Time, n int) bool {
|
func (lim *TokenLimiter) AllowN(now time.Time, n int) bool {
|
||||||
return lim.reserveN(now, n)
|
return lim.reserveN(now, n)
|
||||||
}
|
}
|
||||||
@@ -112,7 +112,8 @@ func (lim *TokenLimiter) reserveN(now time.Time, n int) bool {
|
|||||||
// Lua boolean false -> r Nil bulk reply
|
// Lua boolean false -> r Nil bulk reply
|
||||||
if err == redis.Nil {
|
if err == redis.Nil {
|
||||||
return false
|
return false
|
||||||
} else if err != nil {
|
}
|
||||||
|
if err != nil {
|
||||||
logx.Errorf("fail to use rate limiter: %s, use in-process limiter for rescue", err)
|
logx.Errorf("fail to use rate limiter: %s, use in-process limiter for rescue", err)
|
||||||
lim.startMonitor()
|
lim.startMonitor()
|
||||||
return lim.rescueLimiter.AllowN(now, n)
|
return lim.rescueLimiter.AllowN(now, n)
|
||||||
|
|||||||
@@ -6,9 +6,9 @@ import (
|
|||||||
|
|
||||||
"github.com/alicebob/miniredis/v2"
|
"github.com/alicebob/miniredis/v2"
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/logx"
|
"github.com/zeromicro/go-zero/core/logx"
|
||||||
"github.com/tal-tech/go-zero/core/stores/redis"
|
"github.com/zeromicro/go-zero/core/stores/redis"
|
||||||
"github.com/tal-tech/go-zero/core/stores/redis/redistest"
|
"github.com/zeromicro/go-zero/core/stores/redis/redistest"
|
||||||
)
|
)
|
||||||
|
|
||||||
func init() {
|
func init() {
|
||||||
|
|||||||
@@ -7,11 +7,11 @@ import (
|
|||||||
"sync/atomic"
|
"sync/atomic"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/collection"
|
"github.com/zeromicro/go-zero/core/collection"
|
||||||
"github.com/tal-tech/go-zero/core/logx"
|
"github.com/zeromicro/go-zero/core/logx"
|
||||||
"github.com/tal-tech/go-zero/core/stat"
|
"github.com/zeromicro/go-zero/core/stat"
|
||||||
"github.com/tal-tech/go-zero/core/syncx"
|
"github.com/zeromicro/go-zero/core/syncx"
|
||||||
"github.com/tal-tech/go-zero/core/timex"
|
"github.com/zeromicro/go-zero/core/timex"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
|
|||||||
@@ -8,11 +8,11 @@ import (
|
|||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/collection"
|
"github.com/zeromicro/go-zero/core/collection"
|
||||||
"github.com/tal-tech/go-zero/core/logx"
|
"github.com/zeromicro/go-zero/core/logx"
|
||||||
"github.com/tal-tech/go-zero/core/mathx"
|
"github.com/zeromicro/go-zero/core/mathx"
|
||||||
"github.com/tal-tech/go-zero/core/stat"
|
"github.com/zeromicro/go-zero/core/stat"
|
||||||
"github.com/tal-tech/go-zero/core/syncx"
|
"github.com/zeromicro/go-zero/core/syncx"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
|
|||||||
@@ -3,7 +3,7 @@ package load
|
|||||||
import (
|
import (
|
||||||
"io"
|
"io"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/syncx"
|
"github.com/zeromicro/go-zero/core/syncx"
|
||||||
)
|
)
|
||||||
|
|
||||||
// A ShedderGroup is a manager to manage key based shedders.
|
// A ShedderGroup is a manager to manage key based shedders.
|
||||||
|
|||||||
@@ -4,8 +4,8 @@ import (
|
|||||||
"sync/atomic"
|
"sync/atomic"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/logx"
|
"github.com/zeromicro/go-zero/core/logx"
|
||||||
"github.com/tal-tech/go-zero/core/stat"
|
"github.com/zeromicro/go-zero/core/stat"
|
||||||
)
|
)
|
||||||
|
|
||||||
type (
|
type (
|
||||||
|
|||||||
@@ -3,10 +3,11 @@ package logx
|
|||||||
// A LogConf is a logging config.
|
// A LogConf is a logging config.
|
||||||
type LogConf struct {
|
type LogConf struct {
|
||||||
ServiceName string `json:",optional"`
|
ServiceName string `json:",optional"`
|
||||||
Mode string `json:",default=console,options=console|file|volume"`
|
Mode string `json:",default=console,options=[console,file,volume]"`
|
||||||
|
Encoding string `json:",default=json,options=[json,plain]"`
|
||||||
TimeFormat string `json:",optional"`
|
TimeFormat string `json:",optional"`
|
||||||
Path string `json:",default=logs"`
|
Path string `json:",default=logs"`
|
||||||
Level string `json:",default=info,options=info|error|severe"`
|
Level string `json:",default=info,options=[info,error,severe]"`
|
||||||
Compress bool `json:",optional"`
|
Compress bool `json:",optional"`
|
||||||
KeepDays int `json:",optional"`
|
KeepDays int `json:",optional"`
|
||||||
StackCooldownMillis int `json:",default=100"`
|
StackCooldownMillis int `json:",default=100"`
|
||||||
|
|||||||
@@ -5,7 +5,7 @@ import (
|
|||||||
"io"
|
"io"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/timex"
|
"github.com/zeromicro/go-zero/core/timex"
|
||||||
)
|
)
|
||||||
|
|
||||||
const durationCallerDepth = 3
|
const durationCallerDepth = 3
|
||||||
@@ -79,8 +79,15 @@ func (l *durationLogger) WithDuration(duration time.Duration) Logger {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (l *durationLogger) write(writer io.Writer, level string, val interface{}) {
|
func (l *durationLogger) write(writer io.Writer, level string, val interface{}) {
|
||||||
l.Timestamp = getTimestamp()
|
switch encoding {
|
||||||
l.Level = level
|
case plainEncodingType:
|
||||||
l.Content = val
|
writePlainAny(writer, level, val, l.Duration)
|
||||||
outputJson(writer, l)
|
default:
|
||||||
|
outputJson(writer, &durationLogger{
|
||||||
|
Timestamp: getTimestamp(),
|
||||||
|
Level: level,
|
||||||
|
Content: val,
|
||||||
|
Duration: l.Duration,
|
||||||
|
})
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -23,6 +23,13 @@ func TestWithDurationErrorf(t *testing.T) {
|
|||||||
assert.True(t, strings.Contains(builder.String(), "duration"), builder.String())
|
assert.True(t, strings.Contains(builder.String(), "duration"), builder.String())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func TestWithDurationErrorv(t *testing.T) {
|
||||||
|
var builder strings.Builder
|
||||||
|
log.SetOutput(&builder)
|
||||||
|
WithDuration(time.Second).Errorv("foo")
|
||||||
|
assert.True(t, strings.Contains(builder.String(), "duration"), builder.String())
|
||||||
|
}
|
||||||
|
|
||||||
func TestWithDurationInfo(t *testing.T) {
|
func TestWithDurationInfo(t *testing.T) {
|
||||||
var builder strings.Builder
|
var builder strings.Builder
|
||||||
log.SetOutput(&builder)
|
log.SetOutput(&builder)
|
||||||
@@ -30,6 +37,19 @@ func TestWithDurationInfo(t *testing.T) {
|
|||||||
assert.True(t, strings.Contains(builder.String(), "duration"), builder.String())
|
assert.True(t, strings.Contains(builder.String(), "duration"), builder.String())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func TestWithDurationInfoConsole(t *testing.T) {
|
||||||
|
old := encoding
|
||||||
|
encoding = plainEncodingType
|
||||||
|
defer func() {
|
||||||
|
encoding = old
|
||||||
|
}()
|
||||||
|
|
||||||
|
var builder strings.Builder
|
||||||
|
log.SetOutput(&builder)
|
||||||
|
WithDuration(time.Second).Info("foo")
|
||||||
|
assert.True(t, strings.Contains(builder.String(), "ms"), builder.String())
|
||||||
|
}
|
||||||
|
|
||||||
func TestWithDurationInfof(t *testing.T) {
|
func TestWithDurationInfof(t *testing.T) {
|
||||||
var builder strings.Builder
|
var builder strings.Builder
|
||||||
log.SetOutput(&builder)
|
log.SetOutput(&builder)
|
||||||
@@ -37,6 +57,13 @@ func TestWithDurationInfof(t *testing.T) {
|
|||||||
assert.True(t, strings.Contains(builder.String(), "duration"), builder.String())
|
assert.True(t, strings.Contains(builder.String(), "duration"), builder.String())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func TestWithDurationInfov(t *testing.T) {
|
||||||
|
var builder strings.Builder
|
||||||
|
log.SetOutput(&builder)
|
||||||
|
WithDuration(time.Second).Infov("foo")
|
||||||
|
assert.True(t, strings.Contains(builder.String(), "duration"), builder.String())
|
||||||
|
}
|
||||||
|
|
||||||
func TestWithDurationSlow(t *testing.T) {
|
func TestWithDurationSlow(t *testing.T) {
|
||||||
var builder strings.Builder
|
var builder strings.Builder
|
||||||
log.SetOutput(&builder)
|
log.SetOutput(&builder)
|
||||||
@@ -50,3 +77,10 @@ func TestWithDurationSlowf(t *testing.T) {
|
|||||||
WithDuration(time.Second).WithDuration(time.Hour).Slowf("foo")
|
WithDuration(time.Second).WithDuration(time.Hour).Slowf("foo")
|
||||||
assert.True(t, strings.Contains(builder.String(), "duration"), builder.String())
|
assert.True(t, strings.Contains(builder.String(), "duration"), builder.String())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func TestWithDurationSlowv(t *testing.T) {
|
||||||
|
var builder strings.Builder
|
||||||
|
log.SetOutput(&builder)
|
||||||
|
WithDuration(time.Second).WithDuration(time.Hour).Slowv("foo")
|
||||||
|
assert.True(t, strings.Contains(builder.String(), "duration"), builder.String())
|
||||||
|
}
|
||||||
|
|||||||
@@ -4,8 +4,8 @@ import (
|
|||||||
"sync/atomic"
|
"sync/atomic"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/syncx"
|
"github.com/zeromicro/go-zero/core/syncx"
|
||||||
"github.com/tal-tech/go-zero/core/timex"
|
"github.com/zeromicro/go-zero/core/timex"
|
||||||
)
|
)
|
||||||
|
|
||||||
type limitedExecutor struct {
|
type limitedExecutor struct {
|
||||||
|
|||||||
@@ -6,7 +6,7 @@ import (
|
|||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/timex"
|
"github.com/zeromicro/go-zero/core/timex"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestLimitedExecutor_logOrDiscard(t *testing.T) {
|
func TestLimitedExecutor_logOrDiscard(t *testing.T) {
|
||||||
|
|||||||
@@ -1,6 +1,7 @@
|
|||||||
package logx
|
package logx
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"bytes"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
@@ -17,9 +18,9 @@ import (
|
|||||||
"sync/atomic"
|
"sync/atomic"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/iox"
|
"github.com/zeromicro/go-zero/core/iox"
|
||||||
"github.com/tal-tech/go-zero/core/sysx"
|
"github.com/zeromicro/go-zero/core/sysx"
|
||||||
"github.com/tal-tech/go-zero/core/timex"
|
"github.com/zeromicro/go-zero/core/timex"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
@@ -31,6 +32,15 @@ const (
|
|||||||
SevereLevel
|
SevereLevel
|
||||||
)
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
jsonEncodingType = iota
|
||||||
|
plainEncodingType
|
||||||
|
|
||||||
|
jsonEncoding = "json"
|
||||||
|
plainEncoding = "plain"
|
||||||
|
plainEncodingSep = '\t'
|
||||||
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
accessFilename = "access.log"
|
accessFilename = "access.log"
|
||||||
errorFilename = "error.log"
|
errorFilename = "error.log"
|
||||||
@@ -65,6 +75,7 @@ var (
|
|||||||
timeFormat = "2006-01-02T15:04:05.000Z07"
|
timeFormat = "2006-01-02T15:04:05.000Z07"
|
||||||
writeConsole bool
|
writeConsole bool
|
||||||
logLevel uint32
|
logLevel uint32
|
||||||
|
encoding = jsonEncodingType
|
||||||
// use uint32 for atomic operations
|
// use uint32 for atomic operations
|
||||||
disableStat uint32
|
disableStat uint32
|
||||||
infoLog io.WriteCloser
|
infoLog io.WriteCloser
|
||||||
@@ -124,6 +135,12 @@ func SetUp(c LogConf) error {
|
|||||||
if len(c.TimeFormat) > 0 {
|
if len(c.TimeFormat) > 0 {
|
||||||
timeFormat = c.TimeFormat
|
timeFormat = c.TimeFormat
|
||||||
}
|
}
|
||||||
|
switch c.Encoding {
|
||||||
|
case plainEncoding:
|
||||||
|
encoding = plainEncodingType
|
||||||
|
default:
|
||||||
|
encoding = jsonEncodingType
|
||||||
|
}
|
||||||
|
|
||||||
switch c.Mode {
|
switch c.Mode {
|
||||||
case consoleMode:
|
case consoleMode:
|
||||||
@@ -217,7 +234,7 @@ func ErrorCaller(callDepth int, v ...interface{}) {
|
|||||||
|
|
||||||
// ErrorCallerf writes v with context in format into error log.
|
// ErrorCallerf writes v with context in format into error log.
|
||||||
func ErrorCallerf(callDepth int, format string, v ...interface{}) {
|
func ErrorCallerf(callDepth int, format string, v ...interface{}) {
|
||||||
errorTextSync(fmt.Sprintf(format, v...), callDepth+callerInnerDepth)
|
errorTextSync(fmt.Errorf(format, v...).Error(), callDepth+callerInnerDepth)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Errorf writes v with format into error log.
|
// Errorf writes v with format into error log.
|
||||||
@@ -407,21 +424,31 @@ func infoTextSync(msg string) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func outputAny(writer io.Writer, level string, val interface{}) {
|
func outputAny(writer io.Writer, level string, val interface{}) {
|
||||||
info := logEntry{
|
switch encoding {
|
||||||
Timestamp: getTimestamp(),
|
case plainEncodingType:
|
||||||
Level: level,
|
writePlainAny(writer, level, val)
|
||||||
Content: val,
|
default:
|
||||||
|
info := logEntry{
|
||||||
|
Timestamp: getTimestamp(),
|
||||||
|
Level: level,
|
||||||
|
Content: val,
|
||||||
|
}
|
||||||
|
outputJson(writer, info)
|
||||||
}
|
}
|
||||||
outputJson(writer, info)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func outputText(writer io.Writer, level, msg string) {
|
func outputText(writer io.Writer, level, msg string) {
|
||||||
info := logEntry{
|
switch encoding {
|
||||||
Timestamp: getTimestamp(),
|
case plainEncodingType:
|
||||||
Level: level,
|
writePlainText(writer, level, msg)
|
||||||
Content: msg,
|
default:
|
||||||
|
info := logEntry{
|
||||||
|
Timestamp: getTimestamp(),
|
||||||
|
Level: level,
|
||||||
|
Content: msg,
|
||||||
|
}
|
||||||
|
outputJson(writer, info)
|
||||||
}
|
}
|
||||||
outputJson(writer, info)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func outputError(writer io.Writer, msg string, callDepth int) {
|
func outputError(writer io.Writer, msg string, callDepth int) {
|
||||||
@@ -565,6 +592,62 @@ func statSync(msg string) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func writePlainAny(writer io.Writer, level string, val interface{}, fields ...string) {
|
||||||
|
switch v := val.(type) {
|
||||||
|
case string:
|
||||||
|
writePlainText(writer, level, v, fields...)
|
||||||
|
case error:
|
||||||
|
writePlainText(writer, level, v.Error(), fields...)
|
||||||
|
case fmt.Stringer:
|
||||||
|
writePlainText(writer, level, v.String(), fields...)
|
||||||
|
default:
|
||||||
|
var buf bytes.Buffer
|
||||||
|
buf.WriteString(getTimestamp())
|
||||||
|
buf.WriteByte(plainEncodingSep)
|
||||||
|
buf.WriteString(level)
|
||||||
|
for _, item := range fields {
|
||||||
|
buf.WriteByte(plainEncodingSep)
|
||||||
|
buf.WriteString(item)
|
||||||
|
}
|
||||||
|
buf.WriteByte(plainEncodingSep)
|
||||||
|
if err := json.NewEncoder(&buf).Encode(val); err != nil {
|
||||||
|
log.Println(err.Error())
|
||||||
|
return
|
||||||
|
}
|
||||||
|
buf.WriteByte('\n')
|
||||||
|
if atomic.LoadUint32(&initialized) == 0 || writer == nil {
|
||||||
|
log.Println(buf.String())
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
if _, err := writer.Write(buf.Bytes()); err != nil {
|
||||||
|
log.Println(err.Error())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func writePlainText(writer io.Writer, level, msg string, fields ...string) {
|
||||||
|
var buf bytes.Buffer
|
||||||
|
buf.WriteString(getTimestamp())
|
||||||
|
buf.WriteByte(plainEncodingSep)
|
||||||
|
buf.WriteString(level)
|
||||||
|
for _, item := range fields {
|
||||||
|
buf.WriteByte(plainEncodingSep)
|
||||||
|
buf.WriteString(item)
|
||||||
|
}
|
||||||
|
buf.WriteByte(plainEncodingSep)
|
||||||
|
buf.WriteString(msg)
|
||||||
|
buf.WriteByte('\n')
|
||||||
|
if atomic.LoadUint32(&initialized) == 0 || writer == nil {
|
||||||
|
log.Println(buf.String())
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
if _, err := writer.Write(buf.Bytes()); err != nil {
|
||||||
|
log.Println(err.Error())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
type logWriter struct {
|
type logWriter struct {
|
||||||
logger *log.Logger
|
logger *log.Logger
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -2,6 +2,7 @@ package logx
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"io"
|
"io"
|
||||||
"io/ioutil"
|
"io/ioutil"
|
||||||
@@ -140,6 +141,78 @@ func TestStructedLogInfov(t *testing.T) {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func TestStructedLogInfoConsoleAny(t *testing.T) {
|
||||||
|
doTestStructedLogConsole(t, func(writer io.WriteCloser) {
|
||||||
|
infoLog = writer
|
||||||
|
}, func(v ...interface{}) {
|
||||||
|
old := encoding
|
||||||
|
encoding = plainEncodingType
|
||||||
|
defer func() {
|
||||||
|
encoding = old
|
||||||
|
}()
|
||||||
|
|
||||||
|
Infov(v)
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestStructedLogInfoConsoleAnyString(t *testing.T) {
|
||||||
|
doTestStructedLogConsole(t, func(writer io.WriteCloser) {
|
||||||
|
infoLog = writer
|
||||||
|
}, func(v ...interface{}) {
|
||||||
|
old := encoding
|
||||||
|
encoding = plainEncodingType
|
||||||
|
defer func() {
|
||||||
|
encoding = old
|
||||||
|
}()
|
||||||
|
|
||||||
|
Infov(fmt.Sprint(v...))
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestStructedLogInfoConsoleAnyError(t *testing.T) {
|
||||||
|
doTestStructedLogConsole(t, func(writer io.WriteCloser) {
|
||||||
|
infoLog = writer
|
||||||
|
}, func(v ...interface{}) {
|
||||||
|
old := encoding
|
||||||
|
encoding = plainEncodingType
|
||||||
|
defer func() {
|
||||||
|
encoding = old
|
||||||
|
}()
|
||||||
|
|
||||||
|
Infov(errors.New(fmt.Sprint(v...)))
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestStructedLogInfoConsoleAnyStringer(t *testing.T) {
|
||||||
|
doTestStructedLogConsole(t, func(writer io.WriteCloser) {
|
||||||
|
infoLog = writer
|
||||||
|
}, func(v ...interface{}) {
|
||||||
|
old := encoding
|
||||||
|
encoding = plainEncodingType
|
||||||
|
defer func() {
|
||||||
|
encoding = old
|
||||||
|
}()
|
||||||
|
|
||||||
|
Infov(ValStringer{
|
||||||
|
val: fmt.Sprint(v...),
|
||||||
|
})
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestStructedLogInfoConsoleText(t *testing.T) {
|
||||||
|
doTestStructedLogConsole(t, func(writer io.WriteCloser) {
|
||||||
|
infoLog = writer
|
||||||
|
}, func(v ...interface{}) {
|
||||||
|
old := encoding
|
||||||
|
encoding = plainEncodingType
|
||||||
|
defer func() {
|
||||||
|
encoding = old
|
||||||
|
}()
|
||||||
|
|
||||||
|
Info(fmt.Sprint(v...))
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
func TestStructedLogSlow(t *testing.T) {
|
func TestStructedLogSlow(t *testing.T) {
|
||||||
doTestStructedLog(t, levelSlow, func(writer io.WriteCloser) {
|
doTestStructedLog(t, levelSlow, func(writer io.WriteCloser) {
|
||||||
slowLog = writer
|
slowLog = writer
|
||||||
@@ -242,6 +315,16 @@ func TestSetLevelWithDuration(t *testing.T) {
|
|||||||
assert.Equal(t, 0, writer.builder.Len())
|
assert.Equal(t, 0, writer.builder.Len())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func TestErrorfWithWrappedError(t *testing.T) {
|
||||||
|
SetLevel(ErrorLevel)
|
||||||
|
const message = "there"
|
||||||
|
writer := new(mockWriter)
|
||||||
|
errorLog = writer
|
||||||
|
atomic.StoreUint32(&initialized, 1)
|
||||||
|
Errorf("hello %w", errors.New(message))
|
||||||
|
assert.True(t, strings.Contains(writer.builder.String(), "hello there"))
|
||||||
|
}
|
||||||
|
|
||||||
func TestMustNil(t *testing.T) {
|
func TestMustNil(t *testing.T) {
|
||||||
Must(nil)
|
Must(nil)
|
||||||
}
|
}
|
||||||
@@ -421,6 +504,17 @@ func doTestStructedLog(t *testing.T, level string, setup func(writer io.WriteClo
|
|||||||
assert.True(t, strings.Contains(val, message))
|
assert.True(t, strings.Contains(val, message))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func doTestStructedLogConsole(t *testing.T, setup func(writer io.WriteCloser),
|
||||||
|
write func(...interface{})) {
|
||||||
|
const message = "hello there"
|
||||||
|
writer := new(mockWriter)
|
||||||
|
setup(writer)
|
||||||
|
atomic.StoreUint32(&initialized, 1)
|
||||||
|
write(message)
|
||||||
|
println(writer.String())
|
||||||
|
assert.True(t, strings.Contains(writer.String(), message))
|
||||||
|
}
|
||||||
|
|
||||||
func testSetLevelTwiceWithMode(t *testing.T, mode string) {
|
func testSetLevelTwiceWithMode(t *testing.T, mode string) {
|
||||||
SetUp(LogConf{
|
SetUp(LogConf{
|
||||||
Mode: mode,
|
Mode: mode,
|
||||||
@@ -445,3 +539,11 @@ func testSetLevelTwiceWithMode(t *testing.T, mode string) {
|
|||||||
ErrorStackf(message)
|
ErrorStackf(message)
|
||||||
assert.Equal(t, 0, writer.builder.Len())
|
assert.Equal(t, 0, writer.builder.Len())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
type ValStringer struct {
|
||||||
|
val string
|
||||||
|
}
|
||||||
|
|
||||||
|
func (v ValStringer) String() string {
|
||||||
|
return v.val
|
||||||
|
}
|
||||||
|
|||||||
@@ -13,9 +13,9 @@ import (
|
|||||||
"sync"
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/fs"
|
"github.com/zeromicro/go-zero/core/fs"
|
||||||
"github.com/tal-tech/go-zero/core/lang"
|
"github.com/zeromicro/go-zero/core/lang"
|
||||||
"github.com/tal-tech/go-zero/core/timex"
|
"github.com/zeromicro/go-zero/core/timex"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
|
|||||||
@@ -8,7 +8,7 @@ import (
|
|||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/fs"
|
"github.com/zeromicro/go-zero/core/fs"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestDailyRotateRuleMarkRotated(t *testing.T) {
|
func TestDailyRotateRuleMarkRotated(t *testing.T) {
|
||||||
|
|||||||
@@ -6,7 +6,7 @@ import (
|
|||||||
"io"
|
"io"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/timex"
|
"github.com/zeromicro/go-zero/core/timex"
|
||||||
"go.opentelemetry.io/otel/trace"
|
"go.opentelemetry.io/otel/trace"
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -77,12 +77,24 @@ func (l *traceLogger) WithDuration(duration time.Duration) Logger {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (l *traceLogger) write(writer io.Writer, level string, val interface{}) {
|
func (l *traceLogger) write(writer io.Writer, level string, val interface{}) {
|
||||||
l.Timestamp = getTimestamp()
|
traceID := traceIdFromContext(l.ctx)
|
||||||
l.Level = level
|
spanID := spanIdFromContext(l.ctx)
|
||||||
l.Content = val
|
|
||||||
l.Trace = traceIdFromContext(l.ctx)
|
switch encoding {
|
||||||
l.Span = spanIdFromContext(l.ctx)
|
case plainEncodingType:
|
||||||
outputJson(writer, l)
|
writePlainAny(writer, level, val, l.Duration, traceID, spanID)
|
||||||
|
default:
|
||||||
|
outputJson(writer, &traceLogger{
|
||||||
|
logEntry: logEntry{
|
||||||
|
Timestamp: getTimestamp(),
|
||||||
|
Level: level,
|
||||||
|
Duration: l.Duration,
|
||||||
|
Content: val,
|
||||||
|
},
|
||||||
|
Trace: traceID,
|
||||||
|
Span: spanID,
|
||||||
|
})
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// WithContext sets ctx to log, for keeping tracing information.
|
// WithContext sets ctx to log, for keeping tracing information.
|
||||||
|
|||||||
@@ -51,6 +51,10 @@ func TestTraceError(t *testing.T) {
|
|||||||
l.WithDuration(time.Second).Errorf(testlog)
|
l.WithDuration(time.Second).Errorf(testlog)
|
||||||
assert.True(t, strings.Contains(buf.String(), traceKey))
|
assert.True(t, strings.Contains(buf.String(), traceKey))
|
||||||
assert.True(t, strings.Contains(buf.String(), spanKey))
|
assert.True(t, strings.Contains(buf.String(), spanKey))
|
||||||
|
buf.Reset()
|
||||||
|
l.WithDuration(time.Second).Errorv(testlog)
|
||||||
|
assert.True(t, strings.Contains(buf.String(), traceKey))
|
||||||
|
assert.True(t, strings.Contains(buf.String(), spanKey))
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestTraceInfo(t *testing.T) {
|
func TestTraceInfo(t *testing.T) {
|
||||||
@@ -72,6 +76,41 @@ func TestTraceInfo(t *testing.T) {
|
|||||||
l.WithDuration(time.Second).Infof(testlog)
|
l.WithDuration(time.Second).Infof(testlog)
|
||||||
assert.True(t, strings.Contains(buf.String(), traceKey))
|
assert.True(t, strings.Contains(buf.String(), traceKey))
|
||||||
assert.True(t, strings.Contains(buf.String(), spanKey))
|
assert.True(t, strings.Contains(buf.String(), spanKey))
|
||||||
|
buf.Reset()
|
||||||
|
l.WithDuration(time.Second).Infov(testlog)
|
||||||
|
assert.True(t, strings.Contains(buf.String(), traceKey))
|
||||||
|
assert.True(t, strings.Contains(buf.String(), spanKey))
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestTraceInfoConsole(t *testing.T) {
|
||||||
|
old := encoding
|
||||||
|
encoding = plainEncodingType
|
||||||
|
defer func() {
|
||||||
|
encoding = old
|
||||||
|
}()
|
||||||
|
|
||||||
|
var buf mockWriter
|
||||||
|
atomic.StoreUint32(&initialized, 1)
|
||||||
|
infoLog = newLogWriter(log.New(&buf, "", flags))
|
||||||
|
otp := otel.GetTracerProvider()
|
||||||
|
tp := sdktrace.NewTracerProvider(sdktrace.WithSampler(sdktrace.AlwaysSample()))
|
||||||
|
otel.SetTracerProvider(tp)
|
||||||
|
defer otel.SetTracerProvider(otp)
|
||||||
|
|
||||||
|
ctx, _ := tp.Tracer("foo").Start(context.Background(), "bar")
|
||||||
|
l := WithContext(ctx).(*traceLogger)
|
||||||
|
SetLevel(InfoLevel)
|
||||||
|
l.WithDuration(time.Second).Info(testlog)
|
||||||
|
assert.True(t, strings.Contains(buf.String(), traceIdFromContext(ctx)))
|
||||||
|
assert.True(t, strings.Contains(buf.String(), spanIdFromContext(ctx)))
|
||||||
|
buf.Reset()
|
||||||
|
l.WithDuration(time.Second).Infof(testlog)
|
||||||
|
assert.True(t, strings.Contains(buf.String(), traceIdFromContext(ctx)))
|
||||||
|
assert.True(t, strings.Contains(buf.String(), spanIdFromContext(ctx)))
|
||||||
|
buf.Reset()
|
||||||
|
l.WithDuration(time.Second).Infov(testlog)
|
||||||
|
assert.True(t, strings.Contains(buf.String(), traceIdFromContext(ctx)))
|
||||||
|
assert.True(t, strings.Contains(buf.String(), spanIdFromContext(ctx)))
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestTraceSlow(t *testing.T) {
|
func TestTraceSlow(t *testing.T) {
|
||||||
@@ -93,6 +132,10 @@ func TestTraceSlow(t *testing.T) {
|
|||||||
l.WithDuration(time.Second).Slowf(testlog)
|
l.WithDuration(time.Second).Slowf(testlog)
|
||||||
assert.True(t, strings.Contains(buf.String(), traceKey))
|
assert.True(t, strings.Contains(buf.String(), traceKey))
|
||||||
assert.True(t, strings.Contains(buf.String(), spanKey))
|
assert.True(t, strings.Contains(buf.String(), spanKey))
|
||||||
|
buf.Reset()
|
||||||
|
l.WithDuration(time.Second).Slowv(testlog)
|
||||||
|
assert.True(t, strings.Contains(buf.String(), traceKey))
|
||||||
|
assert.True(t, strings.Contains(buf.String(), spanKey))
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestTraceWithoutContext(t *testing.T) {
|
func TestTraceWithoutContext(t *testing.T) {
|
||||||
|
|||||||
@@ -3,7 +3,7 @@ package mapping
|
|||||||
import (
|
import (
|
||||||
"io"
|
"io"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/jsonx"
|
"github.com/zeromicro/go-zero/core/jsonx"
|
||||||
)
|
)
|
||||||
|
|
||||||
const jsonTagKey = "json"
|
const jsonTagKey = "json"
|
||||||
@@ -15,6 +15,11 @@ func UnmarshalJsonBytes(content []byte, v interface{}) error {
|
|||||||
return unmarshalJsonBytes(content, v, jsonUnmarshaler)
|
return unmarshalJsonBytes(content, v, jsonUnmarshaler)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// UnmarshalJsonMap unmarshals content from m into v.
|
||||||
|
func UnmarshalJsonMap(m map[string]interface{}, v interface{}) error {
|
||||||
|
return jsonUnmarshaler.Unmarshal(m, v)
|
||||||
|
}
|
||||||
|
|
||||||
// UnmarshalJsonReader unmarshals content from reader into v.
|
// UnmarshalJsonReader unmarshals content from reader into v.
|
||||||
func UnmarshalJsonReader(reader io.Reader, v interface{}) error {
|
func UnmarshalJsonReader(reader io.Reader, v interface{}) error {
|
||||||
return unmarshalJsonReader(reader, v, jsonUnmarshaler)
|
return unmarshalJsonReader(reader, v, jsonUnmarshaler)
|
||||||
|
|||||||
@@ -871,3 +871,50 @@ func TestUnmarshalReaderError(t *testing.T) {
|
|||||||
assert.NotNil(t, err)
|
assert.NotNil(t, err)
|
||||||
assert.True(t, strings.Contains(err.Error(), payload))
|
assert.True(t, strings.Contains(err.Error(), payload))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func TestUnmarshalMap(t *testing.T) {
|
||||||
|
t.Run("nil map and valid", func(t *testing.T) {
|
||||||
|
var m map[string]interface{}
|
||||||
|
var v struct {
|
||||||
|
Any string `json:",optional"`
|
||||||
|
}
|
||||||
|
|
||||||
|
err := UnmarshalJsonMap(m, &v)
|
||||||
|
assert.Nil(t, err)
|
||||||
|
assert.True(t, len(v.Any) == 0)
|
||||||
|
})
|
||||||
|
|
||||||
|
t.Run("empty map but not valid", func(t *testing.T) {
|
||||||
|
m := map[string]interface{}{}
|
||||||
|
var v struct {
|
||||||
|
Any string
|
||||||
|
}
|
||||||
|
|
||||||
|
err := UnmarshalJsonMap(m, &v)
|
||||||
|
assert.NotNil(t, err)
|
||||||
|
})
|
||||||
|
|
||||||
|
t.Run("empty map and valid", func(t *testing.T) {
|
||||||
|
m := map[string]interface{}{}
|
||||||
|
var v struct {
|
||||||
|
Any string `json:",optional"`
|
||||||
|
}
|
||||||
|
|
||||||
|
err := UnmarshalJsonMap(m, &v)
|
||||||
|
assert.Nil(t, err)
|
||||||
|
assert.True(t, len(v.Any) == 0)
|
||||||
|
})
|
||||||
|
|
||||||
|
t.Run("valid map", func(t *testing.T) {
|
||||||
|
m := map[string]interface{}{
|
||||||
|
"Any": "foo",
|
||||||
|
}
|
||||||
|
var v struct {
|
||||||
|
Any string
|
||||||
|
}
|
||||||
|
|
||||||
|
err := UnmarshalJsonMap(m, &v)
|
||||||
|
assert.Nil(t, err)
|
||||||
|
assert.Equal(t, "foo", v.Any)
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|||||||
@@ -7,12 +7,11 @@ import (
|
|||||||
"reflect"
|
"reflect"
|
||||||
"strings"
|
"strings"
|
||||||
"sync"
|
"sync"
|
||||||
"sync/atomic"
|
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/jsonx"
|
"github.com/zeromicro/go-zero/core/jsonx"
|
||||||
"github.com/tal-tech/go-zero/core/lang"
|
"github.com/zeromicro/go-zero/core/lang"
|
||||||
"github.com/tal-tech/go-zero/core/stringx"
|
"github.com/zeromicro/go-zero/core/stringx"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
@@ -25,15 +24,17 @@ var (
|
|||||||
errValueNotSettable = errors.New("value is not settable")
|
errValueNotSettable = errors.New("value is not settable")
|
||||||
errValueNotStruct = errors.New("value type is not struct")
|
errValueNotStruct = errors.New("value type is not struct")
|
||||||
keyUnmarshaler = NewUnmarshaler(defaultKeyName)
|
keyUnmarshaler = NewUnmarshaler(defaultKeyName)
|
||||||
cacheKeys atomic.Value
|
|
||||||
cacheKeysLock sync.Mutex
|
|
||||||
durationType = reflect.TypeOf(time.Duration(0))
|
durationType = reflect.TypeOf(time.Duration(0))
|
||||||
|
cacheKeys map[string][]string
|
||||||
|
cacheKeysLock sync.Mutex
|
||||||
|
defaultCache map[string]interface{}
|
||||||
|
defaultCacheLock sync.Mutex
|
||||||
emptyMap = map[string]interface{}{}
|
emptyMap = map[string]interface{}{}
|
||||||
emptyValue = reflect.ValueOf(lang.Placeholder)
|
emptyValue = reflect.ValueOf(lang.Placeholder)
|
||||||
)
|
)
|
||||||
|
|
||||||
type (
|
type (
|
||||||
// A Unmarshaler is used to unmarshal with given tag key.
|
// Unmarshaler is used to unmarshal with given tag key.
|
||||||
Unmarshaler struct {
|
Unmarshaler struct {
|
||||||
key string
|
key string
|
||||||
opts unmarshalOptions
|
opts unmarshalOptions
|
||||||
@@ -46,12 +47,11 @@ type (
|
|||||||
fromString bool
|
fromString bool
|
||||||
canonicalKey func(key string) string
|
canonicalKey func(key string) string
|
||||||
}
|
}
|
||||||
|
|
||||||
keyCache map[string][]string
|
|
||||||
)
|
)
|
||||||
|
|
||||||
func init() {
|
func init() {
|
||||||
cacheKeys.Store(make(keyCache))
|
cacheKeys = make(map[string][]string)
|
||||||
|
defaultCache = make(map[string]interface{})
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewUnmarshaler returns a Unmarshaler.
|
// NewUnmarshaler returns a Unmarshaler.
|
||||||
@@ -207,6 +207,8 @@ func (u *Unmarshaler) processFieldNotFromString(field reflect.StructField, value
|
|||||||
switch {
|
switch {
|
||||||
case valueKind == reflect.Map && typeKind == reflect.Struct:
|
case valueKind == reflect.Map && typeKind == reflect.Struct:
|
||||||
return u.processFieldStruct(field, value, mapValue, fullName)
|
return u.processFieldStruct(field, value, mapValue, fullName)
|
||||||
|
case valueKind == reflect.Map && typeKind == reflect.Map:
|
||||||
|
return u.fillMap(field, value, mapValue)
|
||||||
case valueKind == reflect.String && typeKind == reflect.Slice:
|
case valueKind == reflect.String && typeKind == reflect.Slice:
|
||||||
return u.fillSliceFromString(fieldType, value, mapValue)
|
return u.fillSliceFromString(fieldType, value, mapValue)
|
||||||
case valueKind == reflect.String && derefedFieldType == durationType:
|
case valueKind == reflect.String && derefedFieldType == durationType:
|
||||||
@@ -386,7 +388,13 @@ func (u *Unmarshaler) processNamedFieldWithoutValue(field reflect.StructField, v
|
|||||||
if derefedType == durationType {
|
if derefedType == durationType {
|
||||||
return fillDurationValue(fieldKind, value, defaultValue)
|
return fillDurationValue(fieldKind, value, defaultValue)
|
||||||
}
|
}
|
||||||
return setValue(fieldKind, value, defaultValue)
|
|
||||||
|
switch fieldKind {
|
||||||
|
case reflect.Array, reflect.Slice:
|
||||||
|
return u.fillSliceWithDefault(derefedType, value, defaultValue)
|
||||||
|
default:
|
||||||
|
return setValue(fieldKind, value, defaultValue)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
switch fieldKind {
|
switch fieldKind {
|
||||||
@@ -500,7 +508,8 @@ func (u *Unmarshaler) fillSliceFromString(fieldType reflect.Type, value reflect.
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (u *Unmarshaler) fillSliceValue(slice reflect.Value, index int, baseKind reflect.Kind, value interface{}) error {
|
func (u *Unmarshaler) fillSliceValue(slice reflect.Value, index int,
|
||||||
|
baseKind reflect.Kind, value interface{}) error {
|
||||||
ithVal := slice.Index(index)
|
ithVal := slice.Index(index)
|
||||||
switch v := value.(type) {
|
switch v := value.(type) {
|
||||||
case json.Number:
|
case json.Number:
|
||||||
@@ -529,6 +538,28 @@ func (u *Unmarshaler) fillSliceValue(slice reflect.Value, index int, baseKind re
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (u *Unmarshaler) fillSliceWithDefault(derefedType reflect.Type, value reflect.Value,
|
||||||
|
defaultValue string) error {
|
||||||
|
baseFieldType := Deref(derefedType.Elem())
|
||||||
|
baseFieldKind := baseFieldType.Kind()
|
||||||
|
defaultCacheLock.Lock()
|
||||||
|
slice, ok := defaultCache[defaultValue]
|
||||||
|
defaultCacheLock.Unlock()
|
||||||
|
if !ok {
|
||||||
|
if baseFieldKind == reflect.String {
|
||||||
|
slice = parseGroupedSegments(defaultValue)
|
||||||
|
} else if err := jsonx.UnmarshalFromString(defaultValue, &slice); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
defaultCacheLock.Lock()
|
||||||
|
defaultCache[defaultValue] = slice
|
||||||
|
defaultCacheLock.Unlock()
|
||||||
|
}
|
||||||
|
|
||||||
|
return u.fillSlice(derefedType, value, slice)
|
||||||
|
}
|
||||||
|
|
||||||
func (u *Unmarshaler) generateMap(keyType, elemType reflect.Type, mapValue interface{}) (reflect.Value, error) {
|
func (u *Unmarshaler) generateMap(keyType, elemType reflect.Type, mapValue interface{}) (reflect.Value, error) {
|
||||||
mapType := reflect.MapOf(keyType, elemType)
|
mapType := reflect.MapOf(keyType, elemType)
|
||||||
valueType := reflect.TypeOf(mapValue)
|
valueType := reflect.TypeOf(mapValue)
|
||||||
@@ -584,6 +615,8 @@ func (u *Unmarshaler) generateMap(keyType, elemType reflect.Type, mapValue inter
|
|||||||
targetValue.SetMapIndex(key, innerValue)
|
targetValue.SetMapIndex(key, innerValue)
|
||||||
default:
|
default:
|
||||||
switch v := keythData.(type) {
|
switch v := keythData.(type) {
|
||||||
|
case bool:
|
||||||
|
targetValue.SetMapIndex(key, reflect.ValueOf(v))
|
||||||
case string:
|
case string:
|
||||||
targetValue.SetMapIndex(key, reflect.ValueOf(v))
|
targetValue.SetMapIndex(key, reflect.ValueOf(v))
|
||||||
case json.Number:
|
case json.Number:
|
||||||
@@ -709,7 +742,9 @@ func getValueWithChainedKeys(m Valuer, keys []string) (interface{}, bool) {
|
|||||||
if len(keys) == 1 {
|
if len(keys) == 1 {
|
||||||
v, ok := m.Value(keys[0])
|
v, ok := m.Value(keys[0])
|
||||||
return v, ok
|
return v, ok
|
||||||
} else if len(keys) > 1 {
|
}
|
||||||
|
|
||||||
|
if len(keys) > 1 {
|
||||||
if v, ok := m.Value(keys[0]); ok {
|
if v, ok := m.Value(keys[0]); ok {
|
||||||
if nextm, ok := v.(map[string]interface{}); ok {
|
if nextm, ok := v.(map[string]interface{}); ok {
|
||||||
return getValueWithChainedKeys(MapValuer(nextm), keys[1:])
|
return getValueWithChainedKeys(MapValuer(nextm), keys[1:])
|
||||||
@@ -720,20 +755,6 @@ func getValueWithChainedKeys(m Valuer, keys []string) (interface{}, bool) {
|
|||||||
return nil, false
|
return nil, false
|
||||||
}
|
}
|
||||||
|
|
||||||
func insertKeys(key string, cache []string) {
|
|
||||||
cacheKeysLock.Lock()
|
|
||||||
defer cacheKeysLock.Unlock()
|
|
||||||
|
|
||||||
keys := cacheKeys.Load().(keyCache)
|
|
||||||
// copy the contents into the new map, to guarantee the old map is immutable
|
|
||||||
newKeys := make(keyCache)
|
|
||||||
for k, v := range keys {
|
|
||||||
newKeys[k] = v
|
|
||||||
}
|
|
||||||
newKeys[key] = cache
|
|
||||||
cacheKeys.Store(newKeys)
|
|
||||||
}
|
|
||||||
|
|
||||||
func join(elem ...string) string {
|
func join(elem ...string) string {
|
||||||
var builder strings.Builder
|
var builder strings.Builder
|
||||||
|
|
||||||
@@ -764,15 +785,19 @@ func newTypeMismatchError(name string) error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func readKeys(key string) []string {
|
func readKeys(key string) []string {
|
||||||
cache := cacheKeys.Load().(keyCache)
|
cacheKeysLock.Lock()
|
||||||
if keys, ok := cache[key]; ok {
|
keys, ok := cacheKeys[key]
|
||||||
|
cacheKeysLock.Unlock()
|
||||||
|
if ok {
|
||||||
return keys
|
return keys
|
||||||
}
|
}
|
||||||
|
|
||||||
keys := strings.FieldsFunc(key, func(c rune) bool {
|
keys = strings.FieldsFunc(key, func(c rune) bool {
|
||||||
return c == delimiter
|
return c == delimiter
|
||||||
})
|
})
|
||||||
insertKeys(key, keys)
|
cacheKeysLock.Lock()
|
||||||
|
cacheKeys[key] = keys
|
||||||
|
cacheKeysLock.Unlock()
|
||||||
|
|
||||||
return keys
|
return keys
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -8,7 +8,7 @@ import (
|
|||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/stringx"
|
"github.com/zeromicro/go-zero/core/stringx"
|
||||||
)
|
)
|
||||||
|
|
||||||
// because json.Number doesn't support strconv.ParseUint(...),
|
// because json.Number doesn't support strconv.ParseUint(...),
|
||||||
@@ -198,6 +198,66 @@ func TestUnmarshalIntWithDefault(t *testing.T) {
|
|||||||
assert.Equal(t, 1, in.Int)
|
assert.Equal(t, 1, in.Int)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func TestUnmarshalBoolSliceWithDefault(t *testing.T) {
|
||||||
|
type inner struct {
|
||||||
|
Bools []bool `key:"bools,default=[true,false]"`
|
||||||
|
}
|
||||||
|
|
||||||
|
var in inner
|
||||||
|
assert.Nil(t, UnmarshalKey(nil, &in))
|
||||||
|
assert.ElementsMatch(t, []bool{true, false}, in.Bools)
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestUnmarshalIntSliceWithDefault(t *testing.T) {
|
||||||
|
type inner struct {
|
||||||
|
Ints []int `key:"ints,default=[1,2,3]"`
|
||||||
|
}
|
||||||
|
|
||||||
|
var in inner
|
||||||
|
assert.Nil(t, UnmarshalKey(nil, &in))
|
||||||
|
assert.ElementsMatch(t, []int{1, 2, 3}, in.Ints)
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestUnmarshalIntSliceWithDefaultHasSpaces(t *testing.T) {
|
||||||
|
type inner struct {
|
||||||
|
Ints []int `key:"ints,default=[1, 2, 3]"`
|
||||||
|
}
|
||||||
|
|
||||||
|
var in inner
|
||||||
|
assert.Nil(t, UnmarshalKey(nil, &in))
|
||||||
|
assert.ElementsMatch(t, []int{1, 2, 3}, in.Ints)
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestUnmarshalFloatSliceWithDefault(t *testing.T) {
|
||||||
|
type inner struct {
|
||||||
|
Floats []float32 `key:"floats,default=[1.1,2.2,3.3]"`
|
||||||
|
}
|
||||||
|
|
||||||
|
var in inner
|
||||||
|
assert.Nil(t, UnmarshalKey(nil, &in))
|
||||||
|
assert.ElementsMatch(t, []float32{1.1, 2.2, 3.3}, in.Floats)
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestUnmarshalStringSliceWithDefault(t *testing.T) {
|
||||||
|
type inner struct {
|
||||||
|
Strs []string `key:"strs,default=[foo,bar,woo]"`
|
||||||
|
}
|
||||||
|
|
||||||
|
var in inner
|
||||||
|
assert.Nil(t, UnmarshalKey(nil, &in))
|
||||||
|
assert.ElementsMatch(t, []string{"foo", "bar", "woo"}, in.Strs)
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestUnmarshalStringSliceWithDefaultHasSpaces(t *testing.T) {
|
||||||
|
type inner struct {
|
||||||
|
Strs []string `key:"strs,default=[foo, bar, woo]"`
|
||||||
|
}
|
||||||
|
|
||||||
|
var in inner
|
||||||
|
assert.Nil(t, UnmarshalKey(nil, &in))
|
||||||
|
assert.ElementsMatch(t, []string{"foo", "bar", "woo"}, in.Strs)
|
||||||
|
}
|
||||||
|
|
||||||
func TestUnmarshalUint(t *testing.T) {
|
func TestUnmarshalUint(t *testing.T) {
|
||||||
type inner struct {
|
type inner struct {
|
||||||
Uint uint `key:"uint"`
|
Uint uint `key:"uint"`
|
||||||
@@ -861,10 +921,12 @@ func TestUnmarshalSliceOfStruct(t *testing.T) {
|
|||||||
func TestUnmarshalWithStringOptionsCorrect(t *testing.T) {
|
func TestUnmarshalWithStringOptionsCorrect(t *testing.T) {
|
||||||
type inner struct {
|
type inner struct {
|
||||||
Value string `key:"value,options=first|second"`
|
Value string `key:"value,options=first|second"`
|
||||||
|
Foo string `key:"foo,options=[bar,baz]"`
|
||||||
Correct string `key:"correct,options=1|2"`
|
Correct string `key:"correct,options=1|2"`
|
||||||
}
|
}
|
||||||
m := map[string]interface{}{
|
m := map[string]interface{}{
|
||||||
"value": "first",
|
"value": "first",
|
||||||
|
"foo": "bar",
|
||||||
"correct": "2",
|
"correct": "2",
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -872,6 +934,7 @@ func TestUnmarshalWithStringOptionsCorrect(t *testing.T) {
|
|||||||
ast := assert.New(t)
|
ast := assert.New(t)
|
||||||
ast.Nil(UnmarshalKey(m, &in))
|
ast.Nil(UnmarshalKey(m, &in))
|
||||||
ast.Equal("first", in.Value)
|
ast.Equal("first", in.Value)
|
||||||
|
ast.Equal("bar", in.Foo)
|
||||||
ast.Equal("2", in.Correct)
|
ast.Equal("2", in.Correct)
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -943,6 +1006,22 @@ func TestUnmarshalStringOptionsWithStringOptionsIncorrect(t *testing.T) {
|
|||||||
ast.NotNil(unmarshaler.Unmarshal(m, &in))
|
ast.NotNil(unmarshaler.Unmarshal(m, &in))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func TestUnmarshalStringOptionsWithStringOptionsIncorrectGrouped(t *testing.T) {
|
||||||
|
type inner struct {
|
||||||
|
Value string `key:"value,options=[first,second]"`
|
||||||
|
Correct string `key:"correct,options=1|2"`
|
||||||
|
}
|
||||||
|
m := map[string]interface{}{
|
||||||
|
"value": "third",
|
||||||
|
"correct": "2",
|
||||||
|
}
|
||||||
|
|
||||||
|
var in inner
|
||||||
|
unmarshaler := NewUnmarshaler(defaultKeyName, WithStringValues())
|
||||||
|
ast := assert.New(t)
|
||||||
|
ast.NotNil(unmarshaler.Unmarshal(m, &in))
|
||||||
|
}
|
||||||
|
|
||||||
func TestUnmarshalWithStringOptionsIncorrect(t *testing.T) {
|
func TestUnmarshalWithStringOptionsIncorrect(t *testing.T) {
|
||||||
type inner struct {
|
type inner struct {
|
||||||
Value string `key:"value,options=first|second"`
|
Value string `key:"value,options=first|second"`
|
||||||
@@ -2518,3 +2597,29 @@ func TestUnmarshalJsonReaderPtrArray(t *testing.T) {
|
|||||||
assert.Nil(t, err)
|
assert.Nil(t, err)
|
||||||
assert.Equal(t, 3, len(res.B))
|
assert.Equal(t, 3, len(res.B))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func TestUnmarshalJsonWithoutKey(t *testing.T) {
|
||||||
|
payload := `{"A": "1", "B": "2"}`
|
||||||
|
var res struct {
|
||||||
|
A string `json:""`
|
||||||
|
B string `json:","`
|
||||||
|
}
|
||||||
|
reader := strings.NewReader(payload)
|
||||||
|
err := UnmarshalJsonReader(reader, &res)
|
||||||
|
assert.Nil(t, err)
|
||||||
|
assert.Equal(t, "1", res.A)
|
||||||
|
assert.Equal(t, "2", res.B)
|
||||||
|
}
|
||||||
|
|
||||||
|
func BenchmarkDefaultValue(b *testing.B) {
|
||||||
|
for i := 0; i < b.N; i++ {
|
||||||
|
var a struct {
|
||||||
|
Ints []int `json:"ints,default=[1,2,3]"`
|
||||||
|
Strs []string `json:"strs,default=[foo,bar,baz]"`
|
||||||
|
}
|
||||||
|
_ = UnmarshalJsonMap(nil, &a)
|
||||||
|
if len(a.Strs) != 3 || len(a.Ints) != 3 {
|
||||||
|
b.Fatal("failed")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|||||||
@@ -10,17 +10,23 @@ import (
|
|||||||
"strings"
|
"strings"
|
||||||
"sync"
|
"sync"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/stringx"
|
"github.com/zeromicro/go-zero/core/stringx"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
defaultOption = "default"
|
defaultOption = "default"
|
||||||
stringOption = "string"
|
stringOption = "string"
|
||||||
optionalOption = "optional"
|
optionalOption = "optional"
|
||||||
optionsOption = "options"
|
optionsOption = "options"
|
||||||
rangeOption = "range"
|
rangeOption = "range"
|
||||||
optionSeparator = "|"
|
optionSeparator = "|"
|
||||||
equalToken = "="
|
equalToken = "="
|
||||||
|
escapeChar = '\\'
|
||||||
|
leftBracket = '('
|
||||||
|
rightBracket = ')'
|
||||||
|
leftSquareBracket = '['
|
||||||
|
rightSquareBracket = ']'
|
||||||
|
segmentSeparator = ','
|
||||||
)
|
)
|
||||||
|
|
||||||
var (
|
var (
|
||||||
@@ -118,7 +124,7 @@ func convertType(kind reflect.Kind, str string) (interface{}, error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func doParseKeyAndOptions(field reflect.StructField, value string) (string, *fieldOptions, error) {
|
func doParseKeyAndOptions(field reflect.StructField, value string) (string, *fieldOptions, error) {
|
||||||
segments := strings.Split(value, ",")
|
segments := parseSegments(value)
|
||||||
key := strings.TrimSpace(segments[0])
|
key := strings.TrimSpace(segments[0])
|
||||||
options := segments[1:]
|
options := segments[1:]
|
||||||
|
|
||||||
@@ -198,6 +204,16 @@ func maybeNewValue(field reflect.StructField, value reflect.Value) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func parseGroupedSegments(val string) []string {
|
||||||
|
val = strings.TrimLeftFunc(val, func(r rune) bool {
|
||||||
|
return r == leftBracket || r == leftSquareBracket
|
||||||
|
})
|
||||||
|
val = strings.TrimRightFunc(val, func(r rune) bool {
|
||||||
|
return r == rightBracket || r == rightSquareBracket
|
||||||
|
})
|
||||||
|
return parseSegments(val)
|
||||||
|
}
|
||||||
|
|
||||||
// don't modify returned fieldOptions, it's cached and shared among different calls.
|
// don't modify returned fieldOptions, it's cached and shared among different calls.
|
||||||
func parseKeyAndOptions(tagName string, field reflect.StructField) (string, *fieldOptions, error) {
|
func parseKeyAndOptions(tagName string, field reflect.StructField) (string, *fieldOptions, error) {
|
||||||
value := field.Tag.Get(tagName)
|
value := field.Tag.Get(tagName)
|
||||||
@@ -309,7 +325,7 @@ func parseOption(fieldOpts *fieldOptions, fieldName, option string) error {
|
|||||||
return fmt.Errorf("field %s has wrong options", fieldName)
|
return fmt.Errorf("field %s has wrong options", fieldName)
|
||||||
}
|
}
|
||||||
|
|
||||||
fieldOpts.Options = strings.Split(segs[1], optionSeparator)
|
fieldOpts.Options = parseOptions(segs[1])
|
||||||
case strings.HasPrefix(option, defaultOption):
|
case strings.HasPrefix(option, defaultOption):
|
||||||
segs := strings.Split(option, equalToken)
|
segs := strings.Split(option, equalToken)
|
||||||
if len(segs) != 2 {
|
if len(segs) != 2 {
|
||||||
@@ -334,6 +350,69 @@ func parseOption(fieldOpts *fieldOptions, fieldName, option string) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// parseOptions parses the given options in tag.
|
||||||
|
// for example: `json:"name,options=foo|bar"` or `json:"name,options=[foo,bar]"`
|
||||||
|
func parseOptions(val string) []string {
|
||||||
|
if len(val) == 0 {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
if val[0] == leftSquareBracket {
|
||||||
|
return parseGroupedSegments(val)
|
||||||
|
}
|
||||||
|
|
||||||
|
return strings.Split(val, optionSeparator)
|
||||||
|
}
|
||||||
|
|
||||||
|
func parseSegments(val string) []string {
|
||||||
|
var segments []string
|
||||||
|
var escaped, grouped bool
|
||||||
|
var buf strings.Builder
|
||||||
|
|
||||||
|
for _, ch := range val {
|
||||||
|
if escaped {
|
||||||
|
buf.WriteRune(ch)
|
||||||
|
escaped = false
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
switch ch {
|
||||||
|
case segmentSeparator:
|
||||||
|
if grouped {
|
||||||
|
buf.WriteRune(ch)
|
||||||
|
} else {
|
||||||
|
// need to trim spaces, but we cannot ignore empty string,
|
||||||
|
// because the first segment stands for the key might be empty.
|
||||||
|
// if ignored, the later tag will be used as the key.
|
||||||
|
segments = append(segments, strings.TrimSpace(buf.String()))
|
||||||
|
buf.Reset()
|
||||||
|
}
|
||||||
|
case escapeChar:
|
||||||
|
if grouped {
|
||||||
|
buf.WriteRune(ch)
|
||||||
|
} else {
|
||||||
|
escaped = true
|
||||||
|
}
|
||||||
|
case leftBracket, leftSquareBracket:
|
||||||
|
buf.WriteRune(ch)
|
||||||
|
grouped = true
|
||||||
|
case rightBracket, rightSquareBracket:
|
||||||
|
buf.WriteRune(ch)
|
||||||
|
grouped = false
|
||||||
|
default:
|
||||||
|
buf.WriteRune(ch)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
last := strings.TrimSpace(buf.String())
|
||||||
|
// ignore last empty string
|
||||||
|
if len(last) > 0 {
|
||||||
|
segments = append(segments, last)
|
||||||
|
}
|
||||||
|
|
||||||
|
return segments
|
||||||
|
}
|
||||||
|
|
||||||
func reprOfValue(val reflect.Value) string {
|
func reprOfValue(val reflect.Value) string {
|
||||||
switch vt := val.Interface().(type) {
|
switch vt := val.Interface().(type) {
|
||||||
case bool:
|
case bool:
|
||||||
|
|||||||
@@ -90,6 +90,82 @@ func TestParseKeyAndOptionWithTagAndOption(t *testing.T) {
|
|||||||
assert.True(t, options.FromString)
|
assert.True(t, options.FromString)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func TestParseSegments(t *testing.T) {
|
||||||
|
tests := []struct {
|
||||||
|
input string
|
||||||
|
expect []string
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
input: "",
|
||||||
|
expect: []string{},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
input: ",",
|
||||||
|
expect: []string{""},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
input: "foo,",
|
||||||
|
expect: []string{"foo"},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
input: ",foo",
|
||||||
|
// the first empty string cannot be ignored, it's the key.
|
||||||
|
expect: []string{"", "foo"},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
input: "foo",
|
||||||
|
expect: []string{"foo"},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
input: "foo,bar",
|
||||||
|
expect: []string{"foo", "bar"},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
input: "foo,bar,baz",
|
||||||
|
expect: []string{"foo", "bar", "baz"},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
input: "foo,options=a|b",
|
||||||
|
expect: []string{"foo", "options=a|b"},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
input: "foo,bar,default=[baz,qux]",
|
||||||
|
expect: []string{"foo", "bar", "default=[baz,qux]"},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
input: "foo,bar,options=[baz,qux]",
|
||||||
|
expect: []string{"foo", "bar", "options=[baz,qux]"},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
input: `foo\,bar,options=[baz,qux]`,
|
||||||
|
expect: []string{`foo,bar`, "options=[baz,qux]"},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
input: `foo,bar,options=\[baz,qux]`,
|
||||||
|
expect: []string{"foo", "bar", "options=[baz", "qux]"},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
input: `foo,bar,options=[baz\,qux]`,
|
||||||
|
expect: []string{"foo", "bar", `options=[baz\,qux]`},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
input: `foo\,bar,options=[baz,qux],default=baz`,
|
||||||
|
expect: []string{`foo,bar`, "options=[baz,qux]", "default=baz"},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
input: `foo\,bar,options=[baz,qux, quux],default=[qux, baz]`,
|
||||||
|
expect: []string{`foo,bar`, "options=[baz,qux, quux]", "default=[qux, baz]"},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, test := range tests {
|
||||||
|
test := test
|
||||||
|
t.Run(test.input, func(t *testing.T) {
|
||||||
|
assert.ElementsMatch(t, test.expect, parseSegments(test.input))
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
func TestValidatePtrWithNonPtr(t *testing.T) {
|
func TestValidatePtrWithNonPtr(t *testing.T) {
|
||||||
var foo string
|
var foo string
|
||||||
rve := reflect.ValueOf(foo)
|
rve := reflect.ValueOf(foo)
|
||||||
|
|||||||
@@ -945,6 +945,70 @@ func TestUnmarshalYamlBadReader(t *testing.T) {
|
|||||||
assert.NotNil(t, err)
|
assert.NotNil(t, err)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func TestUnmarshalYamlMapBool(t *testing.T) {
|
||||||
|
text := `machine:
|
||||||
|
node1: true
|
||||||
|
node2: true
|
||||||
|
node3: true
|
||||||
|
`
|
||||||
|
var v struct {
|
||||||
|
Machine map[string]bool `json:"machine,optional"`
|
||||||
|
}
|
||||||
|
reader := strings.NewReader(text)
|
||||||
|
assert.Nil(t, UnmarshalYamlReader(reader, &v))
|
||||||
|
assert.True(t, v.Machine["node1"])
|
||||||
|
assert.True(t, v.Machine["node2"])
|
||||||
|
assert.True(t, v.Machine["node3"])
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestUnmarshalYamlMapInt(t *testing.T) {
|
||||||
|
text := `machine:
|
||||||
|
node1: 1
|
||||||
|
node2: 2
|
||||||
|
node3: 3
|
||||||
|
`
|
||||||
|
var v struct {
|
||||||
|
Machine map[string]int `json:"machine,optional"`
|
||||||
|
}
|
||||||
|
reader := strings.NewReader(text)
|
||||||
|
assert.Nil(t, UnmarshalYamlReader(reader, &v))
|
||||||
|
assert.Equal(t, 1, v.Machine["node1"])
|
||||||
|
assert.Equal(t, 2, v.Machine["node2"])
|
||||||
|
assert.Equal(t, 3, v.Machine["node3"])
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestUnmarshalYamlMapByte(t *testing.T) {
|
||||||
|
text := `machine:
|
||||||
|
node1: 1
|
||||||
|
node2: 2
|
||||||
|
node3: 3
|
||||||
|
`
|
||||||
|
var v struct {
|
||||||
|
Machine map[string]byte `json:"machine,optional"`
|
||||||
|
}
|
||||||
|
reader := strings.NewReader(text)
|
||||||
|
assert.Nil(t, UnmarshalYamlReader(reader, &v))
|
||||||
|
assert.Equal(t, byte(1), v.Machine["node1"])
|
||||||
|
assert.Equal(t, byte(2), v.Machine["node2"])
|
||||||
|
assert.Equal(t, byte(3), v.Machine["node3"])
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestUnmarshalYamlMapRune(t *testing.T) {
|
||||||
|
text := `machine:
|
||||||
|
node1: 1
|
||||||
|
node2: 2
|
||||||
|
node3: 3
|
||||||
|
`
|
||||||
|
var v struct {
|
||||||
|
Machine map[string]rune `json:"machine,optional"`
|
||||||
|
}
|
||||||
|
reader := strings.NewReader(text)
|
||||||
|
assert.Nil(t, UnmarshalYamlReader(reader, &v))
|
||||||
|
assert.Equal(t, rune(1), v.Machine["node1"])
|
||||||
|
assert.Equal(t, rune(2), v.Machine["node2"])
|
||||||
|
assert.Equal(t, rune(3), v.Machine["node3"])
|
||||||
|
}
|
||||||
|
|
||||||
type badReader struct{}
|
type badReader struct{}
|
||||||
|
|
||||||
func (b *badReader) Read(p []byte) (n int, err error) {
|
func (b *badReader) Read(p []byte) (n int, err error) {
|
||||||
|
|||||||
@@ -4,7 +4,7 @@ import (
|
|||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/stringx"
|
"github.com/zeromicro/go-zero/core/stringx"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestMaxInt(t *testing.T) {
|
func TestMaxInt(t *testing.T) {
|
||||||
|
|||||||
@@ -2,7 +2,7 @@ package metric
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
prom "github.com/prometheus/client_golang/prometheus"
|
prom "github.com/prometheus/client_golang/prometheus"
|
||||||
"github.com/tal-tech/go-zero/core/proc"
|
"github.com/zeromicro/go-zero/core/proc"
|
||||||
)
|
)
|
||||||
|
|
||||||
type (
|
type (
|
||||||
|
|||||||
@@ -2,7 +2,7 @@ package metric
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
prom "github.com/prometheus/client_golang/prometheus"
|
prom "github.com/prometheus/client_golang/prometheus"
|
||||||
"github.com/tal-tech/go-zero/core/proc"
|
"github.com/zeromicro/go-zero/core/proc"
|
||||||
)
|
)
|
||||||
|
|
||||||
type (
|
type (
|
||||||
|
|||||||
@@ -2,7 +2,7 @@ package metric
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
prom "github.com/prometheus/client_golang/prometheus"
|
prom "github.com/prometheus/client_golang/prometheus"
|
||||||
"github.com/tal-tech/go-zero/core/proc"
|
"github.com/zeromicro/go-zero/core/proc"
|
||||||
)
|
)
|
||||||
|
|
||||||
type (
|
type (
|
||||||
|
|||||||
@@ -1,14 +1,14 @@
|
|||||||
package mr
|
package mr
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"context"
|
||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"sync"
|
"sync"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/errorx"
|
"github.com/zeromicro/go-zero/core/errorx"
|
||||||
"github.com/tal-tech/go-zero/core/lang"
|
"github.com/zeromicro/go-zero/core/lang"
|
||||||
"github.com/tal-tech/go-zero/core/syncx"
|
"github.com/zeromicro/go-zero/core/threading"
|
||||||
"github.com/tal-tech/go-zero/core/threading"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
@@ -24,12 +24,12 @@ var (
|
|||||||
)
|
)
|
||||||
|
|
||||||
type (
|
type (
|
||||||
|
// ForEachFunc is used to do element processing, but no output.
|
||||||
|
ForEachFunc func(item interface{})
|
||||||
// GenerateFunc is used to let callers send elements into source.
|
// GenerateFunc is used to let callers send elements into source.
|
||||||
GenerateFunc func(source chan<- interface{})
|
GenerateFunc func(source chan<- interface{})
|
||||||
// MapFunc is used to do element processing and write the output to writer.
|
// MapFunc is used to do element processing and write the output to writer.
|
||||||
MapFunc func(item interface{}, writer Writer)
|
MapFunc func(item interface{}, writer Writer)
|
||||||
// VoidMapFunc is used to do element processing, but no output.
|
|
||||||
VoidMapFunc func(item interface{})
|
|
||||||
// MapperFunc is used to do element processing and write the output to writer,
|
// MapperFunc is used to do element processing and write the output to writer,
|
||||||
// use cancel func to cancel the processing.
|
// use cancel func to cancel the processing.
|
||||||
MapperFunc func(item interface{}, writer Writer, cancel func(error))
|
MapperFunc func(item interface{}, writer Writer, cancel func(error))
|
||||||
@@ -43,6 +43,7 @@ type (
|
|||||||
Option func(opts *mapReduceOptions)
|
Option func(opts *mapReduceOptions)
|
||||||
|
|
||||||
mapReduceOptions struct {
|
mapReduceOptions struct {
|
||||||
|
ctx context.Context
|
||||||
workers int
|
workers int
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -68,7 +69,6 @@ func Finish(fns ...func() error) error {
|
|||||||
cancel(err)
|
cancel(err)
|
||||||
}
|
}
|
||||||
}, func(pipe <-chan interface{}, cancel func(error)) {
|
}, func(pipe <-chan interface{}, cancel func(error)) {
|
||||||
drain(pipe)
|
|
||||||
}, WithWorkers(len(fns)))
|
}, WithWorkers(len(fns)))
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -78,7 +78,7 @@ func FinishVoid(fns ...func()) {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
MapVoid(func(source chan<- interface{}) {
|
ForEach(func(source chan<- interface{}) {
|
||||||
for _, fn := range fns {
|
for _, fn := range fns {
|
||||||
source <- fn
|
source <- fn
|
||||||
}
|
}
|
||||||
@@ -88,27 +88,35 @@ func FinishVoid(fns ...func()) {
|
|||||||
}, WithWorkers(len(fns)))
|
}, WithWorkers(len(fns)))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// ForEach maps all elements from given generate but no output.
|
||||||
|
func ForEach(generate GenerateFunc, mapper ForEachFunc, opts ...Option) {
|
||||||
|
drain(Map(generate, func(item interface{}, writer Writer) {
|
||||||
|
mapper(item)
|
||||||
|
}, opts...))
|
||||||
|
}
|
||||||
|
|
||||||
// Map maps all elements generated from given generate func, and returns an output channel.
|
// Map maps all elements generated from given generate func, and returns an output channel.
|
||||||
func Map(generate GenerateFunc, mapper MapFunc, opts ...Option) chan interface{} {
|
func Map(generate GenerateFunc, mapper MapFunc, opts ...Option) chan interface{} {
|
||||||
options := buildOptions(opts...)
|
options := buildOptions(opts...)
|
||||||
source := buildSource(generate)
|
source := buildSource(generate)
|
||||||
collector := make(chan interface{}, options.workers)
|
collector := make(chan interface{}, options.workers)
|
||||||
done := syncx.NewDoneChan()
|
done := make(chan lang.PlaceholderType)
|
||||||
|
|
||||||
go executeMappers(mapper, source, collector, done.Done(), options.workers)
|
go executeMappers(options.ctx, mapper, source, collector, done, options.workers)
|
||||||
|
|
||||||
return collector
|
return collector
|
||||||
}
|
}
|
||||||
|
|
||||||
// MapReduce maps all elements generated from given generate func,
|
// MapReduce maps all elements generated from given generate func,
|
||||||
// and reduces the output elements with given reducer.
|
// and reduces the output elements with given reducer.
|
||||||
func MapReduce(generate GenerateFunc, mapper MapperFunc, reducer ReducerFunc, opts ...Option) (interface{}, error) {
|
func MapReduce(generate GenerateFunc, mapper MapperFunc, reducer ReducerFunc,
|
||||||
|
opts ...Option) (interface{}, error) {
|
||||||
source := buildSource(generate)
|
source := buildSource(generate)
|
||||||
return MapReduceWithSource(source, mapper, reducer, opts...)
|
return MapReduceChan(source, mapper, reducer, opts...)
|
||||||
}
|
}
|
||||||
|
|
||||||
// MapReduceWithSource maps all elements from source, and reduce the output elements with given reducer.
|
// MapReduceChan maps all elements from source, and reduce the output elements with given reducer.
|
||||||
func MapReduceWithSource(source <-chan interface{}, mapper MapperFunc, reducer ReducerFunc,
|
func MapReduceChan(source <-chan interface{}, mapper MapperFunc, reducer ReducerFunc,
|
||||||
opts ...Option) (interface{}, error) {
|
opts ...Option) (interface{}, error) {
|
||||||
options := buildOptions(opts...)
|
options := buildOptions(opts...)
|
||||||
output := make(chan interface{})
|
output := make(chan interface{})
|
||||||
@@ -119,13 +127,13 @@ func MapReduceWithSource(source <-chan interface{}, mapper MapperFunc, reducer R
|
|||||||
}()
|
}()
|
||||||
|
|
||||||
collector := make(chan interface{}, options.workers)
|
collector := make(chan interface{}, options.workers)
|
||||||
done := syncx.NewDoneChan()
|
done := make(chan lang.PlaceholderType)
|
||||||
writer := newGuardedWriter(output, done.Done())
|
writer := newGuardedWriter(options.ctx, output, done)
|
||||||
var closeOnce sync.Once
|
var closeOnce sync.Once
|
||||||
var retErr errorx.AtomicError
|
var retErr errorx.AtomicError
|
||||||
finish := func() {
|
finish := func() {
|
||||||
closeOnce.Do(func() {
|
closeOnce.Do(func() {
|
||||||
done.Close()
|
close(done)
|
||||||
close(output)
|
close(output)
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
@@ -154,17 +162,22 @@ func MapReduceWithSource(source <-chan interface{}, mapper MapperFunc, reducer R
|
|||||||
reducer(collector, writer, cancel)
|
reducer(collector, writer, cancel)
|
||||||
}()
|
}()
|
||||||
|
|
||||||
go executeMappers(func(item interface{}, w Writer) {
|
go executeMappers(options.ctx, func(item interface{}, w Writer) {
|
||||||
mapper(item, w, cancel)
|
mapper(item, w, cancel)
|
||||||
}, source, collector, done.Done(), options.workers)
|
}, source, collector, done, options.workers)
|
||||||
|
|
||||||
value, ok := <-output
|
select {
|
||||||
if err := retErr.Load(); err != nil {
|
case <-options.ctx.Done():
|
||||||
return nil, err
|
cancel(context.DeadlineExceeded)
|
||||||
} else if ok {
|
return nil, context.DeadlineExceeded
|
||||||
return value, nil
|
case value, ok := <-output:
|
||||||
} else {
|
if err := retErr.Load(); err != nil {
|
||||||
return nil, ErrReduceNoOutput
|
return nil, err
|
||||||
|
} else if ok {
|
||||||
|
return value, nil
|
||||||
|
} else {
|
||||||
|
return nil, ErrReduceNoOutput
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -173,18 +186,19 @@ func MapReduceWithSource(source <-chan interface{}, mapper MapperFunc, reducer R
|
|||||||
func MapReduceVoid(generate GenerateFunc, mapper MapperFunc, reducer VoidReducerFunc, opts ...Option) error {
|
func MapReduceVoid(generate GenerateFunc, mapper MapperFunc, reducer VoidReducerFunc, opts ...Option) error {
|
||||||
_, err := MapReduce(generate, mapper, func(input <-chan interface{}, writer Writer, cancel func(error)) {
|
_, err := MapReduce(generate, mapper, func(input <-chan interface{}, writer Writer, cancel func(error)) {
|
||||||
reducer(input, cancel)
|
reducer(input, cancel)
|
||||||
// We need to write a placeholder to let MapReduce to continue on reducer done,
|
|
||||||
// otherwise, all goroutines are waiting. The placeholder will be discarded by MapReduce.
|
|
||||||
writer.Write(lang.Placeholder)
|
|
||||||
}, opts...)
|
}, opts...)
|
||||||
|
if errors.Is(err, ErrReduceNoOutput) {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
// MapVoid maps all elements from given generate but no output.
|
// WithContext customizes a mapreduce processing accepts a given ctx.
|
||||||
func MapVoid(generate GenerateFunc, mapper VoidMapFunc, opts ...Option) {
|
func WithContext(ctx context.Context) Option {
|
||||||
drain(Map(generate, func(item interface{}, writer Writer) {
|
return func(opts *mapReduceOptions) {
|
||||||
mapper(item)
|
opts.ctx = ctx
|
||||||
}, opts...))
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// WithWorkers customizes a mapreduce processing with given workers.
|
// WithWorkers customizes a mapreduce processing with given workers.
|
||||||
@@ -224,8 +238,8 @@ func drain(channel <-chan interface{}) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func executeMappers(mapper MapFunc, input <-chan interface{}, collector chan<- interface{},
|
func executeMappers(ctx context.Context, mapper MapFunc, input <-chan interface{},
|
||||||
done <-chan lang.PlaceholderType, workers int) {
|
collector chan<- interface{}, done <-chan lang.PlaceholderType, workers int) {
|
||||||
var wg sync.WaitGroup
|
var wg sync.WaitGroup
|
||||||
defer func() {
|
defer func() {
|
||||||
wg.Wait()
|
wg.Wait()
|
||||||
@@ -233,9 +247,11 @@ func executeMappers(mapper MapFunc, input <-chan interface{}, collector chan<- i
|
|||||||
}()
|
}()
|
||||||
|
|
||||||
pool := make(chan lang.PlaceholderType, workers)
|
pool := make(chan lang.PlaceholderType, workers)
|
||||||
writer := newGuardedWriter(collector, done)
|
writer := newGuardedWriter(ctx, collector, done)
|
||||||
for {
|
for {
|
||||||
select {
|
select {
|
||||||
|
case <-ctx.Done():
|
||||||
|
return
|
||||||
case <-done:
|
case <-done:
|
||||||
return
|
return
|
||||||
case pool <- lang.Placeholder:
|
case pool <- lang.Placeholder:
|
||||||
@@ -261,6 +277,7 @@ func executeMappers(mapper MapFunc, input <-chan interface{}, collector chan<- i
|
|||||||
|
|
||||||
func newOptions() *mapReduceOptions {
|
func newOptions() *mapReduceOptions {
|
||||||
return &mapReduceOptions{
|
return &mapReduceOptions{
|
||||||
|
ctx: context.Background(),
|
||||||
workers: defaultWorkers,
|
workers: defaultWorkers,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -275,12 +292,15 @@ func once(fn func(error)) func(error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
type guardedWriter struct {
|
type guardedWriter struct {
|
||||||
|
ctx context.Context
|
||||||
channel chan<- interface{}
|
channel chan<- interface{}
|
||||||
done <-chan lang.PlaceholderType
|
done <-chan lang.PlaceholderType
|
||||||
}
|
}
|
||||||
|
|
||||||
func newGuardedWriter(channel chan<- interface{}, done <-chan lang.PlaceholderType) guardedWriter {
|
func newGuardedWriter(ctx context.Context, channel chan<- interface{},
|
||||||
|
done <-chan lang.PlaceholderType) guardedWriter {
|
||||||
return guardedWriter{
|
return guardedWriter{
|
||||||
|
ctx: ctx,
|
||||||
channel: channel,
|
channel: channel,
|
||||||
done: done,
|
done: done,
|
||||||
}
|
}
|
||||||
@@ -288,6 +308,8 @@ func newGuardedWriter(channel chan<- interface{}, done <-chan lang.PlaceholderTy
|
|||||||
|
|
||||||
func (gw guardedWriter) Write(v interface{}) {
|
func (gw guardedWriter) Write(v interface{}) {
|
||||||
select {
|
select {
|
||||||
|
case <-gw.ctx.Done():
|
||||||
|
return
|
||||||
case <-gw.done:
|
case <-gw.done:
|
||||||
return
|
return
|
||||||
default:
|
default:
|
||||||
|
|||||||
@@ -1,6 +1,7 @@
|
|||||||
package mr
|
package mr
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"context"
|
||||||
"errors"
|
"errors"
|
||||||
"io/ioutil"
|
"io/ioutil"
|
||||||
"log"
|
"log"
|
||||||
@@ -10,8 +11,9 @@ import (
|
|||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/tal-tech/go-zero/core/stringx"
|
"github.com/zeromicro/go-zero/core/stringx"
|
||||||
"github.com/tal-tech/go-zero/core/syncx"
|
"github.com/zeromicro/go-zero/core/syncx"
|
||||||
|
"go.uber.org/goleak"
|
||||||
)
|
)
|
||||||
|
|
||||||
var errDummy = errors.New("dummy")
|
var errDummy = errors.New("dummy")
|
||||||
@@ -21,6 +23,8 @@ func init() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestFinish(t *testing.T) {
|
func TestFinish(t *testing.T) {
|
||||||
|
defer goleak.VerifyNone(t)
|
||||||
|
|
||||||
var total uint32
|
var total uint32
|
||||||
err := Finish(func() error {
|
err := Finish(func() error {
|
||||||
atomic.AddUint32(&total, 2)
|
atomic.AddUint32(&total, 2)
|
||||||
@@ -38,14 +42,20 @@ func TestFinish(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestFinishNone(t *testing.T) {
|
func TestFinishNone(t *testing.T) {
|
||||||
|
defer goleak.VerifyNone(t)
|
||||||
|
|
||||||
assert.Nil(t, Finish())
|
assert.Nil(t, Finish())
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestFinishVoidNone(t *testing.T) {
|
func TestFinishVoidNone(t *testing.T) {
|
||||||
|
defer goleak.VerifyNone(t)
|
||||||
|
|
||||||
FinishVoid()
|
FinishVoid()
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestFinishErr(t *testing.T) {
|
func TestFinishErr(t *testing.T) {
|
||||||
|
defer goleak.VerifyNone(t)
|
||||||
|
|
||||||
var total uint32
|
var total uint32
|
||||||
err := Finish(func() error {
|
err := Finish(func() error {
|
||||||
atomic.AddUint32(&total, 2)
|
atomic.AddUint32(&total, 2)
|
||||||
@@ -62,6 +72,8 @@ func TestFinishErr(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestFinishVoid(t *testing.T) {
|
func TestFinishVoid(t *testing.T) {
|
||||||
|
defer goleak.VerifyNone(t)
|
||||||
|
|
||||||
var total uint32
|
var total uint32
|
||||||
FinishVoid(func() {
|
FinishVoid(func() {
|
||||||
atomic.AddUint32(&total, 2)
|
atomic.AddUint32(&total, 2)
|
||||||
@@ -74,7 +86,57 @@ func TestFinishVoid(t *testing.T) {
|
|||||||
assert.Equal(t, uint32(10), atomic.LoadUint32(&total))
|
assert.Equal(t, uint32(10), atomic.LoadUint32(&total))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func TestForEach(t *testing.T) {
|
||||||
|
const tasks = 1000
|
||||||
|
|
||||||
|
t.Run("all", func(t *testing.T) {
|
||||||
|
defer goleak.VerifyNone(t)
|
||||||
|
|
||||||
|
var count uint32
|
||||||
|
ForEach(func(source chan<- interface{}) {
|
||||||
|
for i := 0; i < tasks; i++ {
|
||||||
|
source <- i
|
||||||
|
}
|
||||||
|
}, func(item interface{}) {
|
||||||
|
atomic.AddUint32(&count, 1)
|
||||||
|
}, WithWorkers(-1))
|
||||||
|
|
||||||
|
assert.Equal(t, tasks, int(count))
|
||||||
|
})
|
||||||
|
|
||||||
|
t.Run("odd", func(t *testing.T) {
|
||||||
|
defer goleak.VerifyNone(t)
|
||||||
|
|
||||||
|
var count uint32
|
||||||
|
ForEach(func(source chan<- interface{}) {
|
||||||
|
for i := 0; i < tasks; i++ {
|
||||||
|
source <- i
|
||||||
|
}
|
||||||
|
}, func(item interface{}) {
|
||||||
|
if item.(int)%2 == 0 {
|
||||||
|
atomic.AddUint32(&count, 1)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
|
||||||
|
assert.Equal(t, tasks/2, int(count))
|
||||||
|
})
|
||||||
|
|
||||||
|
t.Run("all", func(t *testing.T) {
|
||||||
|
defer goleak.VerifyNone(t)
|
||||||
|
|
||||||
|
ForEach(func(source chan<- interface{}) {
|
||||||
|
for i := 0; i < tasks; i++ {
|
||||||
|
source <- i
|
||||||
|
}
|
||||||
|
}, func(item interface{}) {
|
||||||
|
panic("foo")
|
||||||
|
})
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
func TestMap(t *testing.T) {
|
func TestMap(t *testing.T) {
|
||||||
|
defer goleak.VerifyNone(t)
|
||||||
|
|
||||||
tests := []struct {
|
tests := []struct {
|
||||||
mapper MapFunc
|
mapper MapFunc
|
||||||
expect int
|
expect int
|
||||||
@@ -127,6 +189,8 @@ func TestMap(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestMapReduce(t *testing.T) {
|
func TestMapReduce(t *testing.T) {
|
||||||
|
defer goleak.VerifyNone(t)
|
||||||
|
|
||||||
tests := []struct {
|
tests := []struct {
|
||||||
mapper MapperFunc
|
mapper MapperFunc
|
||||||
reducer ReducerFunc
|
reducer ReducerFunc
|
||||||
@@ -202,7 +266,22 @@ func TestMapReduce(t *testing.T) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func TestMapReducePanicBothMapperAndReducer(t *testing.T) {
|
||||||
|
defer goleak.VerifyNone(t)
|
||||||
|
|
||||||
|
_, _ = MapReduce(func(source chan<- interface{}) {
|
||||||
|
source <- 0
|
||||||
|
source <- 1
|
||||||
|
}, func(item interface{}, writer Writer, cancel func(error)) {
|
||||||
|
panic("foo")
|
||||||
|
}, func(pipe <-chan interface{}, writer Writer, cancel func(error)) {
|
||||||
|
panic("bar")
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
func TestMapReduceWithReduerWriteMoreThanOnce(t *testing.T) {
|
func TestMapReduceWithReduerWriteMoreThanOnce(t *testing.T) {
|
||||||
|
defer goleak.VerifyNone(t)
|
||||||
|
|
||||||
assert.Panics(t, func() {
|
assert.Panics(t, func() {
|
||||||
MapReduce(func(source chan<- interface{}) {
|
MapReduce(func(source chan<- interface{}) {
|
||||||
for i := 0; i < 10; i++ {
|
for i := 0; i < 10; i++ {
|
||||||
@@ -219,6 +298,8 @@ func TestMapReduceWithReduerWriteMoreThanOnce(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestMapReduceVoid(t *testing.T) {
|
func TestMapReduceVoid(t *testing.T) {
|
||||||
|
defer goleak.VerifyNone(t)
|
||||||
|
|
||||||
var value uint32
|
var value uint32
|
||||||
tests := []struct {
|
tests := []struct {
|
||||||
mapper MapperFunc
|
mapper MapperFunc
|
||||||
@@ -295,6 +376,8 @@ func TestMapReduceVoid(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestMapReduceVoidWithDelay(t *testing.T) {
|
func TestMapReduceVoidWithDelay(t *testing.T) {
|
||||||
|
defer goleak.VerifyNone(t)
|
||||||
|
|
||||||
var result []int
|
var result []int
|
||||||
err := MapReduceVoid(func(source chan<- interface{}) {
|
err := MapReduceVoid(func(source chan<- interface{}) {
|
||||||
source <- 0
|
source <- 0
|
||||||
@@ -318,9 +401,11 @@ func TestMapReduceVoidWithDelay(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestMapVoid(t *testing.T) {
|
func TestMapVoid(t *testing.T) {
|
||||||
|
defer goleak.VerifyNone(t)
|
||||||
|
|
||||||
const tasks = 1000
|
const tasks = 1000
|
||||||
var count uint32
|
var count uint32
|
||||||
MapVoid(func(source chan<- interface{}) {
|
ForEach(func(source chan<- interface{}) {
|
||||||
for i := 0; i < tasks; i++ {
|
for i := 0; i < tasks; i++ {
|
||||||
source <- i
|
source <- i
|
||||||
}
|
}
|
||||||
@@ -332,6 +417,8 @@ func TestMapVoid(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestMapReducePanic(t *testing.T) {
|
func TestMapReducePanic(t *testing.T) {
|
||||||
|
defer goleak.VerifyNone(t)
|
||||||
|
|
||||||
v, err := MapReduce(func(source chan<- interface{}) {
|
v, err := MapReduce(func(source chan<- interface{}) {
|
||||||
source <- 0
|
source <- 0
|
||||||
source <- 1
|
source <- 1
|
||||||
@@ -349,6 +436,8 @@ func TestMapReducePanic(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestMapReduceVoidCancel(t *testing.T) {
|
func TestMapReduceVoidCancel(t *testing.T) {
|
||||||
|
defer goleak.VerifyNone(t)
|
||||||
|
|
||||||
var result []int
|
var result []int
|
||||||
err := MapReduceVoid(func(source chan<- interface{}) {
|
err := MapReduceVoid(func(source chan<- interface{}) {
|
||||||
source <- 0
|
source <- 0
|
||||||
@@ -370,6 +459,8 @@ func TestMapReduceVoidCancel(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestMapReduceVoidCancelWithRemains(t *testing.T) {
|
func TestMapReduceVoidCancelWithRemains(t *testing.T) {
|
||||||
|
defer goleak.VerifyNone(t)
|
||||||
|
|
||||||
var done syncx.AtomicBool
|
var done syncx.AtomicBool
|
||||||
var result []int
|
var result []int
|
||||||
err := MapReduceVoid(func(source chan<- interface{}) {
|
err := MapReduceVoid(func(source chan<- interface{}) {
|
||||||
@@ -395,6 +486,8 @@ func TestMapReduceVoidCancelWithRemains(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestMapReduceWithoutReducerWrite(t *testing.T) {
|
func TestMapReduceWithoutReducerWrite(t *testing.T) {
|
||||||
|
defer goleak.VerifyNone(t)
|
||||||
|
|
||||||
uids := []int{1, 2, 3}
|
uids := []int{1, 2, 3}
|
||||||
res, err := MapReduce(func(source chan<- interface{}) {
|
res, err := MapReduce(func(source chan<- interface{}) {
|
||||||
for _, uid := range uids {
|
for _, uid := range uids {
|
||||||
@@ -410,6 +503,54 @@ func TestMapReduceWithoutReducerWrite(t *testing.T) {
|
|||||||
assert.Nil(t, res)
|
assert.Nil(t, res)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func TestMapReduceVoidPanicInReducer(t *testing.T) {
|
||||||
|
defer goleak.VerifyNone(t)
|
||||||
|
|
||||||
|
const message = "foo"
|
||||||
|
var done syncx.AtomicBool
|
||||||
|
err := MapReduceVoid(func(source chan<- interface{}) {
|
||||||
|
for i := 0; i < defaultWorkers*2; i++ {
|
||||||
|
source <- i
|
||||||
|
}
|
||||||
|
done.Set(true)
|
||||||
|
}, func(item interface{}, writer Writer, cancel func(error)) {
|
||||||
|
i := item.(int)
|
||||||
|
writer.Write(i)
|
||||||
|
}, func(pipe <-chan interface{}, cancel func(error)) {
|
||||||
|
panic(message)
|
||||||
|
}, WithWorkers(1))
|
||||||
|
assert.NotNil(t, err)
|
||||||
|
assert.Equal(t, message, err.Error())
|
||||||
|
assert.True(t, done.True())
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestMapReduceWithContext(t *testing.T) {
|
||||||
|
defer goleak.VerifyNone(t)
|
||||||
|
|
||||||
|
var done syncx.AtomicBool
|
||||||
|
var result []int
|
||||||
|
ctx, cancel := context.WithCancel(context.Background())
|
||||||
|
err := MapReduceVoid(func(source chan<- interface{}) {
|
||||||
|
for i := 0; i < defaultWorkers*2; i++ {
|
||||||
|
source <- i
|
||||||
|
}
|
||||||
|
done.Set(true)
|
||||||
|
}, func(item interface{}, writer Writer, c func(error)) {
|
||||||
|
i := item.(int)
|
||||||
|
if i == defaultWorkers/2 {
|
||||||
|
cancel()
|
||||||
|
}
|
||||||
|
writer.Write(i)
|
||||||
|
}, func(pipe <-chan interface{}, cancel func(error)) {
|
||||||
|
for item := range pipe {
|
||||||
|
i := item.(int)
|
||||||
|
result = append(result, i)
|
||||||
|
}
|
||||||
|
}, WithContext(ctx))
|
||||||
|
assert.NotNil(t, err)
|
||||||
|
assert.Equal(t, context.DeadlineExceeded, err)
|
||||||
|
}
|
||||||
|
|
||||||
func BenchmarkMapReduce(b *testing.B) {
|
func BenchmarkMapReduce(b *testing.B) {
|
||||||
b.ReportAllocs()
|
b.ReportAllocs()
|
||||||
|
|
||||||
|
|||||||
89
core/mr/readme-cn.md
Normal file
89
core/mr/readme-cn.md
Normal file
@@ -0,0 +1,89 @@
|
|||||||
|
# mapreduce
|
||||||
|
|
||||||
|
[English](readme.md) | 简体中文
|
||||||
|
|
||||||
|
## 为什么需要 MapReduce
|
||||||
|
|
||||||
|
在实际的业务场景中我们常常需要从不同的 rpc 服务中获取相应属性来组装成复杂对象。
|
||||||
|
|
||||||
|
比如要查询商品详情:
|
||||||
|
|
||||||
|
1. 商品服务-查询商品属性
|
||||||
|
2. 库存服务-查询库存属性
|
||||||
|
3. 价格服务-查询价格属性
|
||||||
|
4. 营销服务-查询营销属性
|
||||||
|
|
||||||
|
如果是串行调用的话响应时间会随着 rpc 调用次数呈线性增长,所以我们要优化性能一般会将串行改并行。
|
||||||
|
|
||||||
|
简单的场景下使用 `WaitGroup` 也能够满足需求,但是如果我们需要对 rpc 调用返回的数据进行校验、数据加工转换、数据汇总呢?继续使用 `WaitGroup` 就有点力不从心了,go 的官方库中并没有这种工具(java 中提供了 CompleteFuture),我们依据 MapReduce 架构思想实现了进程内的数据批处理 MapReduce 并发工具类。
|
||||||
|
|
||||||
|
## 设计思路
|
||||||
|
|
||||||
|
我们尝试把自己代入到作者的角色梳理一下并发工具可能的业务场景:
|
||||||
|
|
||||||
|
1. 查询商品详情:支持并发调用多个服务来组合产品属性,支持调用错误可以立即结束。
|
||||||
|
2. 商品详情页自动推荐用户卡券:支持并发校验卡券,校验失败自动剔除,返回全部卡券。
|
||||||
|
|
||||||
|
以上实际都是在进行对输入数据进行处理最后输出清洗后的数据,针对数据处理有个非常经典的异步模式:生产者消费者模式。于是我们可以抽象一下数据批处理的生命周期,大致可以分为三个阶段:
|
||||||
|
|
||||||
|
<img src="https://raw.githubusercontent.com/zeromicro/zero-doc/main/doc/images/mapreduce-serial-cn.png" width="500">
|
||||||
|
|
||||||
|
1. 数据生产 generate
|
||||||
|
2. 数据加工 mapper
|
||||||
|
3. 数据聚合 reducer
|
||||||
|
|
||||||
|
其中数据生产是不可或缺的阶段,数据加工、数据聚合是可选阶段,数据生产与加工支持并发调用,数据聚合基本属于纯内存操作单协程即可。
|
||||||
|
|
||||||
|
再来思考一下不同阶段之间数据应该如何流转,既然不同阶段的数据处理都是由不同 goroutine 执行的,那么很自然的可以考虑采用 channel 来实现 goroutine 之间的通信。
|
||||||
|
|
||||||
|
<img src="https://raw.githubusercontent.com/zeromicro/zero-doc/main/doc/images/mapreduce-cn.png" width="500">
|
||||||
|
|
||||||
|
|
||||||
|
如何实现随时终止流程呢?
|
||||||
|
|
||||||
|
`goroutine` 中监听一个全局的结束 `channel` 和调用方提供的 `ctx` 就行。
|
||||||
|
|
||||||
|
## 简单示例
|
||||||
|
|
||||||
|
并行求平方和(不要嫌弃示例简单,只是模拟并发)
|
||||||
|
|
||||||
|
```go
|
||||||
|
package main
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"log"
|
||||||
|
|
||||||
|
"github.com/tal-tech/go-zero/core/mr"
|
||||||
|
)
|
||||||
|
|
||||||
|
func main() {
|
||||||
|
val, err := mr.MapReduce(func(source chan<- interface{}) {
|
||||||
|
// generator
|
||||||
|
for i := 0; i < 10; i++ {
|
||||||
|
source <- i
|
||||||
|
}
|
||||||
|
}, func(item interface{}, writer mr.Writer, cancel func(error)) {
|
||||||
|
// mapper
|
||||||
|
i := item.(int)
|
||||||
|
writer.Write(i * i)
|
||||||
|
}, func(pipe <-chan interface{}, writer mr.Writer, cancel func(error)) {
|
||||||
|
// reducer
|
||||||
|
var sum int
|
||||||
|
for i := range pipe {
|
||||||
|
sum += i.(int)
|
||||||
|
}
|
||||||
|
writer.Write(sum)
|
||||||
|
})
|
||||||
|
if err != nil {
|
||||||
|
log.Fatal(err)
|
||||||
|
}
|
||||||
|
fmt.Println("result:", val)
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
更多示例:[https://github.com/zeromicro/zero-examples/tree/main/mapreduce](https://github.com/zeromicro/zero-examples/tree/main/mapreduce)
|
||||||
|
|
||||||
|
## 欢迎 star!⭐
|
||||||
|
|
||||||
|
如果你正在使用或者觉得这个项目对你有帮助,请 **star** 支持,感谢!
|
||||||
90
core/mr/readme.md
Normal file
90
core/mr/readme.md
Normal file
@@ -0,0 +1,90 @@
|
|||||||
|
<img align="right" width="150px" src="https://raw.githubusercontent.com/zeromicro/zero-doc/main/doc/images/go-zero.png">
|
||||||
|
|
||||||
|
# mapreduce
|
||||||
|
|
||||||
|
English | [简体中文](readme-cn.md)
|
||||||
|
|
||||||
|
## Why MapReduce is needed
|
||||||
|
|
||||||
|
In practical business scenarios we often need to get the corresponding properties from different rpc services to assemble complex objects.
|
||||||
|
|
||||||
|
For example, to query product details.
|
||||||
|
|
||||||
|
1. product service - query product attributes
|
||||||
|
2. inventory service - query inventory properties
|
||||||
|
3. price service - query price attributes
|
||||||
|
4. marketing service - query marketing properties
|
||||||
|
|
||||||
|
If it is a serial call, the response time will increase linearly with the number of rpc calls, so we will generally change serial to parallel to optimize response time.
|
||||||
|
|
||||||
|
Simple scenarios using `WaitGroup` can also meet the needs, but what if we need to check the data returned by the rpc call, data processing, data aggregation? The official go library does not have such a tool (CompleteFuture is provided in java), so we implemented an in-process data batching MapReduce concurrent tool based on the MapReduce architecture.
|
||||||
|
|
||||||
|
## Design ideas
|
||||||
|
|
||||||
|
Let's try to put ourselves in the author's shoes and sort out the possible business scenarios for the concurrency tool:
|
||||||
|
|
||||||
|
1. querying product details: supporting concurrent calls to multiple services to combine product attributes, and supporting call errors that can be ended immediately.
|
||||||
|
2. automatic recommendation of user card coupons on product details page: support concurrently verifying card coupons, automatically rejecting them if they fail, and returning all of them.
|
||||||
|
|
||||||
|
The above is actually processing the input data and finally outputting the cleaned data. There is a very classic asynchronous pattern for data processing: the producer-consumer pattern. So we can abstract the life cycle of data batch processing, which can be roughly divided into three phases.
|
||||||
|
|
||||||
|
<img src="https://raw.githubusercontent.com/zeromicro/zero-doc/main/doc/images/mapreduce-serial-en.png" width="500">
|
||||||
|
|
||||||
|
1. data production generate
|
||||||
|
2. data processing mapper
|
||||||
|
3. data aggregation reducer
|
||||||
|
|
||||||
|
Data producing is an indispensable stage, data processing and data aggregation are optional stages, data producing and processing support concurrent calls, data aggregation is basically a pure memory operation, so a single concurrent process can do it.
|
||||||
|
|
||||||
|
Since different stages of data processing are performed by different goroutines, it is natural to consider the use of channel to achieve communication between goroutines.
|
||||||
|
|
||||||
|
<img src="https://raw.githubusercontent.com/zeromicro/zero-doc/main/doc/images/mapreduce-en.png" width="500">
|
||||||
|
|
||||||
|
How can I terminate the process at any time?
|
||||||
|
|
||||||
|
It's simple, just receive from a channel or the given context in the goroutine.
|
||||||
|
|
||||||
|
## A simple example
|
||||||
|
|
||||||
|
Calculate the sum of squares, simulating the concurrency.
|
||||||
|
|
||||||
|
```go
|
||||||
|
package main
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"log"
|
||||||
|
|
||||||
|
"github.com/tal-tech/go-zero/core/mr"
|
||||||
|
)
|
||||||
|
|
||||||
|
func main() {
|
||||||
|
val, err := mr.MapReduce(func(source chan<- interface{}) {
|
||||||
|
// generator
|
||||||
|
for i := 0; i < 10; i++ {
|
||||||
|
source <- i
|
||||||
|
}
|
||||||
|
}, func(item interface{}, writer mr.Writer, cancel func(error)) {
|
||||||
|
// mapper
|
||||||
|
i := item.(int)
|
||||||
|
writer.Write(i * i)
|
||||||
|
}, func(pipe <-chan interface{}, writer mr.Writer, cancel func(error)) {
|
||||||
|
// reducer
|
||||||
|
var sum int
|
||||||
|
for i := range pipe {
|
||||||
|
sum += i.(int)
|
||||||
|
}
|
||||||
|
writer.Write(sum)
|
||||||
|
})
|
||||||
|
if err != nil {
|
||||||
|
log.Fatal(err)
|
||||||
|
}
|
||||||
|
fmt.Println("result:", val)
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
More examples: [https://github.com/zeromicro/zero-examples/tree/main/mapreduce](https://github.com/zeromicro/zero-examples/tree/main/mapreduce)
|
||||||
|
|
||||||
|
## Give a Star! ⭐
|
||||||
|
|
||||||
|
If you like or are using this project to learn or start your solution, please give it a star. Thanks!
|
||||||
@@ -11,7 +11,7 @@ import (
|
|||||||
"syscall"
|
"syscall"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/logx"
|
"github.com/zeromicro/go-zero/core/logx"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
|
|||||||
@@ -15,7 +15,7 @@ import (
|
|||||||
"syscall"
|
"syscall"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/logx"
|
"github.com/zeromicro/go-zero/core/logx"
|
||||||
)
|
)
|
||||||
|
|
||||||
// DefaultMemProfileRate is the default memory profiling rate.
|
// DefaultMemProfileRate is the default memory profiling rate.
|
||||||
|
|||||||
@@ -10,7 +10,8 @@ import (
|
|||||||
"syscall"
|
"syscall"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/logx"
|
"github.com/zeromicro/go-zero/core/logx"
|
||||||
|
"github.com/zeromicro/go-zero/core/threading"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
@@ -46,10 +47,10 @@ func gracefulStop(signals chan os.Signal) {
|
|||||||
signal.Stop(signals)
|
signal.Stop(signals)
|
||||||
|
|
||||||
logx.Info("Got signal SIGTERM, shutting down...")
|
logx.Info("Got signal SIGTERM, shutting down...")
|
||||||
wrapUpListeners.notifyListeners()
|
go wrapUpListeners.notifyListeners()
|
||||||
|
|
||||||
time.Sleep(wrapUpTime)
|
time.Sleep(wrapUpTime)
|
||||||
shutdownListeners.notifyListeners()
|
go shutdownListeners.notifyListeners()
|
||||||
|
|
||||||
time.Sleep(delayTimeBeforeForceQuit - wrapUpTime)
|
time.Sleep(delayTimeBeforeForceQuit - wrapUpTime)
|
||||||
logx.Infof("Still alive after %v, going to force kill the process...", delayTimeBeforeForceQuit)
|
logx.Infof("Still alive after %v, going to force kill the process...", delayTimeBeforeForceQuit)
|
||||||
@@ -81,7 +82,9 @@ func (lm *listenerManager) notifyListeners() {
|
|||||||
lm.lock.Lock()
|
lm.lock.Lock()
|
||||||
defer lm.lock.Unlock()
|
defer lm.lock.Unlock()
|
||||||
|
|
||||||
|
group := threading.NewRoutineGroup()
|
||||||
for _, listener := range lm.listeners {
|
for _, listener := range lm.listeners {
|
||||||
listener()
|
group.RunSafe(listener)
|
||||||
}
|
}
|
||||||
|
group.Wait()
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -8,7 +8,7 @@ import (
|
|||||||
"os/signal"
|
"os/signal"
|
||||||
"syscall"
|
"syscall"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/logx"
|
"github.com/zeromicro/go-zero/core/logx"
|
||||||
)
|
)
|
||||||
|
|
||||||
const timeFormat = "0102150405"
|
const timeFormat = "0102150405"
|
||||||
|
|||||||
@@ -8,8 +8,8 @@ import (
|
|||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/olekukonko/tablewriter"
|
"github.com/olekukonko/tablewriter"
|
||||||
"github.com/tal-tech/go-zero/core/logx"
|
"github.com/zeromicro/go-zero/core/logx"
|
||||||
"github.com/tal-tech/go-zero/core/threading"
|
"github.com/zeromicro/go-zero/core/threading"
|
||||||
)
|
)
|
||||||
|
|
||||||
type (
|
type (
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
package prof
|
package prof
|
||||||
|
|
||||||
import "github.com/tal-tech/go-zero/core/utils"
|
import "github.com/zeromicro/go-zero/core/utils"
|
||||||
|
|
||||||
type (
|
type (
|
||||||
// A ProfilePoint is a profile time point.
|
// A ProfilePoint is a profile time point.
|
||||||
|
|||||||
@@ -3,7 +3,7 @@ package prof
|
|||||||
import (
|
import (
|
||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"github.com/tal-tech/go-zero/core/utils"
|
"github.com/zeromicro/go-zero/core/utils"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestProfiler(t *testing.T) {
|
func TestProfiler(t *testing.T) {
|
||||||
|
|||||||
@@ -6,9 +6,9 @@ import (
|
|||||||
"sync"
|
"sync"
|
||||||
|
|
||||||
"github.com/prometheus/client_golang/prometheus/promhttp"
|
"github.com/prometheus/client_golang/prometheus/promhttp"
|
||||||
"github.com/tal-tech/go-zero/core/logx"
|
"github.com/zeromicro/go-zero/core/logx"
|
||||||
"github.com/tal-tech/go-zero/core/syncx"
|
"github.com/zeromicro/go-zero/core/syncx"
|
||||||
"github.com/tal-tech/go-zero/core/threading"
|
"github.com/zeromicro/go-zero/core/threading"
|
||||||
)
|
)
|
||||||
|
|
||||||
var (
|
var (
|
||||||
|
|||||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user