milvus
milvus copied to clipboard
[Bug]: v2.1.1 Service often goes down
Is there an existing issue for this?
- [X] I have searched the existing issues
Environment
- Milvus version:2.1.1
- Deployment mode(standalone or cluster):standalone
- SDK version(e.g. pymilvus v2.0.0rc2):
- OS(Ubuntu or CentOS):
- CPU/Memory:
- GPU:
- Others:
Current Behavior
[2022/09/04 22:29:18.505 +00:00] [ERROR] [datanode/flow_graph_delete_node.go:248] ["failed to flush delete data, err = All attempts results:\nattempt #1:invalid start/end timestamp\nattempt #2:invalid start/end timestamp\nattempt #3:invalid start/end timestamp\nattempt #4:invalid start/end timestamp\nattempt #5:invalid start/end timestamp\n"] [stack="github.com/milvus-io/milvus/internal/datanode.(*deleteNode).Operate\n\t/go/src/github.com/milvus-io/milvus/internal/datanode/flow_graph_delete_node.go:248\ngithub.com/milvus-io/milvus/internal/util/flowgraph.(*nodeCtx).work\n\t/go/src/github.com/milvus-io/milvus/internal/util/flowgraph/node.go:102"] panic: failed to flush delete data, err = All attempts results: attempt #1:invalid start/end timestamp attempt #2:invalid start/end timestamp attempt #3:invalid start/end timestamp attempt #4:invalid start/end timestamp attempt #5:invalid start/end timestamp goroutine 543862 [running]: github.com/milvus-io/milvus/internal/datanode.(*deleteNode).Operate(0xc004045ef0, 0xc00a835c50, 0x1, 0x1, 0x0, 0x0, 0x0) /go/src/github.com/milvus-io/milvus/internal/datanode/flow_graph_delete_node.go:249 +0x10c5 github.com/milvus-io/milvus/internal/util/flowgraph.(*nodeCtx).work(0xc03219ab80) /go/src/github.com/milvus-io/milvus/internal/util/flowgraph/node.go:102 +0x23b created by github.com/milvus-io/milvus/internal/util/flowgraph.(*nodeCtx).Start /go/src/github.com/milvus-io/milvus/internal/util/flowgraph/node.go:70 +0x70
Expected Behavior
No response
Steps To Reproduce
No response
Milvus Log
[2022/09/04 22:29:18.505 +00:00] [ERROR] [datanode/flow_graph_delete_node.go:248] ["failed to flush delete data, err = All attempts results:\nattempt #1:invalid start/end timestamp\nattempt #2:invalid start/end timestamp\nattempt #3:invalid start/end timestamp\nattempt #4:invalid start/end timestamp\nattempt #5:invalid start/end timestamp\n"] [stack="github.com/milvus-io/milvus/internal/datanode.(*deleteNode).Operate\n\t/go/src/github.com/milvus-io/milvus/internal/datanode/flow_graph_delete_node.go:248\ngithub.com/milvus-io/milvus/internal/util/flowgraph.(*nodeCtx).work\n\t/go/src/github.com/milvus-io/milvus/internal/util/flowgraph/node.go:102"] panic: failed to flush delete data, err = All attempts results: attempt #1:invalid start/end timestamp attempt #2:invalid start/end timestamp attempt #3:invalid start/end timestamp attempt #4:invalid start/end timestamp attempt #5:invalid start/end timestamp goroutine 543862 [running]: github.com/milvus-io/milvus/internal/datanode.(*deleteNode).Operate(0xc004045ef0, 0xc00a835c50, 0x1, 0x1, 0x0, 0x0, 0x0) /go/src/github.com/milvus-io/milvus/internal/datanode/flow_graph_delete_node.go:249 +0x10c5 github.com/milvus-io/milvus/internal/util/flowgraph.(*nodeCtx).work(0xc03219ab80) /go/src/github.com/milvus-io/milvus/internal/util/flowgraph/node.go:102 +0x23b created by github.com/milvus-io/milvus/internal/util/flowgraph.(*nodeCtx).Start /go/src/github.com/milvus-io/milvus/internal/util/flowgraph/node.go:70 +0x70
Anything else?
No response
@bestRiven thank you for the issue, could you please share more details about the reproduce steps? Also please refer this script to export the whole Milvus logs for investigation.
/assign @bestRiven /unassign
log: 链接: https://pan.baidu.com/s/1TKyDEk0FNLc8VhP3Nd66ag 提取码: 9h6q
I think this is a duplicate of https://github.com/milvus-io/milvus/issues/18565
/assign @XuanYang-cn /unassign @bestRiven
@bestRiven Milvus 2.1.2 is released. You can this new version to check whether your problem is solved.
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions.
Rotten issues close after 30d of inactivity. Reopen the issue with /reopen
.