Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Bug]: Loading collection stuck at 0% after reboot #34531

Open
1 task done
wjddd opened this issue Jul 9, 2024 · 10 comments
Open
1 task done

[Bug]: Loading collection stuck at 0% after reboot #34531

wjddd opened this issue Jul 9, 2024 · 10 comments
Assignees
Labels
kind/bug Issues or changes related a bug triage/accepted Indicates an issue or PR is ready to be actively worked on.

Comments

@wjddd
Copy link

wjddd commented Jul 9, 2024

Is there an existing issue for this?

  • I have searched the existing issues

Environment

- Milvus version: 2.3.5
- Deployment mode(standalone or cluster): standalone
- SDK version(e.g. pymilvus v2.0.0rc2): pymilvus v2.3.5
- OS(Ubuntu or CentOS): centos7
- CPU/Memory: 64c128g

Current Behavior

65 collection in total but only 4 of them successfully loaded, others just keep loading.

Milvus Log

milvus.zip

@wjddd wjddd added kind/bug Issues or changes related a bug needs-triage Indicates an issue or PR lacks a `triage/foo` label and requires one. labels Jul 9, 2024
@yanliang567
Copy link
Contributor

/assign @longjiquan
/unassign

@yanliang567 yanliang567 added triage/accepted Indicates an issue or PR is ready to be actively worked on. and removed needs-triage Indicates an issue or PR lacks a `triage/foo` label and requires one. labels Jul 10, 2024
@wjddd
Copy link
Author

wjddd commented Jul 12, 2024

@longjiquan Hi, is there any solution for this?

@xiaofan-luan
Copy link
Contributor

/assign @bigsheeper

@xiaofan-luan
Copy link
Contributor

@longjiquan Hi, is there any solution for this?

What happened after reboot?
Did your disk ever be full?

@xiaofan-luan
Copy link
Contributor

what's the reason of it's reboot?
please describe what happens to this cluster

@wjddd
Copy link
Author

wjddd commented Jul 16, 2024

what's the reason of it's reboot? please describe what happens to this cluster

It's just a routine maintenance, we have enough free space for milvus.
Everything is ok except milvus got crashed.

@bigsheeper
Copy link
Contributor

what's the reason of it's reboot? please describe what happens to this cluster

It's just a routine maintenance, we have enough free space for milvus. Everything is ok except milvus got crashed.

@wjddd Could you provide the logs from just before the Milvus crash? Were there any unusual occurrences at the time of the crash?

@wjddd
Copy link
Author

wjddd commented Jul 16, 2024

what's the reason of it's reboot? please describe what happens to this cluster

It's just a routine maintenance, we have enough free space for milvus. Everything is ok except milvus got crashed.

@wjddd Could you provide the logs from just before the Milvus crash? Were there any unusual occurrences at the time of the crash?

We don't keep them for so long. All the logs that we can find are provided in milvus.zip. Everything was fine before reboot. After reboot, the loading got stuck, so we deleted more than 5000 empty collections and restarted all the containers.

@xiaofan-luan
Copy link
Contributor

Let's try the reproduce like this:
create 10000 collections,
each collection write 100 entities.
kill the cluster and see if it can recovery @wjddd any suggestions?

BTW, we don't recommend to create so much collections

@wjddd
Copy link
Author

wjddd commented Jul 16, 2024

Let's try the reproduce like this: create 10000 collections, each collection write 100 entities. kill the cluster and see if it can recovery @wjddd any suggestions?

BTW, we don't recommend to create so much collections

OK, I'll give it a try.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
kind/bug Issues or changes related a bug triage/accepted Indicates an issue or PR is ready to be actively worked on.
5 participants