Replies: 1 comment
-
We did fix the situation by keeping the pool but sending a fresh copy of the dataset. The corruption has not resurfaced on the new dataset. I see several closed issues about send/receive corruption with 1M datasets. It was hopefully one of those we hit and the bug is already fixed. Unfortunately, I cannot know for sure since the replication history of the original dataset is not fully documented. As this discussion has not received any response I am now going to destroy the faulty dataset and further investigation will not be possible. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
We have experienced silent data corruption on a replication target. I first thought the problem was related to #6224 but this happened on TrueNAS CORE 12.0-U5 which should already have a fix for that issue.
Most files are okay, some are holes (i.e. zeroed) and other files look like this:
That looks like 1MB of data in a 128k record? (The dataset currently has a 1M recordsize but this particular file has 128k records on the replication source)
I can
zfs send
the bad dataset, with a resulting stream like this:but the kernel panics on
zfs receive
:Is there some known bug/gotcha that can cause this situation?
Can I fix the situation by destroying the affected dataset or should I rebuild the entire pool?
Beta Was this translation helpful? Give feedback.
All reactions