NetBSD-Users archive
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index][Old Index]
Re: dk for z/lvm volumes?
On Wed, 26 Feb 2020 at 21:16, Michael van Elst <mlelstv%serpens.de@localhost> wrote:
>
> ci4ic4%gmail.com@localhost (Chavdar Ivanov) writes:
>
> ># vnconfig -c vnd0 /dev/zvol/rdsk/pail/nbsd1
> >vnconfig: /dev/rvnd0: VNDIOCSET: Operation not supported
>
> If zvols aren't broken then this should work:
>
> ccdconfig -c ccd0 0 /dev/zvol/rdsk/pail/nbsd1
>
> and then access /dev/ccd0[a-p]
# ccdconfig -c ccd0 0 /dev/zvol/rdsk/pail/nbsd1
ccdconfig: ioctl (CCDIOCSET): /dev/ccd0d: Block device required
# ccdconfig -c ccd0 0 /dev/zvol/dsk/pail/nbsd1
ccdconfig: ioctl (CCDIOCSET): /dev/ccd0d: Inappropriate ioctl for device
Apparently not, in this case. I think the comments in the thread
suggest that already.
FWIW I tried the sequence below, which could be used to accomplish the
original question -
Under NetBSD-current:
- created a zvol
- iSCSI exported it (now using net/istgt)
- attached it from a W10 machine, initialized, copied a few files
- created a snapshot of the zvol, compressed with xz (slow...)
- deleted some of the files from the W10 mounted iSCSI volume
- zfs send the snapshot of the volume to a FreeNAS server
On the FeeNAS server:
- xzcat | zfs receive the stream sent from the NetBSD box (obviously
'zfs send | ssh ... zfs receive' will work as well)
- created an extent, target and associated target on the FreeNAS
server (it did not work properly when I added it as a second LUN to an
existing target, it was seen as read-only and attempt to online it
failed, stating the disk was invalid)
On the W10 iSCSI initiator:
- as I was using the same W10 initiator, I had to offline the volume
as mounted from the NetBSD iSCSI server - otherwise it complains about
the disks having the same GUIDs
- attached to the newly created target from the W10 box, onlined the
disk, found the contents before the deletion, as expected.
Basically a quick test to give me some additional confidence in the
ZFS implementation. Looking good IMHO.
>
> --
> --
> Michael van Elst
> Internet: mlelstv%serpens.de@localhost
> "A potential Snark may lurk in every tree."
And another hit, just a moment ago.
On OmniOS I can:
...
root@omni2:/export/home/xci
# zfs create -V 4G data/testv
20-02-26 - 21:58:59
root@omni2:/export/home/xci
# zpool create asd /dev/zvol/dsk/data/testv
20-02-26 - 21:59:04
root@omni2:/export/home/xci
# zpool list
20-02-26 - 21:59:36
NAME SIZE ALLOC FREE CKPOINT EXPANDSZ FRAG CAP DEDUP
HEALTH ALTROOT
asd 3.75G 648K 3.75G - - 0% 0% 1.00x ONLINE -
data 15.9G 85.1M 15.8G - - 0% 0% 1.00x ONLINE -
rpool 31.8G 3.28G 28.5G - - 5% 10% 1.00x ONLINE -
....
So I can create a pool backed by a zvol; I thought this was not
allowed, bit seems to work there.
Under NetBSD-current, I got a panic doing the same:
...
(gdb) target kvm netbsd.6.core
0xffffffff80224225 in cpu_reboot ()
(gdb) bt
#0 0xffffffff80224225 in cpu_reboot ()
#1 0xffffffff809fdbff in kern_reboot ()
#2 0xffffffff80a3fff9 in vpanic ()
#3 0xffffffff80a400bd in panic ()
#4 0xffffffff80a36cc4 in lockdebug_abort ()
#5 0xffffffff809f2367 in mutex_vector_enter ()
#6 0xffffffff81efda2c in zfsdev_close ()
#7 0xffffffff80abd94c in spec_close ()
#8 0xffffffff80ab0808 in VOP_CLOSE ()
#9 0xffffffff80aa8119 in vn_close ()
#10 0xffffffff81ee55f6 in vdev_disk_close ()
#11 0xffff92870dd15688 in ?? ()
#12 0x0000000000000004 in ?? ()
#13 0xffff9981951269f0 in ?? ()
#14 0xffffffff81ee2f6d in vdev_set_state ()
Backtrace stopped: frame did not save the PC
....
Chavdar
--
----
Home |
Main Index |
Thread Index |
Old Index