Re: Re-exporting RBD images via iSCSI

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

i have done a short look into RBD + iSCSI, and i found TGT + librbd.

https://github.com/fujita/tgt
http://stgt.sourceforge.net/

i didn't take a deeper look into it but i like to test it in the next month or so, it looks easy to me https://github.com/fujita/tgt/blob/master/doc/README.rbd

cheers
Ansgar



2013/3/16 Bond, Darryl <dbond@xxxxxxxxxxxxx>
I have a small 3 node ceph cluster with 6 OSDs on each node
I would like to re-export some rbd images via LIO.
Is it recommended to run RBD/LIO on one of the cluster nodes?

Preliminary tests show that it works fine. I have seen reports (that I can't find) that it is not recommended to run the RBD kernel module on an OSD node.

Has anyone used multiple hosts to do iSCSI multipathing to a singe RBD image for vmware?
My thoughts are to export the same RBD image via LIO from 2 hosts. It is easy to configure LIO to use the same iSCSI target address on both hosts.

I could then configure vmware storage with the two ceph nodes as a primary/secondary failover.

Regards
Darryl


The contents of this electronic message and any attachments are intended only for the addressee and may contain legally privileged, personal, sensitive or confidential information. If you are not the intended addressee, and have received this email, any transmission, distribution, downloading, printing or photocopying of the contents of this message or attachments is strictly prohibited. Any legal privilege or confidentiality attached to this message and attachments is not waived, lost or destroyed by reason of delivery to any person other than intended addressee. If you have received this message and are not the intended addressee you should notify the sender by return email and destroy all copies of the message and any attachments. Unless expressly attributed, the views expressed in this email do not necessarily represent the views of the company.
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux