iSCSI don't start automatically after all servers reboot (using 1.3.1 version)
clsaad
8 Posts
August 15, 2017, 6:23 pmQuote from clsaad on August 15, 2017, 6:23 pmHello,
I have now a LAB with 3 physical servers and all have 4 discs with 250GB SATA.
I have created the LAB with 2 iSCSI connections and after the 3 servers restart the iSCSI don't start automatically.
I try to find any information about iSCSI start service in node log and don't find anything. If I try to start manually works fine.
Bellow the node log from ps-node-02
Hello,
I have now a LAB with 3 physical servers and all have 4 discs with 250GB SATA.
I have created the LAB with 2 iSCSI connections and after the 3 servers restart the iSCSI don't start automatically.
I try to find any information about iSCSI start service in node log and don't find anything. If I try to start manually works fine.
Bellow the node log from ps-node-02
admin
2,930 Posts
August 16, 2017, 2:38 pmQuote from admin on August 16, 2017, 2:38 pmThanks for your feedback. It is quite important for us to know how to improve PetaSAN.
Currently in PetaSAN we support nodes going down and up and we automatically handle the distribution of iSCSI load, however this assumes that there is running cluster. We do not auto start iSCSI disks if the entire cluster was down, the admin must manually do this.
In a traditional 2 node active/passive storage system it is easy to support this, a node can operate on its local storage and independent on the backup. In PetaSAN storage is provided by Ceph cluster and it must be up, in addition paths are not owned by the nodes but distributed via the Consul service platform which again needs to be up for things to work. Moreover even if both Ceph and Consul are up we do not want a node taking all the iSCSI load while many other iSCSI server nodes are being started. Most of the technology used in PetaSAN is geared for cloud usage and is not designed to deal with regular restart of the entire system without a user doing so.
But what you mention is definitely something we will try to make it easier from the user point of view.
Thanks for your feedback. It is quite important for us to know how to improve PetaSAN.
Currently in PetaSAN we support nodes going down and up and we automatically handle the distribution of iSCSI load, however this assumes that there is running cluster. We do not auto start iSCSI disks if the entire cluster was down, the admin must manually do this.
In a traditional 2 node active/passive storage system it is easy to support this, a node can operate on its local storage and independent on the backup. In PetaSAN storage is provided by Ceph cluster and it must be up, in addition paths are not owned by the nodes but distributed via the Consul service platform which again needs to be up for things to work. Moreover even if both Ceph and Consul are up we do not want a node taking all the iSCSI load while many other iSCSI server nodes are being started. Most of the technology used in PetaSAN is geared for cloud usage and is not designed to deal with regular restart of the entire system without a user doing so.
But what you mention is definitely something we will try to make it easier from the user point of view.
Last edited on August 16, 2017, 2:41 pm by admin · #2
clsaad
8 Posts
August 21, 2017, 6:38 pmQuote from clsaad on August 21, 2017, 6:38 pmHello,
Thanks for the explanation. I guess that I have a BUG. After shutdown all servers and start all servers and wait the replication, when I try to start the iSCSI connection sound that OK (started) bug none of the 2 IP's responds. I try to stop and restart the iSCSI configuration (for only 1 mapped disk). I found this error in the Log in the node:
BTW, I'm using the latest version (1.4). The upgrade process works very fine 😉
Hello,
Thanks for the explanation. I guess that I have a BUG. After shutdown all servers and start all servers and wait the replication, when I try to start the iSCSI connection sound that OK (started) bug none of the 2 IP's responds. I try to stop and restart the iSCSI configuration (for only 1 mapped disk). I found this error in the Log in the node:
BTW, I'm using the latest version (1.4). The upgrade process works very fine 😉
admin
2,930 Posts
August 22, 2017, 1:08 pmQuote from admin on August 22, 2017, 1:08 pmHi there, and thanks you like the upgrade process.
Re the iSCSI disk not starting, we do not see this problem. we also try to follow your case: shutdown 1.3.1 cluster, restart, upgrade and all worked. Can you give us info how to re-produce the issue. The error listed shows that the kernel cannot map the rbd image: have you upgraded the system outside PetaSAN or created/modified the rbd image manually ? Again any more info to reproduce the problem to help us look into it.
Hi there, and thanks you like the upgrade process.
Re the iSCSI disk not starting, we do not see this problem. we also try to follow your case: shutdown 1.3.1 cluster, restart, upgrade and all worked. Can you give us info how to re-produce the issue. The error listed shows that the kernel cannot map the rbd image: have you upgraded the system outside PetaSAN or created/modified the rbd image manually ? Again any more info to reproduce the problem to help us look into it.
clsaad
8 Posts
August 23, 2017, 2:29 pmQuote from clsaad on August 23, 2017, 2:29 pmHello,
I used the auto update proccess (using USB boot -> identity the old version -> make update -> reboot).
After that initiate the problem. I'm try to search more details about the problem and return back to you.
Hello,
I used the auto update proccess (using USB boot -> identity the old version -> make update -> reboot).
After that initiate the problem. I'm try to search more details about the problem and return back to you.
admin
2,930 Posts
August 24, 2017, 12:09 pmQuote from admin on August 24, 2017, 12:09 pm
Yes these are the steps we followed but it is working here ...so additional info will help:
- Are all your disks not starting or only disk 00003 ?
- Could you start/stop this disk in ver 1.3.1 before you upgraded to 1.4 ?
- Have you used any cli commands either to add/upgrade packages or modified this disk outside PetaSAN ?
- What is the output of
rbd info image-00003 --cluster CLUSTER_NAME
- Also if this is not a production environment, try to map the disk manually but first stop the PetaSAN iscsi service, otherwise it will be unmapped automatically:
systemctl stop petasan-iscsi
rbd map image-00003 --cluster CLUSTER_NAME
rbd showmapped --cluster CLUSTER_NAME
Is the disk mapped successfully ?
when done
rbd unmap image-00003 --cluster CLUSTER_NAME
systemctl start petasan-iscsi
Thanks very much for your help.
Yes these are the steps we followed but it is working here ...so additional info will help:
- Are all your disks not starting or only disk 00003 ?
- Could you start/stop this disk in ver 1.3.1 before you upgraded to 1.4 ?
- Have you used any cli commands either to add/upgrade packages or modified this disk outside PetaSAN ?
- What is the output of
rbd info image-00003 --cluster CLUSTER_NAME
- Also if this is not a production environment, try to map the disk manually but first stop the PetaSAN iscsi service, otherwise it will be unmapped automatically:
systemctl stop petasan-iscsi
rbd map image-00003 --cluster CLUSTER_NAME
rbd showmapped --cluster CLUSTER_NAME
Is the disk mapped successfully ?
when done
rbd unmap image-00003 --cluster CLUSTER_NAME
systemctl start petasan-iscsi
Thanks very much for your help.
Last edited on August 24, 2017, 12:12 pm by admin · #6
iSCSI don't start automatically after all servers reboot (using 1.3.1 version)
clsaad
8 Posts
Quote from clsaad on August 15, 2017, 6:23 pmHello,
I have now a LAB with 3 physical servers and all have 4 discs with 250GB SATA.
I have created the LAB with 2 iSCSI connections and after the 3 servers restart the iSCSI don't start automatically.
I try to find any information about iSCSI start service in node log and don't find anything. If I try to start manually works fine.
Bellow the node log from ps-node-02
Hello,
I have now a LAB with 3 physical servers and all have 4 discs with 250GB SATA.
I have created the LAB with 2 iSCSI connections and after the 3 servers restart the iSCSI don't start automatically.
I try to find any information about iSCSI start service in node log and don't find anything. If I try to start manually works fine.
Bellow the node log from ps-node-02
admin
2,930 Posts
Quote from admin on August 16, 2017, 2:38 pmThanks for your feedback. It is quite important for us to know how to improve PetaSAN.
Currently in PetaSAN we support nodes going down and up and we automatically handle the distribution of iSCSI load, however this assumes that there is running cluster. We do not auto start iSCSI disks if the entire cluster was down, the admin must manually do this.
In a traditional 2 node active/passive storage system it is easy to support this, a node can operate on its local storage and independent on the backup. In PetaSAN storage is provided by Ceph cluster and it must be up, in addition paths are not owned by the nodes but distributed via the Consul service platform which again needs to be up for things to work. Moreover even if both Ceph and Consul are up we do not want a node taking all the iSCSI load while many other iSCSI server nodes are being started. Most of the technology used in PetaSAN is geared for cloud usage and is not designed to deal with regular restart of the entire system without a user doing so.
But what you mention is definitely something we will try to make it easier from the user point of view.
Thanks for your feedback. It is quite important for us to know how to improve PetaSAN.
Currently in PetaSAN we support nodes going down and up and we automatically handle the distribution of iSCSI load, however this assumes that there is running cluster. We do not auto start iSCSI disks if the entire cluster was down, the admin must manually do this.
In a traditional 2 node active/passive storage system it is easy to support this, a node can operate on its local storage and independent on the backup. In PetaSAN storage is provided by Ceph cluster and it must be up, in addition paths are not owned by the nodes but distributed via the Consul service platform which again needs to be up for things to work. Moreover even if both Ceph and Consul are up we do not want a node taking all the iSCSI load while many other iSCSI server nodes are being started. Most of the technology used in PetaSAN is geared for cloud usage and is not designed to deal with regular restart of the entire system without a user doing so.
But what you mention is definitely something we will try to make it easier from the user point of view.
clsaad
8 Posts
Quote from clsaad on August 21, 2017, 6:38 pmHello,
Thanks for the explanation. I guess that I have a BUG. After shutdown all servers and start all servers and wait the replication, when I try to start the iSCSI connection sound that OK (started) bug none of the 2 IP's responds. I try to stop and restart the iSCSI configuration (for only 1 mapped disk). I found this error in the Log in the node:
BTW, I'm using the latest version (1.4). The upgrade process works very fine 😉
Hello,
Thanks for the explanation. I guess that I have a BUG. After shutdown all servers and start all servers and wait the replication, when I try to start the iSCSI connection sound that OK (started) bug none of the 2 IP's responds. I try to stop and restart the iSCSI configuration (for only 1 mapped disk). I found this error in the Log in the node:
BTW, I'm using the latest version (1.4). The upgrade process works very fine 😉
admin
2,930 Posts
Quote from admin on August 22, 2017, 1:08 pmHi there, and thanks you like the upgrade process.
Re the iSCSI disk not starting, we do not see this problem. we also try to follow your case: shutdown 1.3.1 cluster, restart, upgrade and all worked. Can you give us info how to re-produce the issue. The error listed shows that the kernel cannot map the rbd image: have you upgraded the system outside PetaSAN or created/modified the rbd image manually ? Again any more info to reproduce the problem to help us look into it.
Hi there, and thanks you like the upgrade process.
Re the iSCSI disk not starting, we do not see this problem. we also try to follow your case: shutdown 1.3.1 cluster, restart, upgrade and all worked. Can you give us info how to re-produce the issue. The error listed shows that the kernel cannot map the rbd image: have you upgraded the system outside PetaSAN or created/modified the rbd image manually ? Again any more info to reproduce the problem to help us look into it.
clsaad
8 Posts
Quote from clsaad on August 23, 2017, 2:29 pmHello,
I used the auto update proccess (using USB boot -> identity the old version -> make update -> reboot).
After that initiate the problem. I'm try to search more details about the problem and return back to you.
Hello,
I used the auto update proccess (using USB boot -> identity the old version -> make update -> reboot).
After that initiate the problem. I'm try to search more details about the problem and return back to you.
admin
2,930 Posts
Quote from admin on August 24, 2017, 12:09 pm
Yes these are the steps we followed but it is working here ...so additional info will help:
- Are all your disks not starting or only disk 00003 ?
- Could you start/stop this disk in ver 1.3.1 before you upgraded to 1.4 ?
- Have you used any cli commands either to add/upgrade packages or modified this disk outside PetaSAN ?
- What is the output of
rbd info image-00003 --cluster CLUSTER_NAME
- Also if this is not a production environment, try to map the disk manually but first stop the PetaSAN iscsi service, otherwise it will be unmapped automatically:
systemctl stop petasan-iscsi
rbd map image-00003 --cluster CLUSTER_NAME
rbd showmapped --cluster CLUSTER_NAME
Is the disk mapped successfully ?
when done
rbd unmap image-00003 --cluster CLUSTER_NAME
systemctl start petasan-iscsi
Thanks very much for your help.
Yes these are the steps we followed but it is working here ...so additional info will help:
- Are all your disks not starting or only disk 00003 ?
- Could you start/stop this disk in ver 1.3.1 before you upgraded to 1.4 ?
- Have you used any cli commands either to add/upgrade packages or modified this disk outside PetaSAN ?
- What is the output of
rbd info image-00003 --cluster CLUSTER_NAME
- Also if this is not a production environment, try to map the disk manually but first stop the PetaSAN iscsi service, otherwise it will be unmapped automatically:
systemctl stop petasan-iscsi
rbd map image-00003 --cluster CLUSTER_NAME
rbd showmapped --cluster CLUSTER_NAME
Is the disk mapped successfully ?
when done
rbd unmap image-00003 --cluster CLUSTER_NAME
systemctl start petasan-iscsi
Thanks very much for your help.