ML
    • Recent
    • Categories
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    Xen Server 6.5 + Xen Orchestra w. HA & SAN

    IT Discussion
    8
    112
    33.3k
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • ntoxicatorN
      ntoxicator
      last edited by

      @scottalanmiller said:

      Thanks scott. makes sense and I understand.

      So again, just have to spec out a 2U server (I assume 2U). With the required disk space which would hold us out for 5+ years. I am going to say we would well over 5TB+ to be safe.

      Could use 600GB or larger SAS drives with hardware raid controller. Or some enterprise level 7200RPM drives? I'm unsure how folks feel about those.

      DashrenderD scottalanmillerS 2 Replies Last reply Reply Quote 0
      • DashrenderD
        Dashrender @ntoxicator
        last edited by

        @ntoxicator said:

        @scottalanmiller said:

        Thanks scott. makes sense and I understand.

        So again, just have to spec out a 2U server (I assume 2U). With the required disk space which would hold us out for 5+ years. I am going to say we would well over 5TB+ to be safe.

        Could use 600GB or larger SAS drives with hardware raid controller. Or some enterprise level 7200RPM drives? I'm unsure how folks feel about those.

        Is there a specific reason you mention the height of a server? Do you have limitations in your cabinet that limit how tall your servers are?
        In my office I have 1 full height cabinet, 42 U worth of space. Could I have gotten away with a half cabinet, definitely, but I have what I have. I'm using 8 U for UPSs, 4U (two 2U servers) for 10 year old EHR, 4U (two 2U servers) for hypervisor hosts, 2U for a Drobo and 1U for a network switch and 1U for a KVM panel. Grand total 20U. I still have over half the rack left over for expansion.

        If I were looking at new servers, the height of the server would be the least of my concerns. Granted you can get 2U servers today that hold nearly 20 disks, it wouldn't matter to me if it was 4U because I have the space.

        The size of the drives you buy will be dependent upon a few factors. What do you need for IOPs? If you have low IOPs needs, why not buy 4 TB drives? 4 of them in RAID 10 would give you 8 TB of usable space. If you need higher IOPS, perhaps eight 2 TB drives in RAID 10 would be better, still leaving you with 8 TB usable space.

        To determine your IOPs requirement, you could get a Dell DPack run against your system. You just have to ignore the sales people trying to sell you a SAN and remember, Dell isn't trying to be your friend, they are trying to extract money from you. Ignore their SAN recommendation, and post the results to a place like ML to get help/suggestions on what to get.

        Another option would be to hire a firm to do all of this specing for you. They will run the tools and then recommend a system. This is a situation where you are paying someone for their opinion, preferably someone who isn't trying to sell you anything else. This way they understand that they are making money on their opinion/suggestion, not on the hopes of selling you hardware.

        1 Reply Last reply Reply Quote 2
        • scottalanmillerS
          scottalanmiller @ntoxicator
          last edited by

          @ntoxicator said:

          Could use 600GB or larger SAS drives with hardware raid controller. Or some enterprise level 7200RPM drives? I'm unsure how folks feel about those.

          The spindle speed is just spindle speed. If 7200 RPM drives have the IOPS that you need, they are just as good (or better since they are cheaper and more reliable.)

          1 Reply Last reply Reply Quote 1
          • ntoxicatorN
            ntoxicator
            last edited by

            Question

            HA-Lizard HA-iSCSI interface between the two(2) servers..

            bonded GigE link?

            or a single DA 10Gbe cable between the hosts? I suppose, bonded GigE is plenty sufficient due to disk write speeds? (non SSD)

            1 Reply Last reply Reply Quote 0
            • scottalanmillerS
              scottalanmiller
              last edited by

              Can't bond SAN links.

              1 Reply Last reply Reply Quote 0
              • ntoxicatorN
                ntoxicator
                last edited by

                I could be incorrect on the terminology or reference. I was reading one of their posted documents.

                This would be the DRBD interface / IP link between the two(2) nodes

                scottalanmillerS 1 Reply Last reply Reply Quote 0
                • ntoxicatorN
                  ntoxicator
                  last edited by

                  http://www.halizard.com/images/pdf/iscsi-ha_2-node_cluster_howto_1.2.12_final.pdf

                  1 Reply Last reply Reply Quote 0
                  • scottalanmillerS
                    scottalanmiller @ntoxicator
                    last edited by

                    @ntoxicator said:

                    This would be the DRBD interface / IP link between the two(2) nodes

                    DRBD is the protocol there, no iSCSI. It's not SAN or anything like that.

                    1 Reply Last reply Reply Quote 0
                    • ntoxicatorN
                      ntoxicator
                      last edited by

                      Ok. Their documentation says the DRBD interface to be bonded from within Xencenter (per documentation). Am I wrong here?

                      So the ethernet link between the two nodes, im sure GigE is plenty enough bandwidth? or 10Gige not hurt?

                      DashrenderD scottalanmillerS 3 Replies Last reply Reply Quote 0
                      • DashrenderD
                        Dashrender @ntoxicator
                        last edited by

                        @ntoxicator said:

                        Ok. Their documentation says the DRBD interface to be bonded from within Xencenter (per documentation). Am I wrong here?

                        So the ethernet link between the two nodes, im sure GigE is plenty enough bandwidth? or 10Gige not hurt?

                        Well 10 Ge never hurts...

                        1 Reply Last reply Reply Quote 0
                        • scottalanmillerS
                          scottalanmiller @ntoxicator
                          last edited by

                          @ntoxicator said:

                          Ok. Their documentation says the DRBD interface to be bonded from within Xencenter (per documentation). Am I wrong here?

                          That seems fine. DRBD works differently than iSCSI. They are not related protocols.

                          1 Reply Last reply Reply Quote 0
                          • scottalanmillerS
                            scottalanmiller @ntoxicator
                            last edited by

                            @ntoxicator said:

                            So the ethernet link between the two nodes, im sure GigE is plenty enough bandwidth? or 10Gige not hurt?

                            What is the bandwidth of the storage? You will be limited to GigE throughput speeds, that's 1Gb/s, for writes. That's a fraction of what SATA and SAS can do.

                            1 Reply Last reply Reply Quote 0
                            • scottalanmillerS
                              scottalanmiller
                              last edited by

                              Are you sure that all of this makes sense in your environment? This is two orders of magnitude from where you have been in the past. It isn't normal to have an LA (low availability) environment and get by for a long time and suddenly leap to HA. Why not just go to standard availability? It's a full order of magnitude safer than where you have been in the past, almost zero effort (and no risk from that lack of effort... simple is your friend) and less than half the price of doing HA.

                              SA is the only clear win... tons safer, tons cheaper. HA is tons safer for sure, but costs more and doesn't make sense given what was deemed acceptable in the past.

                              ntoxicatorN 1 Reply Last reply Reply Quote 1
                              • ntoxicatorN
                                ntoxicator
                                last edited by

                                Looking to develop hardware costs and quotes for new equipment. Company wants to grow employee's to 500+ by year 2020. Need to have reliable servers hosting VM's

                                If the primary xenserver host fails.. then what? We have a day + of downtime waiting for server to come back online?

                                Looking to future proof so can be in production for next 5 years. I may not be here with company in next 5 years, so want to leave behind a good setup.

                                scottalanmillerS 3 Replies Last reply Reply Quote 0
                                • ntoxicatorN
                                  ntoxicator @scottalanmiller
                                  last edited by

                                  @scottalanmiller said:

                                  Are you sure that all of this makes sense in your environment? This is two orders of magnitude from where you have been in the past. It isn't normal to have an LA (low availability) environment and get by for a long time and suddenly leap to HA. Why not just go to standard availability? It's a full order of magnitude safer than where you have been in the past, almost zero effort (and no risk from that lack of effort... simple is your friend) and less than half the price of doing HA.

                                  SA is the only clear win... tons safer, tons cheaper. HA is tons safer for sure, but costs more and doesn't make sense given what was deemed acceptable in the past.

                                  also to mention, I've been hammering HA setup for awhile to management; for peace of mind and rest-easy at night. Yes, we've been getting along with low availability type setup for now. But as the resource usages increase; i feel the need for HA setup with dual nodes.

                                  scottalanmillerS 1 Reply Last reply Reply Quote 0
                                  • scottalanmillerS
                                    scottalanmiller @ntoxicator
                                    last edited by

                                    @ntoxicator said:

                                    Looking to develop hardware costs and quotes for new equipment. Company wants to grow employee's to 500+ by year 2020. Need to have reliable servers hosting VM's

                                    That's fine. But that doesn't suggest HA at all.

                                    1 Reply Last reply Reply Quote 1
                                    • scottalanmillerS
                                      scottalanmiller @ntoxicator
                                      last edited by

                                      @ntoxicator said:

                                      If the primary xenserver host fails.. then what? We have a day + of downtime waiting for server to come back online?

                                      This is not how you discuss risk. This tells me that HA is not needed. This isn't how a "we need HA" discussion would start.

                                      1 Reply Last reply Reply Quote 0
                                      • scottalanmillerS
                                        scottalanmiller @ntoxicator
                                        last edited by

                                        @ntoxicator said:

                                        Looking to future proof so can be in production for next 5 years. I may not be here with company in next 5 years, so want to leave behind a good setup.

                                        This strongly says that HA is a bad idea as it requires a lot more skill, documentation, knowledge, etc.

                                        scottalanmillerS 1 Reply Last reply Reply Quote 0
                                        • scottalanmillerS
                                          scottalanmiller @ntoxicator
                                          last edited by

                                          @ntoxicator said:

                                          But as the resource usages increase; i feel the need for HA setup with dual nodes.

                                          Resource utilization does not lead directly to needing higher availability. Wall St. firms don't need HA on billion dollar trading systems, why do you feel that moving to 500 users warrants in for you but hundreds of thousands of users does not warrant it for them?

                                          That's not to say that HA can't be right for you, I'm saying that you aren't thinking about it in the way that you should if you were going to determine that HA was needed.

                                          1 Reply Last reply Reply Quote 0
                                          • scottalanmillerS
                                            scottalanmiller @scottalanmiller
                                            last edited by

                                            @scottalanmiller said:

                                            @ntoxicator said:

                                            Looking to future proof so can be in production for next 5 years. I may not be here with company in next 5 years, so want to leave behind a good setup.

                                            This strongly says that HA is a bad idea as it requires a lot more skill, documentation, knowledge, etc.

                                            And if HA is determined to be needed, it needs to be an appliance, not a solution that they will be unable to support.

                                            1 Reply Last reply Reply Quote 0
                                            • 1
                                            • 2
                                            • 3
                                            • 4
                                            • 5
                                            • 6
                                            • 2 / 6
                                            • First post
                                              Last post