[ale] ISCSI array on virtual machine
DJ-Pfulio
DJPfulio at jdpfu.com
Thu Apr 28 09:12:35 EDT 2016
Without LVM, getting a consistent backup from storage with running
processes and open files (like DBMS) is harder. Snapshots. For
document-based storage, snapshots aren't that important.
Then there is the speed of LVM - creating a new ext4 file system on an
LV is almost instantaneous, almost. The days of waiting 20 sec or 10
min or 4 hours are over with LVM.
There are lots of reasons to use LVM beyond the base grow/shrink
capabilities.
On 04/28/16 08:46, Lightner, Jeff wrote:
> 1) I’m a RHEL guy and I don’t use xfs for most purposes but rather
> continue to use ext4 because it suits our needs. We do allow the base
> VolGroup00 LVMs we use for base filesystems to do xfs on RHEL7 simply
> because that is the default and we seldom need to change those
> filesystems. For apps and dbs though we use ext4.
>
> 2) We use LVM for everything. The benefit to using LVM is you can
> increase or decrease LVs on the fly without having to worry about other
> LVs in the same VG. With partitioning you have to adjust existing
> partitions to add new ones. For a single partition with root on it
> this would be a concern if you later decided you wanted to decrease that
> to add another partition. For multiple partitions it would be even
> more of a concern because you might have to adjust multiple partitions.
>
> 3) We insulate filesystems by creating multiple LVs. Typically we
> have /var, /tmp, /usr, /home and /opt at a minimum as separate
> filesystems in VolGroup00. This helps to insure / itself doesn’t get
> filled up by errant processes as well as insuring the other filesystems
> are unique to their purposes. If we add applications or databases that
> require significant space we usually put them on their own
> LVs/filesystems as well.
>
>
>
> We use mostly SAN storage for everything except VolGroup00.
>
>
>
> I was confused by the original post as it was talking about iSCSI then
> VMWare. If the iSCSI attachment is to the VMWare hypervisor doesn’t it
> just present storage to the virtual guests which is seen as if it were
> local disks there? If so you wouldn’t need to do any RAID at guest
> level as I’d expect it would have been presented to the VMWare
> hypervisor in a RAID configuration so the space it allocates to guests
> are already protected by that underlying RAID. Here we use MS Hyper-V
> for the most part and that is how it works on those hypervisors but I’d
> think it was the same from any kind of hypervisor.
>
>
>
> *From:*ale-bounces at ale.org [mailto:ale-bounces at ale.org] *On Behalf Of
> *Jim Kinney
> *Sent:* Thursday, April 28, 2016 7:17 AM
> *To:* Atlanta Linux Enthusiasts - Yes! We run Linux!
> *Subject:* Re: [ale] ISCSI array on virtual machine
>
>
>
> I have a large drive array for my department. I use LVM to carve it up.
> I leave a huge chunk unallocated so I can extend logical partitions as
> required. That dodges the need to shrink existing partitions and allows
> XFS as filesystem.
>
> On Apr 28, 2016 3:27 AM, "Todor Fassl" <fassl.tod at gmail.com
> <mailto:fassl.tod at gmail.com>> wrote:
>
> With respect to your question about using LVM ... I guess that was sort
> of my original question. If I just allocate the whole 8T to one big
> partition, I'd have no reason to use LVM. But I can see the need to use
> LVM if I continue with the scheme where I split the drive into
> partitions for faculty, grads, and staff.
>
> On 04/27/2016 02:27 PM, Jim Kinney wrote:
>
> If you need de-dup, ZFS is the only choice and be ready to throw a lot
> of RAM into the server so it can do it's job. I was looking at dedupe
> on 80TB and the RAM hit was 250GB.
> XFS vs EXT4.
> XFS is the better choice.
> XFS does everything EXT4 does except shrink. It was designed for (then
> very) large files (video) and works quite well with smaller files. It's
> as fast as EXT4 but will handle larger files and many, many more of
> them. I want to say exabytes but not certain. Petabytes are OK
> filesystem sizes with XFS right now. I have no experience with a
> filesystem of that size but I expect there to be some level of metadata
> performance hit.
> If there's the slightest chance of a need to shrink a partition (You
> _are_ using LVM, right?) then XFS will bite you and require relocation,
> tear down, rebuild, relocation. Not a fun process.
> A while back, an install onto a 24 TB RAID6 array refused to budge
> using EXT4. While EXT4 is supposed to address that kind of size, it had
> bugs and unimplemented plans for expansion features that were blockers.
> I used XFS instead and never looked back. XFS has a very complete
> toolset for maintenance/repair needs.
> On Wed, 2016-04-27 at 13:54 -0500, Todor Fassl wrote:
>
> I need to setup a new file server on a virtual machine with an
> attached
> ISCSI array. Two things I am obsessing over -- 1. Which file system
> to
> use and 2. Partitioning scheme.
>
> The ISCSI array is attached to a ubuntu 16.04 virtual machine. To
> tell
> you the truth, I don't even know how that is done. I do not manage
> the
> VMware cluster. In fact, I think the Dell technitian actually ddid
> that
> for us. It looks like a normal 8T hard drive on /dev/sdb to the
> virtual
> machine. The ISCSI array is configured for RAID6 so from what I
> understand, all I have to do is choose a file system appropriate for
> my
> end user's needs. Even though the array looks like a single hard
> drive,
> I don't have to worry about software RAID or anyhthing like that.
>
> Googling shows me no clear advantage to ext4, xfs, or zfs. I haven't
> been able to find a page that says any one of those is an obvious
> choice
> in my situation. I have about 150 end-users with nfs mounted home
> directories. We also have a handful of people using Windows so the
> file
> server will have samba installed. It's a pretty good mix of large
> files
> and small files since different users are doing drastically
> different
> things. There are users who never do anything but read email and
> browse
> the web and others doing fluid dynamic simulations on small
> supercomputers.
>
> Secondthing I've been going back and forth on in my own mind is
> whether
> to do away with seperate partitions for faculty, staff, and grad
> students. My co-worker says that's probably an artifact of the days
> when
> partition sizes were limited. That was before my time here. The last
> 2
> times we rebuilt our file server, we just maintained the
> partitioning
> scheme and just made the sizes times larger. But sometimes the
> faculty
> partition got filled up while there was still plenty of space left
> on
> the grad partition. Or it might be the other way around. If we
> munged
> them all together, that wouldn't happen. The only downside I see to
> doing that is that if the faculty partition gets hosed, the grad
> partition wouldn't be effected. But that seems like a pretty
> arbitrary
> choice. We could just assign users randomly to one partition or
> another.
> When you're setting up a NAS for use by a lot of users, is it
> considered
> best practice to split it up to limit the damage from a messed up
> file
> system? I mean, hopefully, that never happens anyway, right?
>
> Right now, I've got it configured as one gigantic 8T ext4 partition.
> But
> we won't be going live with it until the end of May so I have plenty
> of
> time to completely rebuild it.
>
>
>
> _______________________________________________
> Ale mailing list
> Ale at ale.org <mailto:Ale at ale.org>
> http://mail.ale.org/mailman/listinfo/ale
> See JOBS, ANNOUNCE and SCHOOLS lists at
> http://mail.ale.org/mailman/listinfo
>
>
> --
> Todd
> _______________________________________________
> Ale mailing list
> Ale at ale.org <mailto:Ale at ale.org>
> http://mail.ale.org/mailman/listinfo/ale
> See JOBS, ANNOUNCE and SCHOOLS lists at
> http://mail.ale.org/mailman/listinfo
>
>
>
> _______________________________________________
> Ale mailing list
> Ale at ale.org
> http://mail.ale.org/mailman/listinfo/ale
> See JOBS, ANNOUNCE and SCHOOLS lists at
> http://mail.ale.org/mailman/listinfo
>
--
Got Linux? Used on smartphones, tablets, desktop computers, media
centers, and servers by kids, Moms, Dads, grandparents and IT
professionals.
More information about the Ale
mailing list