Skip to main content

Lab - SAN options

Update: Going to wait and see what the next few months brings for new products. Hesitant to purchase the Iomega units as they are several years old now.

I had a sobering thought this afternoon. My entire VMware lab is basing iSCSI SAN usage off of our Synology DS410j. The same NAS that houses all of our pictures, data, etc. If my lab kills it, my wife kills me. (and frankly I'd be super-upset as well) We do have backups (that I need to test) done to an external, but it would be a giant pain if it broke.

Thus, I consider my options.

Dedicated Openfiler box (~$1800)
This would be a stand alone machine running the latest version of Openfiler. Dual RAID cards with 8 spindles each, two dual NICs for MPIO to two locations, thus mimicking two SANs.

Pro: Does MPIO, high spindle count.
Con: Cannot test SAN replication. Power hippo.

Two NAS that support iSCSI (~$2200)
Basically buy two DS411+ and fill with 1TB or 2TB drives.

Pro: Simple to configure/maintain. Can do synchronization between units.
Con: Only four spindles. Only one NIC (no MPIO). Relatively high cost per unit ($690 per bare chassis).

ESX host running Openfiler VM (~$2200)
Basically an ESX host with a LOT of local storage (dual RAID arrays, 16 spindles).

Pro: VM appliance. Able to run other VMs on same machine.
Con: More complex to maintain. Less ports usable for ESX purposes as they are tied to Openfiler VM. More layers so more latency.

All in all, pricy, and none of them really give me what I'm looking for.

However, there is one more option that I've been avoiding:

NAS Option #2 (~$1300)
IOmega IX4-200d (4x1TB). A lot of the VM bloggers use these, so maybe it's wise to just jump on board. VMware HCL is no small thing either. I would have liked some sort of MPIO option. According to the manual, it does a limited version of MPIO, so maybe this is the best way.

Pro: On the VMware HCL. Dual NICs. 4x1TB is plenty of space.
Con: Bad support (many people not happy). Cannot buy diskless.


Popular posts from this blog

DFSR - eventid 4312 - replication just won't work

This warning isn't documented that well on the googles, so here's some google fodder:

You are trying to set up replication for a DFS folder (no existing replication)Source server is 2008R2, 'branch office' server is 2012R2 (I'm moving all our infra to 2012R2)You have no issues getting replication configuredYou see the DFSR folders get created on the other end, but nothing stagesFinally you get EventID 4312:
The DFS Replication service failed to get folder information when walking the file system on a journal wrap or loss recovery due to repeated sharing violations encountered on a folder. The service cannot replicate the folder and files in that folder until the sharing violation is resolved.  Additional Information:  Folder: F:\Users$\\Desktop\Random Folder Name\  Replicated Folder Root: F:\Users$  File ID: {00000000-0000-0000-0000-000000000000}-v0  Replicated Folder Name: Users  Replicated Folder ID: 33F0449D-5E67-4DA1-99AC-681B5BACC7E5  Replication Group…

Fixing duplicate SPNs (service principal name)

This is a pretty handy thing to know:

SPNs are used when a specific service/daemon uses Kerberos to authenticate against AD. They map a specific service, port, and object together with this convention: class/host:port/name

If you use a computer object to auth (such as local service):

If you use a user object to auth (such as a service account, or admin account):

Why do we care about duplicate SPNs? If you have two entries trying to auth using the same Kerberos ticket (I think that's right...), they will conflict, and cause errors and service failures.

To check for duplicate SPNs:
The command "setspn.exe -X

C:\Windows\system32>setspn -X
Processing entry 7
MSSQLSvc/ is registered on these accounts:
CN=SQL Admin,OU=service accounts,OU=resources,DC=company,DC=local

found 1 groups of duplicate SPNs. (truncated/sanitized)

Note that y…

Logstash to Nagios - alerting based on Windows Event ID

This took way longer than it should have to get here's a config and brain dump...

You want to have a central place to analyze Windows Event/IIS/local application logs, alert off specific events, alert off specific situations.  You don't have the budget for a boxed solution.  You want pretty graphs.  You don't particularly care about individual server states.  (see rationale below - although you certainly have all the tools here to care, I haven't provided that configuration)

ELK stack, OMD, NXlog agent, and Rsyslog.  The premise here is as follows:

Event generated on server into EventLogNXlog ships to Logstash inputLogstash filter adds fields and tags to specified eventsLogstash output sends to a passive Nagios service via the Nagios NSCA outputThe passive service on Nagios (Check_MK c/o OMD) does its thing w. alerting
Open Monitoring Distribution, but the real point here is Check_MK (IIRC Icinga uses this...).  It makes Nagios easy to use and main…