Skip to main content

Home lab - arguments for a 'production' environment

As I've used my lab over the last year or so, I've come to a pretty strong conviction:  If you are simply running 'test' VMs in a throwaway environment you will lose out on some key "free" 'production' experience.  A lot of quotes, but let me try and explain.

Suppose we have a test lab with a few ESXi hosts, vCenter, and whatever flavour OS environment you like.  You have an IIS/Apache test site, vCenter, SQL/mysql cluster, AD domain, vcb, that sort of stuff.   All well and good.  One day something in the ESXi/vCenter infrastructure breaks and you get frustrated trying to solve the problem.  No big deal, wipe and replace, right?  Restore from backup with no thoughts.

Now, suppose we have the same test lab, but you are running your own website with SQL back-end, monitoring software for both your internal lab environment and a few of your clients' sites, your mail comes in to an Exchange box, your AD site is used by the home PCs for authentication, etc.  One day something in the infrastructure breaks.  Can't just wipe and replace!  Have to figure it out, and fast!

It's the latter scenario that will help hone your 'under pressure' troubleshooting skills without getting you fired (you were wise enough to keep your wife's things somewhere safe, right?), and change your outlook on troubleshooting in the workplace.  As others have noted, admins these days find it too easy to just say 'blow it away and start fresh'.  It's not easy to actually figure out what the root cause is, but sometimes for legal/financial reasons you must figure it out.  So, the faster you can do this, the better!  This can vary in different environments, but larger companies may be upset when you can't provide a solid reason why something broke.

All this to say, when you come home from work and one of your hosts is acting up, your cluster is broken, while it can be a stumbling block, you'll usually come out ahead with a few more tricks under your belt, and a little experience goes a long way.

P.S.  If you were wondering, my lab is currently undergoing issues, and I run things like the PTC wiki off it, so taking it down I consider 'downtime' to be avoided.


Popular posts from this blog

DFSR - eventid 4312 - replication just won't work

This warning isn't documented that well on the googles, so here's some google fodder:

You are trying to set up replication for a DFS folder (no existing replication)Source server is 2008R2, 'branch office' server is 2012R2 (I'm moving all our infra to 2012R2)You have no issues getting replication configuredYou see the DFSR folders get created on the other end, but nothing stagesFinally you get EventID 4312:
The DFS Replication service failed to get folder information when walking the file system on a journal wrap or loss recovery due to repeated sharing violations encountered on a folder. The service cannot replicate the folder and files in that folder until the sharing violation is resolved.  Additional Information:  Folder: F:\Users$\\Desktop\Random Folder Name\  Replicated Folder Root: F:\Users$  File ID: {00000000-0000-0000-0000-000000000000}-v0  Replicated Folder Name: Users  Replicated Folder ID: 33F0449D-5E67-4DA1-99AC-681B5BACC7E5  Replication Group…

Fixing duplicate SPNs (service principal name)

This is a pretty handy thing to know:

SPNs are used when a specific service/daemon uses Kerberos to authenticate against AD. They map a specific service, port, and object together with this convention: class/host:port/name

If you use a computer object to auth (such as local service):

If you use a user object to auth (such as a service account, or admin account):

Why do we care about duplicate SPNs? If you have two entries trying to auth using the same Kerberos ticket (I think that's right...), they will conflict, and cause errors and service failures.

To check for duplicate SPNs:
The command "setspn.exe -X

C:\Windows\system32>setspn -X
Processing entry 7
MSSQLSvc/ is registered on these accounts:
CN=SQL Admin,OU=service accounts,OU=resources,DC=company,DC=local

found 1 groups of duplicate SPNs. (truncated/sanitized)

Note that y…

Logstash to Nagios - alerting based on Windows Event ID

This took way longer than it should have to get here's a config and brain dump...

You want to have a central place to analyze Windows Event/IIS/local application logs, alert off specific events, alert off specific situations.  You don't have the budget for a boxed solution.  You want pretty graphs.  You don't particularly care about individual server states.  (see rationale below - although you certainly have all the tools here to care, I haven't provided that configuration)

ELK stack, OMD, NXlog agent, and Rsyslog.  The premise here is as follows:

Event generated on server into EventLogNXlog ships to Logstash inputLogstash filter adds fields and tags to specified eventsLogstash output sends to a passive Nagios service via the Nagios NSCA outputThe passive service on Nagios (Check_MK c/o OMD) does its thing w. alerting
Open Monitoring Distribution, but the real point here is Check_MK (IIRC Icinga uses this...).  It makes Nagios easy to use and main…