Skip to main content

Exmerge and NT

So, lesson for y'all!

When you are running exmerge on an NT box with Exchange 5.5, here's how it works:

You give it the list of boxes you want out, and it then exports each box to a PST file wherever you want it to go.

The hidden catch is: It removes the data from the Exchange store, and creates a PST file! So say you just want to test the utility, but not do anything, while it's functioning users are losing Outlook data left right and centre! Crazy fun times.

Also, when you try to exmerge a user's mailbox that is larger than 2GB, exmerge crashes. If you restart exmerge, everything you took out prior to that will be overwritten (since you say 'sure, why not start from scratch'), but since you didn't know that exmerge removes the data from the Exchange store, you are effectively deleting all their emails. It's awesome.

So then, after you've tried restarting the exmerge process a few times (deleting all users' emails earlier in line than the 2GB+ user), you give up and move on with a new exmerge starting after that user.

This is fine, except then you find out all the users are losing their mailbox content and decide to reverse the process, and discover the above. Fancy!

Anyways, in summary:

- Exmerge for 5.5 does not allow for duplicate copies of the mailbox, so deletes the source once exported.
- You cannot have a PST file larger than 2GB, even when exmerging.
- Users will continue to send/receive email even though their mailbox has been effectively deleted from the store, making your restore process much more complicated than it needed to be two days out from a major migration.

To note, the 2003 exmerge creates copies when it exports to PST, hence the confusion when this happened.

Comments

Popular posts from this blog

DFSR - eventid 4312 - replication just won't work

This warning isn't documented that well on the googles, so here's some google fodder:


You are trying to set up replication for a DFS folder (no existing replication)Source server is 2008R2, 'branch office' server is 2012R2 (I'm moving all our infra to 2012R2)You have no issues getting replication configuredYou see the DFSR folders get created on the other end, but nothing stagesFinally you get EventID 4312:
The DFS Replication service failed to get folder information when walking the file system on a journal wrap or loss recovery due to repeated sharing violations encountered on a folder. The service cannot replicate the folder and files in that folder until the sharing violation is resolved.  Additional Information:  Folder: F:\Users$\user.name\Desktop\Random Folder Name\  Replicated Folder Root: F:\Users$  File ID: {00000000-0000-0000-0000-000000000000}-v0  Replicated Folder Name: Users  Replicated Folder ID: 33F0449D-5E67-4DA1-99AC-681B5BACC7E5  Replication Group…

Fixing duplicate SPNs (service principal name)

This is a pretty handy thing to know:

SPNs are used when a specific service/daemon uses Kerberos to authenticate against AD. They map a specific service, port, and object together with this convention: class/host:port/name

If you use a computer object to auth (such as local service):
MSSQLSVC/tor-sql-01.domain.local:1433

If you use a user object to auth (such as a service account, or admin account):
MSSQLSVC/username:1433

Why do we care about duplicate SPNs? If you have two entries trying to auth using the same Kerberos ticket (I think that's right...), they will conflict, and cause errors and service failures.

To check for duplicate SPNs:
The command "setspn.exe -X

C:\Windows\system32>setspn -X
Processing entry 7
MSSQLSvc/server1.company.local:1433 is registered on these accounts:
CN=SERVER1,OU=servers,OU=resources,DC=company,DC=local
CN=SQL Admin,OU=service accounts,OU=resources,DC=company,DC=local

found 1 groups of duplicate SPNs. (truncated/sanitized)

Note that y…

Logstash to Nagios - alerting based on Windows Event ID

This took way longer than it should have to get going...so here's a config and brain dump...

Why?
You want to have a central place to analyze Windows Event/IIS/local application logs, alert off specific events, alert off specific situations.  You don't have the budget for a boxed solution.  You want pretty graphs.  You don't particularly care about individual server states.  (see rationale below - although you certainly have all the tools here to care, I haven't provided that configuration)

How?
ELK stack, OMD, NXlog agent, and Rsyslog.  The premise here is as follows:

Event generated on server into EventLogNXlog ships to Logstash inputLogstash filter adds fields and tags to specified eventsLogstash output sends to a passive Nagios service via the Nagios NSCA outputThe passive service on Nagios (Check_MK c/o OMD) does its thing w. alerting
OMD
Open Monitoring Distribution, but the real point here is Check_MK (IIRC Icinga uses this...).  It makes Nagios easy to use and main…