Skip to main content

Your site is hacked - how to handle the situation properly

One of my client's sites was hacked the other day - new experience for me, so thought I'd share in a somewhat linear format.

  1. The developer who wrote the site did so using Drupal, and apparently a few weeks ago a pretty big vulnerability was found (something to do w. PHP).
  2. The site is hosted by Network Solutions (UNIX web host package) - (NS hereforth).
  3. There is no customer data held on the site, it's just informational & a few PDFs
  4. NS detected that the site was hacked, shut it down, and sent an email to the primary contact.
  5. Primary contact was someone who used to manage the account - I was never made primary (for whatever reason - obviously THIS is why having it set correctly is important)
  6. Client called saying their customers were reporting an offline site.
  7. The NS splash page simply said 'invalid domain'.
  8. I called NS after triple-checking NS account status & domain validity.
  9. NS confirmed account status was OK, transferred me to web host division - didn't have time to wait on hold.
  10. I submitted an email ticket requesting an update.
  11. I called NS again later that morning, got through to someone who explained that I needed to delete all the PHP files and let them know.
  12. (The site is written in PHP.)
  13. I told them this was a little crazy and to please turn our site back on.
  14. They told me, sorry, you violated the acceptable use policy, as we stated in the email.
  15. I said, what email?
  16. They informed me the primary account holder was notified 4 days ago.
  17. I requested they please send it to me.
  18. Asked to speak to a manager.
  19. After reading the email that just came in, I hung up and contacted stakeholders.
Admittedly I should have asked for ALL the information up front before getting upset (I was rather incredulous with the tech's answers).  It's reasonable to take a site down for hosting malicious content.  I would suggest it's NOT reasonable to take it offline without calling the customer first, or notifying ALL contact parties the site is being taken down.

Further, their policy is hands-off - they would only hint at what was wrong, would not provide full details - so you're on your own as to fixing the issue.  As a last issue, you have to wait 24-48 hours for 'approval' once you've notified them about the fix.

Thankfully our developer pulled an all-nighter to get the code side resolved, so now we're just waiting on approval to bring it back up (he can then put the final patch in place).  Still, that'll be almost a week of downtime by the time it's back online.  A week of downtime and they send one email.  Yeesh.

Points to consider:
  • Is your account/domain contact information valid?
  • What's your policy about security checks for code issues?
  • Who is responsible for security fixes/patches? (i.e. maintenance schedule)
  • Are you monitoring the site (independent of the hosting provider)?  Are you doing HTTP content checks?
  • Do you have emergency contact info for all stakeholders?

Comments

Popular posts from this blog

DFSR - eventid 4312 - replication just won't work

This warning isn't documented that well on the googles, so here's some google fodder:


You are trying to set up replication for a DFS folder (no existing replication)Source server is 2008R2, 'branch office' server is 2012R2 (I'm moving all our infra to 2012R2)You have no issues getting replication configuredYou see the DFSR folders get created on the other end, but nothing stagesFinally you get EventID 4312:
The DFS Replication service failed to get folder information when walking the file system on a journal wrap or loss recovery due to repeated sharing violations encountered on a folder. The service cannot replicate the folder and files in that folder until the sharing violation is resolved.  Additional Information:  Folder: F:\Users$\user.name\Desktop\Random Folder Name\  Replicated Folder Root: F:\Users$  File ID: {00000000-0000-0000-0000-000000000000}-v0  Replicated Folder Name: Users  Replicated Folder ID: 33F0449D-5E67-4DA1-99AC-681B5BACC7E5  Replication Group…

Fixing duplicate SPNs (service principal name)

This is a pretty handy thing to know:

SPNs are used when a specific service/daemon uses Kerberos to authenticate against AD. They map a specific service, port, and object together with this convention: class/host:port/name

If you use a computer object to auth (such as local service):
MSSQLSVC/tor-sql-01.domain.local:1433

If you use a user object to auth (such as a service account, or admin account):
MSSQLSVC/username:1433

Why do we care about duplicate SPNs? If you have two entries trying to auth using the same Kerberos ticket (I think that's right...), they will conflict, and cause errors and service failures.

To check for duplicate SPNs:
The command "setspn.exe -X

C:\Windows\system32>setspn -X
Processing entry 7
MSSQLSvc/server1.company.local:1433 is registered on these accounts:
CN=SERVER1,OU=servers,OU=resources,DC=company,DC=local
CN=SQL Admin,OU=service accounts,OU=resources,DC=company,DC=local

found 1 groups of duplicate SPNs. (truncated/sanitized)

Note that y…

Logstash to Nagios - alerting based on Windows Event ID

This took way longer than it should have to get going...so here's a config and brain dump...

Why?
You want to have a central place to analyze Windows Event/IIS/local application logs, alert off specific events, alert off specific situations.  You don't have the budget for a boxed solution.  You want pretty graphs.  You don't particularly care about individual server states.  (see rationale below - although you certainly have all the tools here to care, I haven't provided that configuration)

How?
ELK stack, OMD, NXlog agent, and Rsyslog.  The premise here is as follows:

Event generated on server into EventLogNXlog ships to Logstash inputLogstash filter adds fields and tags to specified eventsLogstash output sends to a passive Nagios service via the Nagios NSCA outputThe passive service on Nagios (Check_MK c/o OMD) does its thing w. alerting
OMD
Open Monitoring Distribution, but the real point here is Check_MK (IIRC Icinga uses this...).  It makes Nagios easy to use and main…