MonsterMegs - Notice history

All systems operational

About This Site

Welcome to MonsterMegs' status page. If you want to keep on top of any disruptions in service to our website, control panel, or hosting platform, this is the page to check. We report minor/major outages and scheduled maintenance to this status page. However, issues affecting a small number of customers may be reported directly to affected customers in MyMonsterMegs (https://my.monstermegs.com). If you're experiencing an issue you do not see reported on this page, please log into MyMonsterMegs to view any alerts our team has added to your account.

100% - uptime

Website - Operational

100% - uptime
Dec 2024 · 100.0%Jan 2025 · 100.0%Feb · 100.0%
Dec 2024
Jan 2025
Feb 2025

Customer Portal - Operational

100% - uptime
Dec 2024 · 100.0%Jan 2025 · 100.0%Feb · 100.0%
Dec 2024
Jan 2025
Feb 2025
100% - uptime

Thunder - Operational

100% - uptime
Dec 2024 · 100.0%Jan 2025 · 100.0%Feb · 100.0%
Dec 2024
Jan 2025
Feb 2025

Hurricane - Operational

100% - uptime
Dec 2024 · 99.74%Jan 2025 · 100.0%Feb · 100.0%
Dec 2024
Jan 2025
Feb 2025

Storm - Operational

99% - uptime
Dec 2024 · 100.0%Jan 2025 · 100.0%Feb · 97.76%
Dec 2024
Jan 2025
Feb 2025

Lightning - Operational

100% - uptime
Dec 2024 · 99.83%Jan 2025 · 100.0%Feb · 100.0%
Dec 2024
Jan 2025
Feb 2025
100% - uptime

DNS-1 - Operational

100% - uptime
Dec 2024 · 100.0%Jan 2025 · 100.0%Feb · 100.0%
Dec 2024
Jan 2025
Feb 2025

DNS-2 - Operational

100% - uptime
Dec 2024 · 100.0%Jan 2025 · 100.0%Feb · 100.0%
Dec 2024
Jan 2025
Feb 2025

DNS-3 - Operational

100% - uptime
Dec 2024 · 100.0%Jan 2025 · 100.0%Feb · 100.0%
Dec 2024
Jan 2025
Feb 2025

DNS-4 - Operational

100% - uptime
Dec 2024 · 100.0%Jan 2025 · 100.0%Feb · 100.0%
Dec 2024
Jan 2025
Feb 2025
100% - uptime

US Backup Storage Daily - Operational

100% - uptime
Dec 2024 · 100.0%Jan 2025 · 100.0%Feb · 100.0%
Dec 2024
Jan 2025
Feb 2025

US Backup Storage Weekly - Operational

100% - uptime
Dec 2024 · 100.0%Jan 2025 · 100.0%Feb · 100.0%
Dec 2024
Jan 2025
Feb 2025

EU Backup Storage Daily - Operational

100% - uptime
Dec 2024 · 100.0%Jan 2025 · 100.0%Feb · 100.0%
Dec 2024
Jan 2025
Feb 2025

EU Backup Storage Weekly - Operational

100% - uptime
Dec 2024 · 100.0%Jan 2025 · 100.0%Feb · 100.0%
Dec 2024
Jan 2025
Feb 2025
100% - uptime

Zabbix Monitoring Server US - Operational

100% - uptime
Dec 2024 · 100.0%Jan 2025 · 100.0%Feb · 100.0%
Dec 2024
Jan 2025
Feb 2025

Zabbix Monitoring Server EU - Operational

100% - uptime
Dec 2024 · 100.0%Jan 2025 · 100.0%Feb · 100.0%
Dec 2024
Jan 2025
Feb 2025

Third Party: Cloudflare → Cloudflare Sites and Services → CDN/Cache - Operational

Notice history

Feb 2025

Failed Hard Drive - Storm
  • Resolved
    Resolved

    The server has now been running for over 4 hours and there have not been any errors. We are now confident that the issue is resolved.

  • Update
    Update

    We are happy to state that the raid rebuild did complete and everything does look to be stable at this time. We do feel that the PCI Adapter was the cause of the hard drives showing as failed or disappearing from bios.

    We are going to keep this incident open for a few more hours and will closely monitor the server for any further issues.

  • Monitoring
    Monitoring

    The PCI Adapter has been replaced and the second hard drive has been added back in. We will continue to monitor the server closely over the next few hours and will update this incident if there are any further issues.

  • Investigating
    Investigating

    After talking with the datacenter, we feel we might have a lead on what is causing the drives too show failed or fall out of connection. It appears this server setup uses a PCI adapter when there is more than 1 hard drive installed. This morning we had the datacenter bypass the PCI adapter and plugin the min drive directly into the motherboard. The server has been running stable now for nearly 3 hours.

    So we feel that the PCI adapter might be failing. Within the hour we are going to take the server offline once again to plug the second drive back in and use a new PCI adapter. If this does snot resolve the issue, then we will have to migrate to a new server, but we have high hopes that it is the PCI adapter that is at fault.

  • Monitoring
    Monitoring

    As you may have noticed, the server is back online. Please note that is most likely temporary. The datacenter did not feel that the problem was with the motherboard, so we have booted the server with just the one drive that has all the data.

    We are gonna let this run for a couple hours and see if it remains stable. If it does, our next plan of action is most likely to setup a second server and then migrate the data to this second server. If the hard drive does fall out of connection or throws any errors, then we will be forced to rebuild the server and restore backups.

    We will update this post, once we have our next plan of action.

  • Identified
    Identified

    We are currently talking with the datacenter to try a motherboard replacement. The fact that the drive does show up fine for a period of time and then disappears, makes us think the motherboard may be malfunctioning.

    This will be our last attempt to rectify this issue and if this does not work, we will have to proceed with disaster recovery.

  • Investigating
    Investigating

    Unfortunately, the resync of the raid array failed at 82%. So we are continuing to explore our options and see if maybe the new hard drive has issues itself.

  • Monitoring
    Monitoring

    We hopefully have some good news, but we are not out of the woods yet. We had the datacenter swap the drives around from their original mounting location. This many times, will refresh the bios to see the drives again. This could have been a case of the main drive getting bumped and having a loose connection or the bios just having issues reading the drives.

    Now that the main drive is showing up, the server booted correctly and the raid has been resyncing for the last 20 minutes or so. Once the raid resync completes, everything should be good to go. We will update once the resync is complete or if there are any further issues along the way.

  • Update
    Update

    After further investigation, it looks like the second drive might have failed during the resync of the raid array. This is extremely rare, but it has been seen to happen. Before we jump to the extreme and reprovision the server, reconfigure it, and start restoring backups...we are gonna still work to try and recover the data without a complete rebuild.

    If we do have to resort to a full rebuild, we will be email all customers with further details. While we work to try and recover the data as it stands, the updates to this incident will be limited. As we wan to put full focus on attempting to recover the data.

    If it does come to the point that we need to rebuild and restore backups, this could realistically take a few days to fully restore the server. So we want everyone to be aware that will not be a quick process and will take time to restore everything correctly without rushing and causing more errors.

  • Investigating
    Investigating

    It appears that was not the cause as we had thought. Again we are seeing services falling off after a period of time. We are continuing to troubleshoot the issue. This is looking to be a bit more complicated than we anticipated. At this point we do not have an accurate timeframe to offer of how soon this will be resolved.

  • Monitoring
    Monitoring

    We found that after the reboot, one of the servers security programs was triggering some extreme ddos protection that was freezing the server and causing processes to fail. We believe we have disabled the features that have caused this and have also reached out to the vendor to find out why this happened.

    If this does happen again, we will disable the service completely, until we can work this out with the vendor. We are going to continue to monitor over the next few hours.

  • Update
    Update

    We are still investigating the what happened during the raid rebuild. At this time we don't have any further information, but hope to have more details very soon.

  • Investigating
    Investigating

    Something has happened during the raid rebuild and has caused several parts of the server to become inaccessible. We are investigating and will update as we have more information.

  • Monitoring
    Monitoring

    We found the newly installed drive has the latest firmware, so there is no need for further reboots. We are awaiting for the raid rebuild to complete and then we will consider this issue closed.

  • Update
    Update

    The hard drive has been replaced and the server is back online. Now we will proceed to rebuild the raid and update the firmware on the drive. Expect 1-2 reboots and then the replacement will be complete.

  • Update
    Update

    The datacenter has already taken the server offline to replace the drive. We will update as things progress.

  • Identified
    Identified

    We have been informed by our monitoring systems, that one of the hard drives on the server has failed. We are going to proceed immediately with a request to have the hard drive replaced. We anticipate the drive to be replace within the next couple hours. During this time, the server will go offline for about 30 minutes. After the replacement is done, there will be a few reboots required to rebuild the raid and update the firmware on the drive.

Jan 2025

No notices reported this month

Dec 2024

Hetzner Emergency Router Maintenance
  • Resolved
    Resolved

    After a server reboot, the server is back online. It appears for some reason, the server did not recapture the incoming network connection and a full server reboot was required.

  • Update
    Update

    Hetzner has listed the maintenance as complete, but still the Hurricane server has no connectivity. We have contacted Hetzner support to find out why and are awaiting an update. We we post back as soon as we have more information.

  • Update
    Update

    The Lightning server is back online. We are currently waiting for connectivity to be restored on the Hurricane server.

  • Identified
    Identified

    It appears they are performing emergency maintenance on several routers. Here is a snippet from their status page:

    During the above-mentioned time frame, we will be carrying out urgent maintenance work on the access router fsn1-dc13-ex9k2, fsn1-dc13-ex9k1, fsn1-dc11-ex9k2, fsn1-dc11-ex9k1, and the connected ToR switches. During this work, network traffic will be interrupted, meaning that affected customer systems will not be accessible during this time. This also affects vSwitches for Dedicated Root Servers. Private Switches and Custom Solutions are excluded from the maintenance, unless explicitly stated.

  • Investigating
    Investigating

    We are currently investigating a network outage in the Hetzner datacenter. We will update when we have more information.

Dec 2024 to Feb 2025

Next