Do you run your own name servers? Want to sign your DNS requests? Check out DNSSEC support at Gandi!
Gandi is pleased to offer .NZ, official ccTLD of New Zealand.
This extension is only available as third-level domains such as: .CO.NZ, .NET.NZ, or .ORG.NZ.
You can purchase a .NZ for $24.00 excl. VAT for A rates to $18.00 excl. VAT for E rates for one year (see all the prices).
Register a .NZ
Gandi is pleased to offer another European extension: .RO, the official extension of Romania.
.RO domains can be registered at Gandi for €36.00 excl. VAT under A rates to €32.00 excl. VAT under E rates for one year.
You can see the information page and prices here.
Register a .RO:
Starting today, you can register your .GG (Guernesey) and .JE (Jersey) domains at Gandi. A great way to round out our collection for the Channel islands, and which should go well with the .IM (Isle of Man) extension, which has been available at Gandi for a couple years already.
The .GG is sold at £55.00 excl. VAT under A rates to £48.00 excl. VAT under E rates, per year (see all the prices).
The .JE extension is also sold at £55.00 excl. VAT under A rates to £48.00 excl. VAT under E rates, per year (see all the prices).
Register a .GG or .JE domain at Gandi:
We are currently experiencing issues with webmail authentication. Our teams are working to resolve the problem as soon as possible. This issue does not impact reception of your emails, but only accessing them.
We apologise for the inconvenience.
Update 29/02/2012: 13:30 GMT: The issue has been stabilised since the end of the afternoon yesterday. We are nevertheless working on the platform to avoid furture recurrence of this issue.
To correct problems we have identified in our storage systems, we need to perform corrective maintenance procedures tonight between 23:30 and 3:30 CET. This will impact some of your server systems, making certain data volumes unavailable for a period of 15 to 20 minutes.
We recommend that you do not restart your servers during this period. All services should return to normal immediately following the termination of the maintenance window.
If you will be affected by this maintenance, you should have received a mail from us advising you of the issues and timining. Please see this page for the contents of that message.
We will update this post to keep you informed of our progress.
Edit 23:30 CET: operation started
Edit 00:32 CET: first reboot done, equipment beeing behaving as expected, we proceed further
Edit 01:00 CET: most reboots are now ongoing, our first upgrades having been successful
Edit 01:30 CET: most of our storage units have been upgraded - if your server recovered from I/O stalls, this issue is fixed for you -- otherwise, this upgrade will be finished in the next hour
Edit 01:34 CEST: a compute node crashed during this operation, we are starting the affected virtual servers on another machine right now.
Edit 02h30: maintenance is finished, thank your for your patience during this operation.
Our Simple Hosting has been available for a month now. Here are some feedback, and some links you may find useful.
A piece of storage equipment is failing, likely due to defective components. Our teams are currently working to restore the situation as soon as possible. We recommend that you do not restart your server if you are impacted. We will keep you informed of our progress of this incident in this article.
[02:28 CET] Recovery sucessful. Components repaired and storage unit now nominal.
We have a temporary emergency halt on the hosting storage system (filers). We recommend that you do NOT attempt to restart your server. The impacted servers should recover in the next few minutes. We will update you with further information as soon as possible.
[edit 00:00] The services are fully restored as of 21:20 CET. Most users were back to nominal function before 19:30, but some took longer to start. Identifiable blocked systems were managed and restarted manually. Please restart your services if they are still unavailable at this time, and contact support if your server is not available and cannot be restarted.
A storage unit is currently experiencing a slowdown. Our teams are currently working on a solution.
Update (09:45 GMT): The situation improved between 07:00 and 08:00 GMT. There were significant slowdowns between 05:00 and 06:50 GMT.
Update (January 25th 09:00 GMT): A storage equipment is currently experiencing slowdown. The incident is similar to the one yesterday. Our technical team is working on solving the issue.
Update (January 25th 10:00 GMT): The I/O situation improved. Our technical team is still working to find a complete fix to the issue.
Update (January 25th 10:22 GMT): A storage equipment is currently experiencing slowdown. The incident is similar to the one this morning. Our technical team is working on solving the issue.
Update (January 26th 11:26 GMT): The I/O situation improved. Our technical team is still working to find a complete fix to the issue.
Update (January 27th 19:11 GMT): A storage equipment is currently experiencing slowdown. The incident is similar to the incident of the week. Our technical team is working on solving the issue.
Update (January 27th 22:00 GMT): The I/O situation is now stabilized. Our technical team is still working to find a complete fix to the issue.
Update (February 2nd 03h30 GMT): Another incident affects one of our storage units. We're now rebooting the faulty equipment. We recently found a few corrective actions that we'll soon be able to take in order to solve this kind of issues.
Update (February 2nd 20:19 GMT): Another incident has occurred, and slowdown was noticed, however the situation is stable right now.
Update (February 6th 02:09 GMT): Slowdown on one of our storage units. Teams working on it.
Two storage units are concerned by these incidents, which are isolated slowdowns in read/write operations. We suspect that the problem is two-fold: a software problem (blocking of operations), and a hardware problem (some disk models are unusually slow).
When these slowdowns occur, the implementation of iSCSI that lets us connect your servers to their disks may be dysfunctional. The result is an "I/O wait" that is artificially high (100%) even if the storage is once again rapid.
We are currently working on these three problems by giving priority to the capacity of our system to re-establish service after a slowdown.