Archive for the ‘Uncategorized’ Category

Lon50/Telecity-hex89.core device outage

Posted: Tuesday, August 18th, 2020 at 05:20 by Naz

Our Co-Location provider for this site has advised of a power outage which is effecting multiple floors within the London Harbour Exchange building. Investigations are being carried out to determine an RCA. At the time of this post the building has been evacuated due to the fire alarm triggering.

More information to be provided as and when we have updates from our Co-Location provider

  1. ljordan says:

    Further updates pending but Talk Talk are currently invesigating further. As soon as they are made available we will advise further.

  2. ljordan says:

    We are still awaiting further updates from the suppliers on this outage. The London building is also still under evacuation. The engineers are on site to investigate as soon as possible. As soon as we have any news, we will advise further.

  3. ljordan says:

    It been advised by Equinix that there is multiple core links experiencing an outage and 36 network to network interface ports causing Business customers connected to the London Harbour and Harbour exchanges to experience a Total Loss of Service. Outage start time was 04:23 (18/08). Full Impact assessment is still ongoing.

    Latest Update
    Equinix have advised that a fire alarm was triggered by the failure of output static switch from their Galaxy UPS system. This has resulted in a loss of power for multiple customers and Equinix IBX Engineers are working to resolve the issue

  4. mpurcell says:

    Latest Update:

    Engineers from Equinix continue to work on restoring power at the datacentre to restore service.

    Escalations have been made to suppliers who hand over services to us at this location with a view of getting more detail and timescales from Equinix. Equinix IBX site staff reports that IBX engineers and the specialist vendor have begun restoring services to customers by migrating to newly installed and commissioned infrastructure. IBX engineers continue to work towards restoring service to all customers

    Root cause has not currently been identified.

    Next update expected at 1000

  5. mpurcell says:

    Update is due shortly. Will update here as soon as it is received.

  6. mpurcell says:

    Latest Update:

    Talk Talk Support Teams have confirmed that services remain impacted but IBX staff are being allowed back on site and have confirmed that the fire detection system has been fully restored. The Galaxy UPS failure remains an ongoing incident and is currently under investigation

  7. mpurcell says:

    Latest Update:

    Equinix Engineers have advised that their IBX team have begun restoring power to affected devices. Unfortunately, at present there remains no estimate resolution time.

    We will continue to provide updates as soon as they come through.

  8. mpurcell says:

    Latest Update:

    Equinix IBX staff have reported that some services have been reinstated and they continue to work towards restoring services to all customers by migrating to the newly installed and commissioned infrastructure. Talk Talk Network teams have also confirmed that some 3rd party services are coming back online.

  9. mpurcell says:

    Latest Update:

    Equinix IBX Site Staff reports that services have been further restored for several more customers and IBX Engineers continue to work towards restoring services to all customers by migrating to the newly installed and commissioned infrastructure

  10. mpurcell says:

    Apologies, no further updates at present. As soon as one comes through we will update here.

  11. mpurcell says:

    Latest Update:

    Equinix IBX Site Staff reports that services have been further restored to more customers and increasing numbers of those affected are now operational along with the majority of Equinix Network Services. IBX Engineers continue to work towards restoring services to all customers by migrating to the newly installed and commissioned infrastructure.

    No root cause has yet been identified.

  12. mpurcell says:

    Latest Update:

    Internally we are beginning to see an increasing number of our services coming back online.

  13. mpurcell says:

    Latest Update:

    Talk Talk reports that Service was restored at 15:36 to the TalkTalk switch, in HEX, on the Legacy TUK Network.

    IBX Engineers continue to work towards restoring services to all customers by migrating to the newly installed and commissioned infrastructure and estimate full restoration by or before 21:00 GMT/BST.

  14. mpurcell says:

    Additional to last:

    Equinix have advised that electrical work is being carried out at the data centre whereby services (under floor sockets) require migration to a new distribution board as one has failed. There are 8 floors that require this work to be carried out. Floors one and two have been completed. As a result we still consider the services to be at risk, despite them being restored currently. Further outages may be seen until the expected completion time of 21:00hrs tonight.

  15. ljordan says:

    Latest Update
    Equinix IBX Site Staff reports that restoration of services continues. The remaining affected customers are in the process of being migrated to the newly installed and commissioned infrastructure. A revised resolution time has not yet been provided.

  16. ljordan says:

    Equinix advised that all services are now restored. All customers are now fed from the new infrastructure and at full redundancy. UPS failure was deemed the cause. The incident will now be resolved as all systems are available with no further issues reported. Monitoring will remain in place to ensure stability.

Telehouse Migrations

Posted: Tuesday, August 11th, 2020 at 14:19 by Iain Beveridge

From Wednesday 12th August 2020 19:00 – 02:00, Nightly until Friday 14th August 2020.

Due to the Telehouse Metro data centre closing, there is an ongoing project to migrate all Vodafone provided services to a new handoff location at Equinix Slough.

A maintenance window is scheduled nightly from 19:00 until 02:00 the following morning, customers will experience a single 30 minute outage whilst their individual circuit’s configuration is migrated and tested.

If you require any additional information, please contact the Service Desk at Service.desk@cityfibre.com

Incident: Virgin Media leased lines

Posted: Monday, August 10th, 2020 at 17:14 by Jonathan Clarke

We are currently seeing a number of Virgin Media leased line circuits as down. We are currently in correspondence with suppliers for diagnostics.

Further updates will be provided when available.

  1. Jonathan Clarke says:

    We can see that the circuits have now reconnected, We would advise rebooting equipment if the issue remains. Diagnostics are still being performed to ascertain why this drop occurred.

    Further updates will be provided when available.

  2. Jonathan Clarke says:

    Following further diagnostics and reports, we can see the circuits have reconnected, however, some circuits are unable to break out.

    We are continuing to investigate this and we will update you again when we have any further information.

  3. Jonathan Clarke says:

    Suppliers has confirmed that they are suffering with packet loss on their link that provides for the circuits having breakout problems.

    This is being investigated by Virgin Media and we are now awaiting an update from them.

    Further updates will be provided when available. We appreciate your patience regarding the fault.

  4. Jonathan Clarke says:

    Suppliers has confirmed an engineer attended site and resolved issue with the packet loss, they are monitoring the connection, however, believe this to be resolved.

    We are now awaiting restoration confirmation and we can update you. We apologise for the inconvenience cause to you and your customers.

  5. Chris McDonald says:

    Virgin Media have confirmed that service is fully restored. There was a card failure and their engineer has been to site to replace the faulty card.

Telehouse Circuit Migration

Posted: Tuesday, July 28th, 2020 at 15:45 by Iain Beveridge

From Friday 31st July 2020 19:00 – 02:00, Nightly until Wednesday 12th August 2020.

Due to the Telehouse Metro data centre closing, there is an ongoing project to migrate all Vodafone provided services to a new handoff location at Equinix Slough.

A maintenance window is scheduled nightly from 19:00 until 02:00 the following morning, customers will experience a single 30 minute outage whilst their individual circuit’s configuration is migrated and tested.

If you require any additional information, please contact the Service Desk at Service.desk@cityfibre.com

Incident: Vodafone leased lines

Posted: Friday, July 24th, 2020 at 16:19 by david labouchardiere

We are currently seeing a number of Vodafone leased line circuits as down. Initial diagnostics indicate the circuits are related to a single Vodafone handoff. Engineers are investigating and engaging with our supplier. Further updates will be provided when available.

  1. Jonathan Clarke says:

    Suppliers have advised that diagnostics show their management kit unreachable due to which services are impacted. Vodafone have proactively escalated to Level 1 to expedite the fault and restoration of the services.

    Further updates will be provided when available.

  2. Jonathan Clarke says:

    Suppliers have advised that a field engineer has been dispatched to investigate, the ETA for the engineer is 18:30.

    Further updates will be provided when available.

  3. Jonathan Clarke says:

    Suppliers have advised that the field engineer recently dispatched to investigate has been delayed and therefore, the ETA is now 19:30

    We apologise for the delay and we will give further updates when provided.

  4. Jonathan Clarke says:

    Suppliers have advised that the field engineer has arrived on site to complete investigations.

    Further updates will be provided when available.

  5. Jonathan Clarke says:

    Suppliers advise investigations are still ongoing and we are awaiting further updates.

    Thank you for you patience.

  6. Jonathan Clarke says:

    Suppliers advise that the field engineer has been investigating with a 2nd Line Engineer. We have been advised that their equipment is unable to pass traffic and they suspect a line card issue. Suppliers transmission team has been engaged and a resolution is being attempted.

  7. Jonathan Clarke says:

    Suppliers advise that the transmission team has ordered a spare line card for replacement. ETA for the spare card to arrive at the site is 60 minutes. We are proactively monitoring for further updates. When a further update is obtained, we will update you again.

    We appreciate your patience regarding the incident.

  8. Jonathan Clarke says:

    Engineers are continuing to work on this incident, we are awaiting a further update. Once obtained, we will update you again.

  9. Jonathan Clarke says:

    Suppliers have replaced the line card and confirmed that the issue is resolved. We are now seeing our managed routers back online.

    If you have any further issues, please reboot the equipment and if after, the issue persists please call us.

    We appreciate your patience and apologise for any inconvenience caused.

At risk: bristol.core / bri165-rtr-1

Posted: Friday, July 3rd, 2020 at 14:43 by Adam Wilson

Tuesday 7th July 2020, 22:00 – 23:59

After a power check within Bristol POP, BT identified an issue with part of the power infrastructure within one of our racks.

Therefore, BT will be undertaking emergency maintenance work in the above window which may require the power feeds to be moved, whilst a UPS is swapped within the rack. The core network router is not expected to experience any loss of service.

  1. This work was postponed due to third party resourcing issues, revised date will be republished shortly.

Incident: Bristol POP

Posted: Thursday, June 25th, 2020 at 03:56 by Robin Posadas

We have experienced a loss of connectivity to our equipment in our Bristol POP. Between 02:50 – 03:00. Currently all services have been restored. Engineers are investigating. Further updates will be provided as and when they become available.

  1. Alistair Patterson says:

    This incident was due to a power issue within the pop. This was raised to our supplier at the time who are continuing to investigate the root cause. A further update will be provided when available.

Emergency Work: interxion2.edge switch stack reboot

Posted: Tuesday, June 23rd, 2020 at 17:48 by Richard Blunt

Tuesday 23 June 23:00 – 23:15

Memory fragmentation has been detected in one of the Interxion stack switches, This is causing intermittent connectivity to some customer firewalls.

The device will have an emergency reboot at 23:00 to clear the issue. This will cause a brief outage with some none HA devices.

 

An update will be posted on completion of the work.

  1. Richard Blunt says:

    Scheduled reboot completed successfully.

At Risk: Columbo POP

Posted: Tuesday, June 2nd, 2020 at 12:17 by Daniel Jones
We have identified an issue via network monitoring. Engineers are working to resolve the issue.
Site is at risk until further notice.
  1. Daniel Jones says:

    An engineer is en-route to site to rectify the issue.

  2. Aiden Spencer says:

    The issue is now resolved, equipment has been replaced and no customers were impacted.

VOIP Routing Table

Posted: Monday, May 4th, 2020 at 16:02 by Iain Beveridge

Monday 4th May 2020 21:00 – 21:10

We are carrying out essential planned maintenance on our VoIP routing table this evening during the above window. All precautions have been taken to ensure this should not be service affecting.