Splunk Observability Cloud US1

Is Splunk Observability Cloud US1 Down Right Now? Check if there is a current outage ongoing.

Splunk Observability Cloud US1 is currently Operational

Last checked from Splunk Observability Cloud US1's official status page

Historical record of incidents for Splunk Observability Cloud US1

Report: "Splunk RUM Monitoring MetricSets, Session Data and Trace Data is being Dropped"

Last update
investigating

We are continuing to investigate this issue.

investigating

A degradation in the performance of the Splunk RUM metrics processing pipeline is causing Monitoring MetricSets, Session Data, and Trace data to be delayed by more than five minutes. Charts and detectors built from Monitoring MetricSets are impacted

Report: "UI degraded, small number of datapoints being dropped"

Last update
resolved

This incident has been resolved.

monitoring

A fix has been implemented and we are monitoring the results.

investigating

Customers may experience intermittent timeout issues logging into app.us1.signalfx.com. We are investigating and will provide an update shortly. A small number of datapoints are not accepted at ingest, which may result in delays in some detectors.

investigating

Datapoint ingest is affected and we are dropping datapoints. We are investigating and will provide an update every 15 mins.

Report: "UI degraded, small number of datapoints being dropped"

Last update
Monitoring

A fix has been implemented and we are monitoring the results.

Update

Customers may experience intermittent timeout issues logging into app.us1.signalfx.com. We are investigating and will provide an update shortly. A small number of datapoints are not accepted at ingest, which may result in delays in some detectors.

Investigating

Datapoint ingest is affected and we are dropping datapoints. We are investigating and will provide an update every 15 mins.

Report: "Splunk Observability Cloud Application Intermittent Connection Issues"

Last update
Investigating

Customers may experience intermittent timeout issues logging into app.us1.signalfx.com. We are investigating and will provide an update shortly.

Report: "Datapoints Being Dropped"

Last update
Investigating

Datapoint ingest is affected and we are dropping datapoints. We are investigating and will provide an update every 15 mins.

Report: "Degraded performance accessing the Splunk APM Interface"

Last update
resolved

This incident has been resolved.

monitoring

A fix has been implemented and we are monitoring the results.

identified

We are investigating a degradation in performance while using the Splunk APM Interface. Parts of the Splunk APM Troubleshooting experience like the service map, map breakdowns, Tag Spotlight and charts within Splunk APM may be impacted. Trace data ingest is not impacted.

Report: "Intermittent login failures for customers using Unified identity"

Last update
resolved

This incident has been resolved.

investigating

Customers using unified identity may experience intermittent failures while logging into Splunk Observability cloud web interface. Datapoint Ingest is not affected. We are investigating and will provide an update shortly.

Report: "Intermittent login failures for customers using Unified identity"

Last update
Investigating

Customers using unified identity may experience intermittent failures while logging into Splunk Observability cloud web interface. Datapoint Ingest is not affected. We are investigating and will provide an update shortly.

Report: "Degraded performance accessing the Splunk APM Interface"

Last update
Identified

We are investigating a degradation in performance while using the Splunk APM Interface. Parts of the Splunk APM Troubleshooting experience like the service map, map breakdowns, Tag Spotlight and charts within Splunk APM may be impacted. Trace data ingest is not impacted.

Report: "Organization metrics(org metrics) data points are being dropped"

Last update
resolved

This incident has been resolved.

monitoring

A fix has been implemented and we are monitoring the results.

identified

The issue has been identified and a fix is being implemented.

investigating

A degradation of the ingest path for Organization metrics data points is causing data points to be dropped and lost. We are investigating the issue and will provide an update as soon as possible.

Report: "Cloud metric data points are delayed"

Last update
resolved

Cloud metrics may be delayed up up 30 minutes in time window: 2:10 - 2:37 AM PST. There is no data loss, but data might be delayed.

Report: "Organization metrics(org metrics) data points are being dropped"

Last update
Identified

The issue has been identified and a fix is being implemented.

Investigating

A degradation of the ingest path for Organization metrics data points is causing data points to be dropped and lost. We are investigating the issue and will provide an update as soon as possible.

Report: "Cloud metric data points are delayed"

Last update
Resolved

Cloud metrics may be delayed up up 30 minutes in time window: 2:10 - 2:37 AM PST. There is no data loss, but data might be delayed.

Report: "Alerts are delayed"

Last update
resolved

This incident has been resolved.

monitoring

The system has recovered and we are continuing to monitor

identified

We are actively implementing backend improvements to address delays in event availability. This multi-step process is expected to take several hours. Engineering teams across regions are coordinating to monitor progress and ensure full functionality is restored. Our next update will be provided by 10:00 AM UTC on May 9, 2025

identified

We are continuing to work on the issue. Alert notifications are being sent out on time. A small percentage of the events behind those notifications are delayed being created and are not available to see in the user interface.

identified

Efforts to resolve the issue are ongoing, and we’ll continue to share updates along the way

identified

We're making steady progress on the fix and will keep you informed as more details emerge

identified

We are continuing to work on a fix and will provide updates as more information becomes available

identified

We are continuing to work on a fix for this issue

identified

Starting at 10:30a PT some small percentage of alerts may have been delayed by up to 2-3 hours. The root cause is known and a fix is being worked on.

Report: "Alerts are delayed"

Last update
Identified

Starting at 10:30a PT some small percentage of alerts may have been delayed by up to 2-3 hours. The root cause is known and a fix is being worked on.

Report: "Splunk APM Monitoring, Troubleshooting MetricSets and Trace Data are Delayed"

Last update
resolved

This incident has been resolved.

monitoring

A fix has been implemented and we are monitoring the results.

investigating

A degradation in the performance of a key backend component of Splunk APM is causing both Troubleshooting, Monitoring MetricSets and Trace Data to be delayed by more than five minutes. No data is being dropped at this time but data for the APM Troubleshooting page, Tag Spotlight experience, Trace Analyzer and other APM pages, as well as metrics created from traces and APM detectors are all delayed.

Report: "Splunk APM Monitoring, Troubleshooting MetricSets and Trace Data are Delayed"

Last update
Investigating

A degradation in the performance of a key backend component of Splunk APM is causing both Troubleshooting, Monitoring MetricSets and Trace Data to be delayed by more than five minutes. No data is being dropped at this time but data for the APM Troubleshooting page, Tag Spotlight experience, Trace Analyzer and other APM pages, as well as metrics created from traces and APM detectors are all delayed.

Report: "Splunk Synthetics Google Chrome Upgrade"

Last update
Scheduled

Splunk Synthetic Monitoring will update Google Chrome and Chromium to version 135.0.7049.84-1 for Browser tests on 4/22 at 8 am PST. We periodically auto-update to newer versions of Google Chrome/Chromium when available. Due to differences between browser versions, Synthetics test behavior or timings can sometimes change and may require updates to your configured steps.

In progress

Scheduled maintenance is currently in progress. We will provide updates as necessary.

Report: "Log Observer Connect Search Failures"

Last update
resolved

This incident has been resolved.

monitoring

A fix has been implemented and we are monitoring the results.

identified

The issue has been identified and a fix is being implemented.

investigating

We're experiencing failures for searches related to Log observer. We will provide an update as soon as possible.

Report: "Log Observer Connect Search Failures"

Last update
Investigating

We're experiencing failures for searches related to Log observer. We will provide an update as soon as possible.

Report: "Charts may have been slow or not loaded"

Last update
resolved

From 10:35 to 10:43 PT charts may have been slow or not loaded. Ingestion and processing of data was not impacted. The issue is resolved at this time.

Report: "Charts may have been slow or not loaded"

Last update
Resolved

From 10:35 to 10:43 PT charts may have been slow or not loaded. Ingestion and processing of data was not impacted. The issue is resolved at this time.

Report: "Unable to fetch Splunk RUM sessions"

Last update
resolved

This incident has been resolved.

identified

The issue has been identified and a fix is being implemented.

investigating

We are currently investigating this issue.

Report: "Unable to fetch Splunk RUM sessions"

Last update
Resolved

This incident has been resolved.

Identified

The issue has been identified and a fix is being implemented.

Investigating

We are currently investigating this issue.

Report: "Delay in Alert notifications"

Last update
resolved

A fix has been applied and systems are now functioning as expected.

identified

We have identified the root cause of this issue and are working on fixing it. We will provide an update once we have applied the fix.

investigating

We are currently investigating this issue.

Report: "Delay in Alert notifications"

Last update
Investigating

We are currently investigating this issue.

Report: "Notifications Delayed"

Last update
resolved

This incident has been resolved.

monitoring

A fix has been implemented and we are monitoring the results.

identified

The issue has been identified and a fix is being implemented.

investigating

Notifications for Alerts by HTTP and email are being delayed. We are actively investigating the issue.

Report: "Notifications Delayed"

Last update
Investigating

Notifications for Alerts by HTTP and email are being delayed. We are actively investigating the issue.

Report: "Event creation delayed"

Last update
resolved

This incident has been resolved.

monitoring

Event creation may be delayed by 2 hours for some customers, alerting notification is not impacted but events for alerts may not be available in the UI until this incident is resolved.

Report: "Event creation delayed"

Last update
Monitoring

Event creation may be delayed by 2 hours for some customers, alerting notification is not impacted but events for alerts may not be available in the UI until this incident is resolved.

Report: "APM Landing Page & Charts Unable to Load"

Last update
resolved

This incident has been resolved.

monitoring

We have identified the issue and are starting to see services return to operational.

investigating

We are currently experiencing issues with Charts and APM and are investigating the issue.

Report: "Splunk APM Troubleshooting MetricSets Delayed"

Last update
resolved

This incident has been resolved.

monitoring

We are continuing to monitor for any further issues.

monitoring

A fix has been implemented and we are monitoring the results.

investigating

A degradation in the performance of the Splunk APM trace processing pipeline is causing Troubleshooting MetricSets to be delayed by more than five minutes. As a result, the APM Troubleshooting experience, service maps and Tag Spotlight do not have access to the most recent data. The processing of metrics for Business Workflows, which also depends on this pipeline, are equally delayed. Trace data ingest is not impacted at this time; service-level and endpoint-level Monitoring MetricSets and the detectors built from them are also not impacted.

Report: "Splunk Synthetics runners are unavailable in the Stockholm region"

Last update
resolved

This incident has been resolved.

monitoring

We are continuing to monitor for any further issues.

monitoring

We are seeing signs of recovery and are continuing to monitor.

identified

Splunk Synthetics runners are unavailable in the Stockholm region due to an AWS EC2 instance outage in the eu-north-1 region.

Report: "Degraded Performance of Web and APIs for Some Customers"

Last update
resolved

Between 11:30 AM and 3:00 PM (PST) on 2/10, customers experienced slow chart loading times and API responses. Some API requests returned errors. Data ingestion remained unaffected. We have identified and resolved the root cause.

Report: "Charts Slow or not Loading for Some Customers"

Last update
resolved

Some customers were experiencing slow chart loading times, and some charts may not load. There may be delays when creating new detectors as well. Data ingestion and the performance of opened charts and existing detectors remained unaffected. We have identified the root cause and implemented fixes. Our team is working on a long-term solution to prevent these performance issues. This issue has occurred multiple times over the past three days: once on 2/4 for 3 minutes, three times on 2/5 totaling 14 minutes, and four times on 2/6 totaling 56 minutes. The last occurrence was 3:53-4:11 PM (PST) on 2/6.

Report: "Charts Were Loading Slowly between 04:05 PM to 04:15 PM Pacific Time"

Last update
resolved

Charts for a subset of customers may be loading slowly between 04:05 PM to 04:15 PM Pacific Time. Portion of catrs may have failed to load between same time window. Datapoint ingest was not affected. Issue has been fully resolved since.

Report: "Seach indexing delays for updates to charts, detectors, and dashboards."

Last update
resolved

Starting at 3:37PM PT until 5:40PM PT our systems experienced an issue in which UI objects such as charts, detectors, and dashboards may not have been indexed when created or updated in the index when updated, resulting in potentially stale results when searching UI objects. The issue has since been fully resolved, no data was lost, all indices are now up-to-date. Alerting and telemetry data were not impacted.

Report: "Splunk RUM Session data Being Dropped"

Last update
resolved

This incident has been resolved.

monitoring

A fix has been implemented and we are monitoring the results.

identified

A degradation of the ingest path for Splunk RUM is causing session data to be dropped and lost. We are investigating the issue and will provide an update as soon as possible.

Report: "Creation and updates to dimensions are delayed"

Last update
resolved

This incident has been resolved.

investigating

We are continuing to investigate this issue.

investigating

We are experiencing delays in creation and updates to dimensions. We are actively investigating the root cause.

Report: "Creation and updates to dimensions are delayed"

Last update
resolved

This incident has been resolved.

monitoring

A fix has been implemented and we are monitoring the results.

identified

The issue has been identified and a fix is being implemented.

investigating

We are continuing to investigate this issue.

investigating

We are continuing to investigate this issue.

investigating

We are experiencing delays in creation and updates to dimensions. We are actively investigating the root cause.

Report: "Elevated Error Rate from the Splunk APM API"

Last update
resolved

This incident has been resolved.

monitoring

A fix has been implemented and we are monitoring the results.

identified

We are investigating an elevated rate of errors occurring while interacting with the Splunk APM API. Trace data ingest is not impacted.

Report: "Charts and Detectors Delayed"

Last update
resolved

This incident has been resolved.

investigating

Customers may be experiencing delays in some charts and detectors for new MTSes. Datapoint ingest is not affected. We are investigating and will provide an update shortly.

Report: "Splunk RUM Session data Being Dropped"

Last update
resolved

This incident has been resolved.

monitoring

A fix has been implemented and we are monitoring the results.

identified

We are continuing to work on a fix for this issue.

identified

The issue has been identified and a fix is being implemented.

investigating

We are continuing to investigate this issue.

investigating

A degradation of the ingest path for Splunk RUM is causing < 1% of session data to be dropped and lost. We are investigating the issue and will provide an update as soon as possible.

Report: "Customers may encounter intermittent issues when logging into unified identity orgs."

Last update
resolved

We are currently investigating this issue.

Report: "Splunk Log Observer Interface Unavailable"

Last update
resolved

Log Observer web application was unavailable for some customers between 3:29am PDT and 3:51am PDT. Log data ingest was not impacted. The system has since recovered and is operating as expected.

Report: "Splunk APM Trace Monitoring, Troubleshooting and Trace Ingestion MetricSets Delayed"

Last update
resolved

This incident has been resolved.

investigating

A degradation in the performance of the Splunk APM trace processing pipeline is causing Troubleshooting MetricSets to be delayed by more than five minutes. As a result, the APM Troubleshooting experience, service maps and Tag Spotlight do not have access to the most recent data. A degradation in the performance of the Splunk APM metrics processing pipeline is causing Monitoring MetricSets to be delayed by more than five minutes. Service, endpoint, and workflow dashboards, and other charts and detectors built from Monitoring MetricSets are impacted. A degradation in the performance of the Splunk APM data ingestion pipeline is causing the processing and storage of raw trace data to be delayed by more than five minutes. The processing of metrics for Business Workflows, which also depends on this pipeline, are equally delayed. Trace data ingest is not impacted at this time; service-level and endpoint-level Monitoring MetricSets and the detectors built from them are also not impacted.

Report: "Delayed Alerting From Detectors"

Last update
resolved

Less than 1% of alerts may have experienced latency from 7:27 am PST – 8:35 am PST.

Report: "Latency in Chart Loading"

Last update
resolved

Some users may have experienced latency in chart loading starting at 8:12 am PST and ending at 8:39 am PST

Report: "Datapoint Ingestion Lag"

Last update
resolved

We were able to resolve this issue. Ingestion was delayed only from AWS tooling due to upstream issues with the cloud provider - https://health.aws.amazon.com/health/status

identified

We have identified the cause and are currently resolving the issue.

investigating

We are currently experiencing lag when ingesting data and are investigating the issue.

Report: "Splunk APM Troubleshooting MetricSets Delayed"

Last update
resolved

This incident has been resolved.

monitoring

A fix has been implemented and we are monitoring the results.

identified

The issue has been identified and a fix is being implemented.

investigating

We are continuing to investigate this issue.

investigating

A degradation in the performance of the Splunk APM trace processing pipeline is causing Troubleshooting MetricSets to be delayed by more than five minutes. As a result, the APM Troubleshooting experience, service maps and Tag Spotlight do not have access to the most recent data. The processing of metrics for Business Workflows, which also depends on this pipeline, are equally delayed. Trace data ingest is not impacted at this time; service-level and endpoint-level Monitoring MetricSets and the detectors built from them are also not impacted.

Report: "Metrics and tags collection is delayed due to AWS outage"

Last update
resolved

This incident has been resolved.

monitoring

A fix has been implemented and we are monitoring the results.

identified

AWS is currently experiencing an outage. We will continue to monitor until the AWS issue is resolved. The latest information can be found here: https://health.aws.amazon.com/health/status

Report: "Splunk APM Troubleshooting MetricSets Delayed"

Last update
resolved

This incident has been resolved.

monitoring

A fix has been implemented and we are monitoring the results.

investigating

A degradation in the performance of the Splunk APM trace processing pipeline is causing 5% - 10% of Troubleshooting MetricSets to be delayed by more than five minutes. As a result, the APM Troubleshooting experience, service maps and Tag Spotlight do not have access to the most recent data. The processing of metrics for Business Workflows, which also depends on this pipeline, are equally delayed. Trace data ingest is not impacted at this time; service-level and endpoint-level Monitoring MetricSets and the detectors built from them are also not impacted.

Report: "Splunk APM Interface performance degraded"

Last update
resolved

We had degraded performance of the Splunk APM web application and UI was failing to load intermittently, between 2024-08-28 9:30AM Pacific and 2024-08-28 9:45AM. We have since recovered and APM UI is now fully available.

Report: "Splunk Observability Cloud Interface degraded"

Last update
resolved

This incident has been resolved.

monitoring

The system is recovering and we are continuing to monitor.

investigating

We are investigating a degradation of the Splunk Observability Cloud application. Data ingest is not impacted at this time. We will provide an update as soon as possible.

Report: "Splunk Synthetic Monitoring updated Google Chrome to version 125"

Last update
resolved

Splunk Synthetic Monitoring updated Google Chrome to version 125 for Browser tests on July 18 at 12:30 PM EDT. We periodically auto-update to newer versions of Google Chrome when available. Due to differences between browser versions, test behavior or timings can sometimes change and may require updates to your test steps.

Report: "Delayed metrics from GCP"

Last update
resolved

The issue with Cloud Monitoring metrics has been resolved for all affected users as of Monday, 2024-07-15 10:52 US/Pacific.

identified

We are experiencing an issue syncing cloud metrics from GCP. Some metrics from GCP could be delayed or dropped. This is due to a GCP issue. See status page here: https://status.cloud.google.com/incidents/ERzzrJqeGR2GCW51XKFv

investigating

Experiencing an issue syncing cloud metrics from GCP. Some metrics from GCP could be delayed or dropped.

investigating

Experiencing an issue syncing cloud metrics from GCP. Some metrics from GCP could be delayed.

Report: "Event alerts are delayed by 2 hours"

Last update
resolved

A fix has been implemented and full functionality has been restored. Past missing events have been filled in and will be visible as expected

monitoring

A fix has been implemented and we are monitoring the results. During the incident, utilizing our APIs or UIs for visualizing alerts, muting rules or session logging events may be missing some information that will be later filled in. However, all notifications should still have been generated and sent externally as expected and on-time. This includes Slack, ServiceNow and JIRA integrations as well as emails. Expected resolution by 21:30 PT

monitoring

A fix has been implemented and we are monitoring the results. During the incident, utilizing our APIs or UIs for visualizing alerts, muting rules or session logging events may be missing some information that will be later filled in. However, all notifications should still have been generated and sent externally as expected and on-time. This includes Slack, ServiceNow and JIRA integrations as well as emails. Expected resolution by 21:00 PT.

monitoring

A fix has been implemented and we are monitoring the results. During the incident, utilizing our APIs or UIs for visualizing alerts, muting rules or session logging events may be missing some information that will be later filled in. However, all notifications should still have been generated and sent externally as expected and on-time. This includes Slack, ServiceNow and JIRA integrations as well as emails. Expected resolution by 21:30 PT.

monitoring

A fix has been implemented and we are monitoring the results.

monitoring

A fix has been implemented and we are monitoring the results.

monitoring

A fix has been implemented and we are monitoring the results.

identified

Event alerts are delayed by 2 hours with some small number of events lost. UI and ingest are unaffected. The problem has been identified and we are working on a resolution. Updates will be provided every 15 minutes.

Report: "Interface loading slowly"

Last update
resolved

This incident has been resolved.

monitoring

We are continuing to monitor the environment.

monitoring

We are continuing to monitor the environment. Some of the write operations may be degraded while systems continue to recover.

monitoring

A fix has been pushed. The interface and log-in latency has improved. We are monitoring the results.

monitoring

A fix has been pushed. The interface and log-in latency has improved. We are monitoring the results.

identified

We are continuing to develop a fix. We will issue another update shortly.

identified

This issue has been identified and a fix is being implemented. We will issue another update shortly.

investigating

We are continuing to investigate this issue. We will issue another update shortly.

investigating

We are aware that the interface and log-ins may be slow or unresponsive. We are actively investigating the issue. We will issue another update shortly.

Report: "Datapoints Ingestion Delayed"

Last update
resolved

This incident has been resolved.

identified

Datapoint ingest is affected and we are experiencing some delay in ingest. We have identified the issue and are working to resolve the problem. Updates will be provided every fifteen minutes.

Report: "Charts not Loading for Some Customers"

Last update
resolved

This incident has been resolved.

investigating

We are continuing to investigate this issue.

investigating

Charts on APM Profiling for a subset of customers may not be loading. Datapoint ingest is not affected. We are investigating and will provide an update shortly.

Report: "Charts not Loading for Some Customers"

Last update
resolved

This incident has been resolved.

monitoring

A fix has been implemented and we are monitoring the results.

investigating

Charts in APM Profiling for a subset of customers may not be loading. Datapoint ingest is not affected. We are investigating and will provide an update shortly.

Report: "Charts not Loading for Some Customers"

Last update
resolved

This incident has been resolved as of 15:13 PDT. Charts and alerts are now operational.

identified

After further review, in addition to charts not loading for a subset of customers, some alerts are delayed, some alerts may be sending false positives, and some alerts may not be sent for some customers. The fix that is being rolled out will fix the alerts as well as charts not loading. We are continuing to monitor the roll out of the fix.

identified

The issue has been identified and a fix is being implemented.

investigating

Starting at 1:50p PT, charts for a subset of customers may not be loading. Datapoint ingest is not affected. We are investigating and will provide an update shortly.

Report: "Charts not Loading and detectors not starting for Some Customers"

Last update
resolved

Following the monitoring period, no further issues have been observed and the incident is now considered as Resolved.

monitoring

A fix has been implemented at 04:14 PT, the services are now recovered and we are monitoring the results.

investigating

There are also issues since 3:21a PT in the Monitoring MetricSets. A degradation in the performance of the Splunk APM metrics processing pipeline is causing Monitoring MetricSets to be corrupted and represent only a portion of the traffic and activity they are meant to measure on our customer’s services, endpoints, traces and workflows.

investigating

As of 3:18a PT this morning, charts for a subset of customers may not be loading. Some detectors may not be able to start. Datapoint ingest is not affected. We are investigating and will provide an update shortly.

Report: "Charts not Loading"

Last update
resolved

Charts for all customers may not be loading from 10:44 am - 10:55 am. Alert notifications are delayed. Datapoint ingest is not affected. This issue is completely resolved now.

Report: "Splunk APM Interface Unavailable"

Last update
resolved

We ran into issues with availability of Splunk APM UI from 2:06am - 2:14 am PDT. Trace data ingest was not impacted and this incident is completely resolved.

Report: "Splunk Observability Cloud Web Interface Intermittently Unavailable"

Last update
resolved

This incident has been resolved.

monitoring

A fix has been implemented and we are monitoring the results.

identified

The issue has been identified and a fix is being implemented.

investigating

We are investigating an intermittent issue with Splunk Observability Cloud web application. Data ingest is not impacted at this time. We will provide an update as soon as possible.

Report: "Splunk APM Interface Unavailable"

Last update
resolved

We ran into issues with availability of Splunk APM UI from 10:54am - 11:04 am PDT. Trace data ingest was not impacted and this incident is completely resolved.