Pipefy

Is Pipefy Down Right Now? Check if there is a current outage ongoing.

Pipefy is currently Degraded

Last checked from Pipefy's official status page

Historical record of incidents for Pipefy

Report: "DNS Propagation in Progress"

Last update
monitoring

We're currently undergoing DNS propagation, which might cause some customers to experience intermittent issues accessing Pipefy. This can depend on your network or geographical region. While this process completes, if you're facing critical access problems, we suggest temporarily switching to Google's public DNS servers. This often helps resolve name resolution issues quickly. - Google DNS: Primary: 8.8.8.8 Secondary: 8.8.4.4 - Steps for Windows users: 1. Open Command Prompt as Administrator: 2. Press the Windows key and type "cmd". 3. Right-click on "Command Prompt" and select "Run as administrator". 4. In the Command Prompt window, type ipconfig /flushdns and press Enter. You should see a message confirming that the DNS cache has been flushed. For those who can't or aren't able to do this yourself, please involve your local IT team. They can either make the changes for individual users or adjust settings on your VPNs. We appreciate your patience as we work to ensure full service availability.

Report: "Unexpected internal issues"

Last update
resolved

We would like to confirm that the issue has been officially resolved. The Root Cause Analysis (RCA) will be shared as soon as we have more details. We sincerely apologize for the inconvenience and for falling short of the level of service you expect from us.

monitoring

The problem has been identified and the fix has been applied by our provider. You may experience some intermittency in the application as the fix may take some time to fully propagate. We sincerely apologize for the inconvenience and for falling short of the level of service you expect from us.

identified

We're still working hard to fix the internal issues that popped up after that recent problem with our third-party provider. Thanks for hanging in there and for your understanding while we work to get everything back to normal.

identified

We're letting you know that the issue with our third-party provider has been resolved. Because of this incident, we ran into some internal problems that impacted our services. Our team is actively working to get everything back to normal now. We appreciate your patience and understanding while we work to fully restore all services.

investigating

We are still waiting for Cloudflare's fix.

identified

We are still waiting for Cloudflare's fix.

identified

More details about the issue on their status page: https://www.cloudflarestatus.com/incidents/25r9t0vz99rp

identified

We are continuing to work on a fix for this issue.

identified

We are continuing to work on a fix for this issue.

identified

The issue has been identified and a fix is being implemented.

monitoring

A fix has been implemented and we are monitoring the results.

Report: "Report and filters slowness"

Last update
identified

We are continuing to work on a fix for this issue.

identified

Users are facing slowness when using reports and filters on their pipes. The issue was caused by a third-party provider, and our team is actively working to restore full functionality as quickly as possible.

Report: "Slowness in Pipefy"

Last update
resolved

This incident has been resolved.

identified

We are aware that some users may be experiencing slowness when extracting reports, applying filters, and using dashboards in Pipefy.

Report: "Slowness in Pipefy"

Last update
Resolved

This incident has been resolved.

Identified

We are aware that some users may be experiencing slowness when extracting reports, applying filters, and using dashboards in Pipefy.

Report: "Some users are not able to send E-mails"

Last update
postmortem

**Root Cause** Some users experienced issues sending emails due to a security measure that blocked certain content in the email body. This was intended to prevent specific types of data from being sent but inadvertently affected regular email usage. **Resolution** The issue was resolved by reversing the initial change and then adjusting the blocking rule to be more precise. This ensured that regular emails could be sent without disruption. **Action Plan** The initial blocking rule was reverted and replaced with a more targeted approach, improving error logging for easier troubleshooting. Ongoing evaluations will continue to refine the email blocking mechanisms to prevent future issues.

resolved

This incident has been resolved.

investigating

We are currently investigating this issue.

Report: "Slowness in filtering and generating reports"

Last update
postmortem

**Root Cause** We experienced issues with several nodes in a specific data center, leading to slow performance and intermittent access to reports. **Resolution** The affected nodes were replaced with new ones to restore normal operations. **Action Plan** Since the issue originated from the data center provider, no preventive actions can be taken on our side.

resolved

This incident has been resolved.

monitoring

The maintenance has been successfully completed and the fix has been applied. The platform is now stable, and we are actively monitoring to ensure everything remains operational.

identified

Our infrastructure team will perform maintenance today at 23:00 UTC to address the slowness issue and improve overall performance. We’ll provide further updates as the work progresses.

investigating

We are currently experiencing performance issues that may impact some users when filtering data or generating reports within Pipefy. Our engineering team has identified the cause and is working on a fix. We will provide updates as progress is made. Thank you for your understanding and patience!

Report: "Instability when opening some cards within Pipes due to a failure in the card's email feature."

Last update
resolved

This incident has been resolved.

monitoring

A fix has been implemented and we are monitoring the results.

identified

The issue has been identified and a fix is being implemented.

investigating

We are currently investigating this issue.

Report: "Instability when opening some cards within Pipes due to a failure in the card's email feature."

Last update
Resolved

This incident has been resolved.

Monitoring

A fix has been implemented and we are monitoring the results.

Identified

The issue has been identified and a fix is being implemented.

Investigating

We are currently investigating this issue.

Report: "Some users are experiencing issues when trying to filter cards by subtitle in Kanban view via advanced filters"

Last update
postmortem

**Root Cause** The advanced filters in the Kanban search bar were not working correctly due to a change in strategy. Initially, the feature was set for a small group of users, but a new setup enabled it for everyone, which might not have been the user's intention. **Resolution** The strategy was removed, and the feature was disabled for all users. After this, the issue was fixed, and the feature was gradually reintroduced. **Action Plan** To prevent similar issues, support will need prior approval before making changes. We will explore adding permission controls and separating operational features from product rollouts.

resolved

This incident has been resolved.

monitoring

A fix has been implemented and we are monitoring the results.

identified

The issue has been identified and a fix is being implemented.

investigating

We are currently investigating this issue.

Report: "Users are not able to log in, and automations with attachments are not working"

Last update
postmortem

**Root Cause** Users were unable to log in due to a security measure that blocked certain IP addresses. This was a preventive action taken because some authentication processes were incorrectly identifying the user's IP, leading to a block. **Resolution** The issue was resolved by removing the IP from the security rule after an incident was identified. This allowed users to log in again without any disruptions. **Action Plan** The plan includes changing the external DNS to direct requests internally, analyzing the authentication flow for better alert responses, reviewing login thresholds, and improving event logging. Additionally, updating support articles with relevant IP information is necessary.

resolved

This incident has been resolved.

monitoring

A fix has been implemented and we are monitoring the results.

investigating

Users are not able to log in, and automations with attachments are not working either

Report: "Pipefy is experiencing slowness, and our engineers are already working on it."

Last update
postmortem

**Root Cause** An Airflow task ran a database query without necessary filters, causing the system to become overloaded. This led to multiple attempts to execute the same heavy query simultaneously, which significantly impacted the database's performance and availability. **Resolution** The problematic queries were identified and stopped manually at the database level to alleviate the overload and restore normal operations. **Action Plan** Since this was a unique and isolated incident, no further actions are planned. The situation is considered unlikely to happen again.

resolved

This incident has been resolved.

monitoring

A fix has been implemented and we are monitoring the results.

identified

The issue has been identified and a fix is being implemented.

investigating

We are currently investigating this issue.

Report: "Emails not being sent"

Last update
postmortem

**Root Cause** The issue arose due to a failure in processing an email template, which happened because an account was compromised. This led to excessive memory usage and a failure in sending emails. **Resolution** The solution involved increasing the memory allocated to the email processing system to handle the issue effectively. **Action Plan** The plan includes consulting with development managers to determine the next steps for further investigation. The error logging system has recorded the incident details for future reference.

resolved

We've identified the problem, and emails are being sent again. We appreciate your patience.

investigating

We are continuing to investigate this issue.

investigating

Pipefy is facing an issue with emails. Our engineering team is investigating the problem and working diligently to resolve it as soon as possible. We understand how important this is for your operations, and we appreciate your patience during this time.

Report: "API parameter update.field.value, phase form and start form is not working properly"

Last update
postmortem

**Root Cause** Users experienced issues with our API due to a mismatch between the database function version and the code version. The discrepancy in data type definitions caused the malfunction. **Resolution** The issue was resolved by rolling back the function version to a previous state, which restored proper functionality to the API. **Action Plan** We plan to review and address database transactions within the code, develop a strategy for managing these transactions, and ensure consistency between production and development environments. Additionally, alerts will be created to identify related errors, and the feasibility of running tests post-migration will be assessed.

resolved

This incident has been resolved.

monitoring

A fix has been implemented and we are monitoring the results.

identified

A database update this morning caused an issue with the API. A manual revert to a previous version with the correct schema was completed, and the issue has been resolved.

investigating

Some users were unable to use the update.field.values parameter from our API, as well as phase forms or start forms.

Report: "Some users may be unable to access Pipefy."

Last update
postmortem

**Root Cause** Last Sunday, there was an issue with the database cluster due to a key rotation process. The Postgres operator rollout was not executed during the key rotation. Later, Kubernetes initiated the rollout, which led to the failover of the database instances. This caused some users to be unable to access Pipefy. **Resolution** The failover process, which is a built-in feature of the database cluster, had to be completed without interruption. The total downtime experienced was 4 minutes. **Action Plan** No immediate action was taken during the incident, as it was expected. However, the necessary actions should have been performed during the scheduled maintenance window on Sunday.

resolved

This incident has been resolved.

identified

Some users may be experiencing slowness and difficulty accessing Pipefy. We have identified the issue and are working to fix it as quickly as possible.

Report: "E-mails are not being sent"

Last update
postmortem

**Root Cause** Our domain was misused to send phishing emails, leading to a review of our email service account. To protect our domain's reputation, the email service provider temporarily disabled our account. **Resolution** We contacted the email service provider with details of our investigation and the steps we have taken to stop the malicious emails. We also informed them about setting up a new account to continue our operations smoothly. **Action Plan** We have implemented several measures to prevent future incidents. These include updating our monitoring systems to detect abuse, blocking harmful IP addresses, removing malicious accounts, and planning to blacklist harmful domains in the future.

resolved

Emails sent between 15:28 and 17:45 UTC may not have been delivered, so we recommend resending them. All other emails sent before or after that time are functioning normally. Thank you for your understanding.

identified

We are continuing to work on a fix for this issue.

identified

We experienced an issue with our third-party email provider, but we are already working on a solution to resume email sending as soon as possible

Report: "Card activities are not being shown"

Last update
postmortem

**Root Cause** The issue arose because a crucial security key was not set up correctly after a previous update. This oversight led to the system's storage becoming full, which caused disruptions in displaying card activities. **Resolution** To resolve the issue, we increased the storage capacity to bring the database back online. Additionally, the security key was updated to ensure that data processing could continue smoothly. **Action Plan** We have implemented new metrics to aid in future troubleshooting efforts. Further actions will be taken to prevent similar incidents from occurring.

resolved

This incident has been resolved.

monitoring

A fix has been implemented and we are monitoring the results.

identified

The issue has been identified and a fix is being implemented.

Report: "Emails sent by Pipefy are been sent to Spam. (Only for Microsoft-Based Email Users)"

Last update
resolved

This incident has been resolved.

identified

We are continuing to work on a fix for this issue.

identified

We are aware that some customers using Microsoft-based email services have received warnings about malicious links sent via Pipefy. Our engineering team is already working to resolve this issue as quickly as possible. Also if you are trying to send tasks by email, some users will not be able to open the link due to warnings about malicious links.

Report: "Some filters may not be working or may be inconsistent."

Last update
resolved

This incident has been resolved.

monitoring

The filters are now working again.

identified

We are reestablishing the connection to ensure that all information generated since 14:45 (UTC) is properly processed

investigating

This may impact features such as advanced search, reports, recurring automations that use filters and interfaces. We are still investigating the issue.

investigating

At 14:45 (UTC), our filters stopped updating some information due to a connection loss. We are still investigating the issue

Report: "Unable to Singup"

Last update
resolved

Users were not able to Signup

Report: "AI Automations doesn't working"

Last update
resolved

We kept monitoring OpenAI and it is stable again , so we will resolve the incident

monitoring

The OpenAI service is working again, so we have successfully processed all the pending AI automation jobs. Meanwhile, we are continuing to monitor the OpenAI operational status.

investigating

We are continuing to investigate this issue.

investigating

We have temporarily paused the AI Automation feature due to an incident with Open AI.

Report: "SOme users might not be able to login"

Last update
postmortem

**Root Cause** Some users were unable to log in because their accounts or IPs were flagged by an overly strict rule that blocked access based on shared IP usage. **Resolution** The restrictive rule was disabled to restore access, resolving the login issue. **Action Plan** The team refined the rule to better align with expected behavior and is monitoring its effectiveness to reduce false positives. They are also improving error handling, organizing a workshop on platform security, and evaluating ways to enhance privacy for specific attack vectors.

resolved

We have successfully updated our internal settings, and users are now able to log in again. More detailed information about the case will be provided in the post-mortem.

monitoring

We are continuing to monitor for any further issues.

monitoring

We noticed some users might have faced login redirects to a block page with the message "The owner of this website has banned you temporarily". We have updated our settings internally and are monitoring the results.

Report: "Pipefy down"

Last update
postmortem

**Root Cause** The primary database server encountered a memory issue, which triggered a recovery process. After the expected recovery time of 10 minutes, a system operator initiated a transfer to a backup server to restore functionality. **Resolution** Once the recovery time elapsed, the system operator successfully switched the operations to a backup server, ensuring the service was restored and operational. **Action Plan** The recovery time will be reduced to 3 minutes to minimize downtime in the future. Additionally, the database service monitoring will be enhanced to ensure prompt alerts in similar situations.

resolved

Pipefy was down (01:54 PM to 02:09 PM) for most of our customers. Now the application is normalized. The RCA is yet to be finished, we'll post more details further on our Post-Mortem.

Report: "Pipefy's website with latency"

Last update
resolved

Pipefy's website (www.pipefy.com) was facing latency and signup/signup for free wasn't working properly.

Report: "Filters are not working properly"

Last update
resolved

Due to an update in our plataform, it was not possible to use the filter in the pipe and in the reports. However, we were able to quickly find the root cause and reverted to the previous version, solving the issue.

Report: "Users weren't able to login into Pipefy"

Last update
postmortem

**Root Cause** A technical issue caused a pod outage, which affected the functioning of other pods, leading to a disruption in service. Users experienced problems logging into the system. **Resolution** The issue was resolved by implementing a rollout of the statefulset, restoring the affected services to normal operation. **Action Plan** A new alert system has been established to monitor and respond to any future restarts, ensuring prompt action to prevent similar incidents.

resolved

Users attempting to log in were encountering an Internal Server Error, while those already logged in were unaffected. Our infrastructure team implemented a rollout that successfully resolved the login issue.

Report: "Pipefy is down"

Last update
postmortem

**Root Cause** The migration performed on Sunday involved moving from one cluster to another, where some data types were changed and columns removed. This affected database statistics and execution plans, leading to poor performance. **Resolution** The solution involved increasing the sampling range significantly, which improved understanding and execution plans. Temporary measures included disabling a database feature globally and pre-loading indexes into memory. Once the permanent fix was applied, the feature was re-enabled. **Action Plan** Initial temporary solutions brought the system back online by disabling a database feature and caching indexes to improve queries. Sampling range was increased for better execution plans. Future actions include improving alerts to monitor database health, increasing load tests before migrations

resolved

With the changes implemented so far, the data retrieval process is smoothly running. RCA will be posted on the post mortem. All good now!

monitoring

Optimized data retrieval has been activated on one of our service replicas, and we're monitoring it. Pipefy performance is all good, but we want to make sure everything runs smoothly after this change.

monitoring

Pipefy's performance is all good now. We're now monitoring the performance of our queue management system to prevent unexpected loads.

monitoring

After setting the index of one of our tables in the cache memory, performance is now back to normal. We're monitoring Pipefy's performance for the next minutes.

monitoring

Systems are operational, but there is a delay in the performance, which translates to a slow navigation inside Pipefy. We're spending all the efforts to investigate the cause of this delay.

monitoring

For what we could test, Pipefy is now loading normally, and the navigation inside the app is running smooth. We'll keep monitoring it for a few more moments.

monitoring

A fix has been implemented and we are monitoring the results.

investigating

While latency persists, Pipefy is loading more smoothly. Our engineering team is actively identifying all resources needed to implement an effective solution to resolve the issue.

investigating

Right now, we are still encountering latency. Our team is dedicated to determinate the issues affecting the platform’s performance and is actively working on a effective solution.

investigating

At the moment, we continue to experience high latency and error messages while loading Pipefy. Our team is committed to uncovering the factors impacting the platform's performance, and we are focused on implementing a robust solution.

investigating

We want to assure you that our investigation into the current latency and errors is ongoing. Our engineers are meticulously examining the system to pinpoint the causes and implement necessary fixes.

investigating

Currently, there is still a high latency and errors showing on screen while loading Pipefy. Our team continues to actively investigating in details the whole platform to find a final solution.

investigating

Pages are now loading, but taking very long to load and showing errors on screen. Our team is actively investigating the situation to implement a solution as quickly as possible.

investigating

We are continuing the investigation to identify the root cause for this issue.

investigating

We are still trying to identify the reason that is causing Pipefy to be down/not loading.

investigating

We are currently investigating this issue.

Report: "Filters are not working properly"

Last update
postmortem

**Root Cause** The issue was caused by a significant update to the system, which led to filters not functioning as expected. This disruption affected the normal operation of the filters, causing inconvenience to users. **Resolution** To resolve the issue, we increased the number of backup systems to ensure smoother operation and prevent similar problems in the future. **Action Plan** We have taken steps to enhance system reliability by increasing backup systems. Additionally, a new monitoring alert has been set up to quickly identify and address any future issues.

resolved

Filters are back to normal now. The actions we took did solve the issue, but the root cause is yet to be found. We'll post more about the root cause in the post-mortem.

monitoring

We manually scaled our VMs to handle the queue, and it is normalized now. We're monitoring the next minutes.

investigating

There was a high load in a queue that manages updates in card attributes (e.g.: title, current phase and labels), and this load led to a delay in the processing of data inside the queue. Even though it is decreasing now, we are investigating the genesis of this event.

identified

For more context on the issue: when we apply a filter, matching cards are not being displayed. There are no pop-up error messages, but the expected results aren't show.

identified

When we filter for data inside Pipefy, it might not come as expected. We are looking for the root cause.

Report: "Some users might not be able to log into Pipefy"

Last update
postmortem

**Root Cause** A technical issue was identified within the cache cluster of the authenticator provider, stemming from previosuly established infrastructure limitations. These constraints hindered the system’s ability to effectively manage cached data, resulting in some users being unable to sign in **Resolution** The system was stabilized by restarting the entire cluster. Once stability was confirmed, further adjustments were made by increasing allocated infra to prevent future issues. **Action Plan** To enhance system reliability, allocated infra limitations were increased. Future steps include enhancing alerts to quickly identify and address potential problems.

resolved

After this rollout, logins are all good now. We'll post more details about this episode in the post-mortem.

monitoring

After a rollout in our infra, the login seems to be more stable. We're now monitoring the next login events.

identified

After a rollout in our infra, the login seems to be more stable. We're now monitoring.

identified

Our authentication tool is unstable, and this instability led to login issues. It works every now and then, but we're investigating more details on what could have caused this issue.

investigating

We are continuing to investigate this issue.

investigating

Users that are already logged in will not be affected, but users that are trying to login now may face difficulties. We're investigating the root cause, and we'll send updates as we progress.

Report: "Some features may experience a delay"

Last update
postmortem

**Root Cause** An incompatibility error occurred during our data replication process, preventing the database from updating properly. As a result, users experienced delays in both manual updates and automated processes. **Resolution** The data replication process was immediately stopped to avoid further problems. **Action Plan** Our alert system was enhanced to better control the acceptable levels of error in logs. The data replication process was adjusted to avoid the same error .

resolved

We encountered a configuration issue while attempting to replicate our database, which blocked partial execution and led to related errors. This impacted card deletions, asynchronous card updates, and automation processes, ultimately resulting in delays on the platform. The root cause of the error will be detailed in the post-mortem report. Additionally, we plan to enhance our alert system to better monitor acceptable error levels in the logs.

Report: "Dashboards and IA responses outdated"

Last update
postmortem

**Root Cause** After the upgrade, the new version failed to use the replication slot. **Resolution** A script was executed to correct the data for the period it was unavailable. **Action Plan** During the incident, a correction script was updated and run.

resolved

The issue has been resolved. The correction has been made and the data is being mirrored. We are still monitoring data performance and will provide further details in our post-mortem report.

monitoring

The correction has been applied, we'll still monitoring the data performance.

identified

With the fix underway, the data is gradually being updated. The dashboard and IA Response will still present outdated data until the correction is completed.

identified

We are applying a fix to resolve the issue. Our team is closely monitoring the situation and will provide more information soon.

identified

We are continuing to work on a fix for this issue.

identified

Due to some complications during the maintenance window, the data reflected in dashboards and IA responses are reflecting until 9 am GMT-3 today. We're working to solve this issue.

Report: "Pipefy is down"

Last update
postmortem

**Root Cause** The service experienced downtime due to a failover in our system's cache instance, which impacted our application. The underlying reason for this failover is still being investigated. **Resolution** The issue was addressed by reverting to a previous setup to restore service functionality. **Action Plan** We are actively communicating with our partners to identify the root cause. If no satisfactory response is received, we will reconsider our cache migration strategy. Additionally, we will enhance monitoring of our systems to prevent future occurrences.

resolved

The issue was with our third-party software. The connection was restored with a parallel fix. The root cause analysis is still in progress, and we will provide more details in our Post-Mortem.

investigating

A fix has been implemented and we are monitoring the results.

investigating

We are currently investigating this issue.

Report: "Filters are partially out of date"

Last update
postmortem

**Root Cause** The delay occurred because we failed to process a specific update, which was unable to be sent to the queue that later goes to the external system. This happened due to a connection field being incorrectly updated through the API. **Resolution** As several updates were taking place during the incident, we had to implement a solution in the application to deal with this exception. Essentially, the payload size is now validated before being sent to the queue, which solved the problem during the incident. Other actions have been mapped for further analysis. **Action Plan** We will create a disaster recovery process and resume discussion on alarm systems.

resolved

With the applied fix, there will be no more issues. Incident is now solved. We'll send more information on post-mortem in the next 2 business days,

monitoring

We're monitoring the queue. It is in a very sharp descent, and it'll soon be all good.

monitoring

Our graphs shows the fix results are coming. We're still monitoring it until it is all good.

monitoring

The fix has been implemented, and we're now monitoring the outcome of it. The fix was adding an extra validation step. Filters are not working yet because there's still a queue to handle, but it'll soon be all good.

identified

We understood the reason for the issue, and we're now working on the fix.

investigating

Filters are not bringing the results we've searched for. We're investigating the issue.

Report: "It is not possible to export reports, via downloads or email"

Last update
postmortem

**Root Cause** The reports stopped generating because we hit a limit in our data system. This happened because we didn't close the data search when an error occurred. The error was due to a problem with how numbers were being read, as some fields were text and not numbers, causing the search to remain open. **Resolution** We increased the data search size. Additionally, we reverted the code to a previous version. **Action Plan** We configured Elasticsearch to handle report data and will implement specific monitoring for reports.

resolved

The incident has been solved. Within two business days, we will publish the postmortem data along with the corrective actions to be implemented.

monitoring

We have successfully rolled back the code responsible for the failure, resolving the issue with exports. As a result, exporting reports is now functioning normally. We are actively monitoring the situation to prevent any further incidents.

identified

We performed an update, and some exports started working again. We are working on a definitive solution. We will give an update within 20 minutes.

identified

We have already identified the root cause and are working on the solution. We will give a new update within 20 minutes

identified

Our functionality to export reports is not working, either by download or by email. We don't have any workaround for this at the moment.

Report: "Pipefy might be inacessible for some users"

Last update
postmortem

**Root Cause** The outage was caused by emergency maintenance on a Third-party service that manages our data processing system. **Resolution** We started investigating the issue and soon after, the Third-party maintenance window was completed, restoring the service. **Action Plan** We initiated an investigation, and the service was restored shortly after. Also, we opened a ticket with the service provider to gather more information.

resolved

The issue was with our connection inside a third-party software. The connection was normalized, but the RCA is yet to be finished, we'll post more details on our Post-Mortem once we have the intel.

monitoring

A fix has been implemented and we are monitoring the results.

investigating

We are currently investigating this issue.

Report: "Inconsistent data in pipe reports"

Last update
postmortem

**Root Cause** The data source used by our apps (reports and filters) became out of sync with the database due to a weekend process required for compliance coupled with database connection issues. **Resolution** We adjusted the syncronization to skip changes from the compliance process and increased its throughput. Subsequently, a manual synchronization of all updated cards was performed to ensure everything was up to date. **Action Plan** Improve alerts to also encompass issues found on the incident, document the changes made for future reference, and improve and streamline the internal process that loads the application during weekends.

resolved

Reports and filters are up to date

identified

We are continuing to work on a fix for this issue.

identified

We are continuing to work on a fix for this issue.

identified

We are continuing to work on a fix for this issue.

identified

We are continuing to work on a fix for this issue.

identified

We are continuing to work on a fix for this issue.

identified

We are continuing to work on a fix for this issue.

identified

We are continuing to work on a fix for this issue.

identified

The issue has been identified and a fix is being implemented.

investigating

We are currently investigating this issue.

Report: "Pipefy is down"

Last update
postmortem

**Root Cause** The system experienced an overload because the implemented scalability strategy of the new model was insufficient to handle the increased load on our new configuration, which resulted in the database being overwhelmed. **Resolution** To solve the problem, a percentage of users were reverted to the application's default configuration and we manually set the scalability strategy temporarily **Action Plan** We have made enhancements to our system to improve its scalability and stability, including but not liited to optimizing the database access mode and adjusting the application scalability. Additionally, we have implemented additional monitoring systems to track performance and ensure stability effectively.

resolved

We will continue to internally monitor our infra load to ensure that all users can use the platform, but everything is back to normal

monitoring

We are monitoring our current infra load to ensure that all users can use the platform. Some webhooks might have faced delays and some users might be having slowness when clicking to open cards/pipes.

monitoring

We are monitoring our current infra load to ensure that all users can use the platform without latency. Some webhooks might have faced delays.

monitoring

A fix has been implemented and we are monitoring the results.

investigating

We are continuing to investigate this issue.

investigating

We are continuing to investigate this issue.

investigating

We are continuing to investigate this issue.

investigating

Some customers might face latency and 503 error pages

Report: "Pipefy experienced some instability"

Last update
resolved

Some customers experienced latency and 503 error pages

Report: "Pipefy down"

Last update
resolved

Now everything is normalizing and we're currently investiganting what happened

investigating

Now everything is normalizing and we're currently investiganting what happened

investigating

We are continuing to investigate this issue.

investigating

We are currently investigating this issue.

Report: "Pipefy is currently facing slowness"

Last update
postmortem

**Root Cause** The filters, reports, and cards were taking a long time to load due to excessive disk usage and increased search latency. **Resolution** We resolved the issue by rolling back to a previous state and reindexing the data from midnight until ten o'clock the following day. **Action Plan** Decrease backup latency to ensure faster rollback in case of need.

resolved

This incident has been resolved.

monitoring

We are continuing to monitor for any further issues.

monitoring

A fix has been implemented and we are monitoring the results.

identified

The issue has been identified and a fix is being implemented.

Report: "Some users are not able to login"

Last update
postmortem

**Root Cause** A change in database permissions, intended to enhance security, inadvertently restricted some users from accessing our platform. **Resolution** Database permissions were adjusted to ensure all users would be able to access the platform **Action Plan** We will implement an automatatic routine to fuirther validate infrastructure permission changes, and create a monitoring dashboard to track these errors.

resolved

This incident has been resolved.

monitoring

A fix has been implemented and we are monitoring the results.

investigating

We are continuing to investigate this issue.

investigating

We are continuing to investigate the issue.

investigating

We are still investigating the issue.

investigating

We are continuing investigating the issue

investigating

We are currently investigating the issue

Report: "New created cards are not appearing in filters and reports"

Last update
postmortem

**Root Cause** The issue was caused by a lack of validation in the API mutation regarding the blank value sent, which led to a lag on data replication. **Resolution** A rollback to the previous version of data replication process was performed. A long-term solution involves fixing the field to handle invalid values and redeploying the new flow. **Action Plan** We'll handle the validation of invalid values, and improve our monitoring system related to events like these.

resolved

The fix has been applied, and the filters are now working fine.

monitoring

A fix has been applied, and we're now monitoring.

identified

We've identified the issue occurs only to a few customer and not the whole base, along with we have found the root cause and we are implementing a fix for it.

investigating

We are currently investigating this issue.

Report: "Pipefy is facing a slowness"

Last update
postmortem

**Root Cause** our autoscaler experienced a prolonged restart process, causing other components of the system to fail to scale appropriately in response to platform usage. This resulted in a degradation of platform performance, impacting user experience. **Resolution** Initial steps involved manual scaling to improve performance. Further investigation led to adjustments in system parameters for stability. **Action Plan** A monitoring system will be developed to oversee scaling operations, ensuring smoother platform performance.

resolved

The incident has been successfully resolved.

monitoring

We've applied the fix, and we're now monitoring it.

identified

We've identified the cause of the issue, and we're now arranging the fix.

investigating

We are currently investigating this issue.

Report: "Pipefy website is offline"

Last update
postmortem

**Root Cause** The issue was due to a deactivated SSL certificate, impacting our website's availability. Unfortunately, specific details are unavailable due to data retention policies. **Resolution** The resolution involved reactivating the SSL certificate to restore service. **Action Plan** We've completed a contingency plan to ensure service continuity and implemented a monitoring system for certificate expiration to maintain secure, uninterrupted service.

resolved

The fix has been applied, and after our monitoring and checking everything is back to normal

monitoring

Fix has been applied, and we're now monitoring it.

identified

We've already identified the issue, and the team is working on resolving it as quickly as possible.

investigating

We are currently investigating the issue

investigating

Pipefy website (pipefy.com) is offline. The app (app.pipefy.com) is working normally.

Report: "We're facing some instability"

Last update
postmortem

**Root Cause** During the setup of our connection manager, the specified image in the configuration file was not present in the container registry. Consequently, when attempting to launch pods, they encountered issues due to the unavailability of the required image. **Resolution** The issue was resolved by reverting to previous settings, improving the changes in settings process to ensure stability. **Action Plan** All necessary details within the connection manager setup have been arranged to facilitate future deployments.

resolved

This incident has been resolved.

monitoring

A fix has been implemented and we are monitoring the results.

identified

We are continuing to work on a fix for this issue.

identified

We've identified the root cause. Working on the fix.

investigating

We are currently investigating this issue.

Report: "Pipefy might be inacessible for some users"

Last update
postmortem

**Root Cause** The database health check failed, prompting an automatic failover that unfortunately did not execute as anticipated. As a result, an unplanned service downtime occurred. **Resolution** To resolve the issue, system adjustments were made to manage database connections and ensure a smooth transition to a new master node. **Action Plan** Improve system monitoring with additional metrics for better oversight. Adjusted the health check and failover parameters.

resolved

This incident has been resolved.

monitoring

A fix has been applied to resolve the outage and Pipefy is now available to all users.

identified

The technical team is still working to restore full access to all users as quickly as possible. Some user might still experience degraded performance.

identified

The technical team is still working to restore full access to all users as quickly as possible.

identified

The technical team is still working to restore full access to all users as quickly as possible.

identified

The technical team has identified the causes of the system unavailability and is currently working to restore full access to all users as quickly as possible.

investigating

We are continuing to investigate this issue.

investigating

We are currently investigating this issue.

Report: "Pipefy is facing some latency"

Last update
postmortem

**Root Cause** A database setup not fully prepared for load balancing, which caused some users to see outdated information. **Resolution** To ensure data accuracy, we adjusted our database connections to only utilize the primary source. **Action Plan** We've implemented additional monitoring for data replication processes to enhance system reliability.

resolved

This incident has been resolved.

monitoring

Performance should be back to normal

identified

Some users might still be experiencing issues moving, creating and fillin in fields in cards

identified

The previous incident resulted in an increased latency . Some users might experienced issues moving, creating and filtering cards

Report: "Pipefy is facing a slowness"

Last update
postmortem

**Root Cause** A task initiated caused an overload, leading to system slowness. **Resolution** The issue was resolved by removing the source causing the overload and addressing the root cause to prevent future occurrences. **Action Plan** Implemented a solution to block problematic processes and explored enhancements in our process to improve system resilience.

resolved

This incident has been successfully resolved. All good!

monitoring

The fix has been applied, and we're now monitoring the post-action events.

identified

We've identified the reason, and we're working on the fix. Automations may have a delay of 1 minute to run.

investigating

We are currently investigating the issue

Report: "Some feature might be temporarily unavailable"

Last update
postmortem

**Root Cause** A feature was temporarily disabled and re-enabled, causing all settings to reset. This was done without realizing the impact it would have. **Resolution** Immediate steps were taken to manually restore the most critical settings for customers, and a backup was used to fully recover the system. **Action Plan** Reviewing internal processes, evaluating essential features, considering an enterprise license, and enhancing feature management governance.

resolved

The fix has been applied, and after our monitoring and checking everything is back to normal

monitoring

We are continuing to monitor for any further issues.

monitoring

Fix has been applied, and we're now monitoring it.

identified

We understood the issue, and we're now taking actions to monitor what's next.

investigating

We are still investigating the issue.

investigating

We realized that some features may not be available, and we're now investigating what happened to work on the fix.

Report: "Pipefy experienced some instability"

Last update
postmortem

**Root Cause** The database health check failed which caused reduced operational capacity. This led to a short downtime. **Resolution** To resolve the issue, automatic system adjustments were made to manage database connections and ensure a smooth transition to a new master node. **Action Plan** Enhanced system monitoring with additional metrics to improve observability and ensure system reliability.

resolved

From 13h30 to 13h35 some users migh have experinced a loading issue in Pipefy caused by an instability in our infrastructure. The issue was quickly mitigated and everything is back to normal.

Report: "PIpefy down"

Last update
resolved

An update was implemented to our system infrastructure last night. Contrary to expectations, these modifications did not transpire as planned, resulting in a brief interruption of platform services for an estimated duration of 5 minutes around 01:00 AM. As soon as the issue was identified, the update was promptly reverted, restoring all systems to their normal operational status.

Report: "White screen when opening public phase form"

Last update
postmortem

**Root Cause** Front end received an unexpected response from the platform and could not handle it correctly. **Resolution** Production was updated to a previous version. **Action Plan** We will discuss further regarding the different formats used on the front-end.

resolved

This incident has been resolved.

identified

The issue has been identified and a fix is being implemented.

investigating

We are currently investigating this issue.

Report: "Pipefy home page might be unacessible for a few orgs"

Last update
postmortem

**Root Cause** An a/b configuration test was made on an infrastructure level that impacted the usage of the home page, dashboards, calendar and sprint for isolated and internal users **Resolution** The test was manually removed from the palatform **Action Plan** We added more steps and approvals to the a/b test configuration deployment and enhanced our testing steps to ensure effective incident communication.

resolved

This incident has been resolved.

monitoring

A fix has been implemented and we are monitoring the results.

investigating

Some users and organizations might be receiving 503 page error on the home page

investigating

Some users and organizations might be receiving 503 page error

Report: "Trouble using the filter when searching for new created cards"

Last update
postmortem

**Root Cause** Every update made on cards were not being properly registered, and thus, the filters inside Pipefy couldn't see what was happening. **Resolution** We restarted the system that stores all the data, and then all the updates started coming to the right place. **Action Plan**

resolved

The system instability has been fixed and the performance of the platform has been restored. As soon as the investigation process is over, we’ll share further details about the causes, implemented fixes and preventive actions to be implemented.

identified

We are still investigating the issue, and one of our actions might result in cards not appearing in pipe reports, specially for cards created after 8:30 pm (UTC 0). It shall normalize soon.

identified

The issue has been identified and a fix is being implemented.

investigating

We are continuing to investigate this issue.

investigating

Pipefy's technical team is still investigating the causes of the issue. A final diagnosis has yet to be established. We regret the inconvenience and assure you that we are working with all our resources to restore normal access and isolate the causes.

investigating

The team is still investigating the problem.

investigating

We are currently investigating this issue.

Report: "Not able to import any file via Importer"

Last update
postmortem

**Root Cause** This incident was caused by a migration of the APP Importer to a new environment. **Resolution** The resolution was made by reverting the Migration. **Action Plan**

resolved

The Importer feature instability has been fixed and the performance has been restored. As soon as the investigation process is over, we’ll share further details about the causes, implemented fixes and preventive actions to be implemented.

monitoring

The unavailability has been fixed and the access to Pipefy’s web and Importer has been restored. We are currently monitoring the system to ensure all features are working as expected. As soon as the monitoring process is over we’ll share further details about the causes, investigation and preventive actions to be implemented.

identified

Pipefy’s technical team is investigating the causes of the platform outage. We’re sorry about the inconvenience the outage has caused and we ensure we’re working with all our resources to restore normal access and isolate the causes of this incident.