CloudWatch Monitoring & Troubleshooting experience

I joined Amazon CloudWatch as one of the senior designers, tasked with the mission of redefining the entire Monitoring and Troubleshooting Experience.


CloudWatch itself is a monitoring, analytics and data visualisation platform for software developers to monitor their applications – and also for management to monitor trends and costs and shape their business strategy. 


CloudWatch runs one of the largest data stores on the planet, and we are solving problems of massive scale. AWS CloudWatch is used by leading businesses like Netflix, LinkedIn, Facebook, WarnerMedia, BBC, Adobe, Twitter etc. 


We have 3 million weekly active accounts – the scale is very big.


I was brought into the team to focus on multiple retention, activation and acquisition metrics. How could we improve a product for engineers that would make them want to use Amazon CloudWatch for Monitoring and Troubleshooting instead of other competitors? I introduced various creative solutions. I pushed our team to prioritise metrics and to understand what a ‘golden path’ might be for our customers.



I ran a workshop to align existing resources and KPIs between the product team and UX team at the start of the project.

We focussed on these metrics


Activation and engagements:

  • Number of created dashboards.
  • Number of created alarms.
  • Funnels completions rate.
  • Average time spent to close a ticket.
  • Average time spent in the platform per user.


  • Number of new AWS accounts
  • Incoming traffic sources


I redefined CloudWatch archetypes, mapped user journeys, validated creative proposals with customers and stakeholders, and defined the MVP.

Engineer journey – Start from Onboarding CloudWatch to the root cause solution 



Main insights

Customers need to find relevant data quickly, they need to be able to correlate, analyse, and compare data points. CloudWatch currently offers a fragmented experience. Customers today jump between services, regions and applications, losing their context and slowing down their root analysis.


I ran a series of experiments designed to find metrics that defined user retention, acquisition and activation.


I launched 3 new features to ensure full stack observability – one place to find everything. I improved existing flows, interactions and data correlation patterns.


I tested and iterated high-fidelity prototypes, and A/B test small improvements, supporting developers to guarantee the best quality.





I analysed feedback after launching the beta version to a group of selected customers, and tracked metrics on the live platform.


3 weeks after launch, we calculated we. had 20,000 new dashboards and 2,000+ new alarm aggregations. We reduced the drop-off to 11% with an increase in positive feedback of 7%.


At the end of the quarter, I put together a data report to communicate the next steps and share the new proposal.




Peer and Manager quotes


“Simona is always willing to improve CloudWatch to satisfy customers. She dives deep to understand customers’ behaviors and provides insightful research reports to backup her decisions. I was very impressed by her alarms research and metrics rework. I like her curious mind and, for example, all the inspiring stuff she shared with us about data visualization.”

— Engineering Manager


“Simona is a real voice of the customer. Her decisions are based on research insights, data and customer feedback. She is always driving the teams to deliver high quality results, while both managing the projects and staying connected to details”

— Sr. Product Manager


“Simona is a true advocate for Customer Obsession, putting the needs of the customer first with excellent communication to develop her point. This can often be difficult with the level of complexity with certain products, which she takes in her stride. She is always open to feedback and I’m  excited to see her impact/contribution to the entire Console going forward”

— UX Manager