Webinar - Page Speed - Who's Doing it Well?
Join Spike’s Head of Performance, Mike Forshaw as we dive into our analysis of the winners and losers on e-commerce website page speed. Mike will...
As performance testing becomes more data-intensive, analysing the results quickly and accurately is crucial for maintaining optimal website speed and user experience. In this blog, I explore how AI chatbots can support this process by comparing several popular options for their ability to analyse a real set of performance testing results. From ease of use and response times to the depth of insights provided, this comparison reveals which AI tools stand out as reliable partners in performance analysis—and which fall short.
Let’s dive into the findings and see which AI chatbot comes out on top.
This article explores the capabilities of various AI chatbots in analysing performance testing results. By comparing tools like ChatGPT, DeepAI, Gemini, Meta AI, and Claude AI, the analysis highlights their strengths, weaknesses, and overall effectiveness in providing insights and recommendations.
All AI tools provided valuable insights and recommendations, but each had its own set of challenges. ChatGPT and Claude AI stood out for their detailed analysis and useful suggestions, while DeepAI and Meta AI offered solid performance with some limitations. Gemini had some issues with date formats and vague recommendations but still provided useful observations. Overall, these AI tools can serve as reliable partners in performance analysis, especially for small data sets.
I started off by creating a very simple JMeter test running against the Spike website at www.wearespike.co.uk. Nothing complicated, just a few users each hitting the various pages of the site. Once I had my base results, I then manually created a few more sets of results and stored them in an Excel spreadsheet, see below. Each set of results presented progressively slower response times apart from the final set which I adjusted to give improved response times.
The results set:
At this point, I started exploring the various AI chatbots available. I only wanted to use free ones for this exercise. It’s important that a tool is confident enough in its own ability that it can give you a decent free offering before it starts to charge. I asked each tool the exact same question and if possible didn’t provide any further prompts.
The prompt I gave each tool was as follows:
"I'm going to upload you a file with several sets of performance test results over a number of days. These results are from a test I've been running against our company website. I'd like you to analyse the results and provide me with any trends you see, areas of concern, areas to improve etc. Are you comfortable with that? I only want you to consider columns A to N. Can you do this?"
Almost all the tools I asked were more than happy to help in the first instance but as soon as I uploaded the file or copied and pasted the data, they told me I’d need to upgrade before they would “help me.” Those tools were immediately discounted. Finally, I was left with the following set of willing helpers:
Let’s take a quick look at how each one did.
ChatGPT is an advanced AI language model developed by OpenAI that generates human-like text responses, making it useful for a variety of applications such as conversation, content creation, and data analysis. Its capabilities include understanding context, providing information, and engaging in interactive dialogue.
DeepAI is an AI platform that offers various tools for generating text, images, and other content, emphasizing accessibility and user-friendliness. Its API allows developers to integrate AI capabilities into applications, facilitating creative and analytical tasks.
Gemini, developed by Google DeepMind, is an advanced AI chatbot designed to understand complex queries and provide insightful responses, with a focus on integrating deep learning for enhanced contextual comprehension and user interaction.
Then came Meta AI, and I was expecting big things….
Meta AI is a conversational assistant developed by Meta, designed to provide real-time information and answer questions, with advanced natural language understanding and integration across Meta's platforms.
Claude AI, developed by Anthropic, is an AI assistant focused on delivering safe, reliable, and conversational support, optimized for in-depth reasoning and complex task assistance.
All in all, I was very impressed with what I saw from each of the AI tools. It’s certainly not the most comprehensive performance test analysis I’ve ever seen but it was all generated in a matter of seconds without any fuss….apart from ChatGPT who was obviously having a bad morning. Granted this was only a small data set but I can’t see why it wouldn’t work on much larger results files and it certainly gives you a good starting point. The errors from Gemini make me a tad nervous but you would always expect to check the outputs before you acted upon them.
I’d be interested to hear your thoughts and experiences. Have I missed any good tools that you can recommend? I’d certainly like to see how much more the paid options could give you and if they are worth the subscription costs.
Join Spike’s Head of Performance, Mike Forshaw as we dive into our analysis of the winners and losers on e-commerce website page speed. Mike will...
For most ecommerce sites 2020 has allowed an ever rising influx of customers to flock online and buy pretty much everything they want or need as...
To quote Top Gun: “I feel the need … the need for speed!” This is what every business and IT leader should be thinking about when it comes to...
Core Web Vitals: What is it and what the hell does it mean for you? If you missed the live session you can view the recording below!