ChatGPT vs. Superusers – What Our Data Showed
Over the past month, we’ve been putting ChatGPT to the test to answer two questions.
1) How might ChatGPT threaten the work of community professionals?
2) How might ChatGPT be useful to the work of community professionals?
The Limitations of ChatGPT
Before we provide answers, let’s highlight the limitations of ChatGPT.
I like Benedict Evans description of AI as: “a ten-year-old that’s read every book in the library and can repeat stuff back to you, but a little garbled and with no idea that Jonathan Swift wasn’t actually proposing, modestly, a new source of income for the poor.”
Specifically, ChatGPT suffers from the following:
1) It’s not able to understand anything. It essentially makes predictions. When you give it a prompt, it gives a set of words and sentences which matches the input with the highest probability.
2) It’s simply attempting to give the most “plausible” answer. This isn’t mean it’s the right answer – just the one which best matches the pattern.
3) It can’t evaluate competing sources of information. The lack of subject matter expertise means it lacks the ability to evaluate the quality of sources.
4) It doesn’t acknowledge uncertainty. ChatGPT will provide a statement of fact without acknowledging any degree of uncertainty. At present, you can’t find the sources of information to evaluate.
Let’s take a look at some examples.
Methodology
To help answer this question, we undertook a few simple tests.
1) Comparing existing answers. We found questions with an accepted solution and posted those same questions to Chat-GPT to evaluate if it could provide a comparable, if not better, answer.
2) Creating new answers. We posted unanswered questions into Chat-GPT and published the answer in the community. We then let members highlight if they found the answer useful or not.
Before we explore the results, it’s useful to highlight the methodological problems.
Comparing ChatGPT responses to accepted solutions suffers from a verification problem. We don’t have the expertise to evaluate whether one response is better than another – but we can look at the nature and structure of the response. Yet, this might suffer from subjective bias. We’re also, by definition, comparing ChatGPT to the ‘best’ answers – not the typical member response.
Creating and publishing answers to unanswered questions in the community also has issues. By nature, if the question is unanswered, it might be a more difficult question to answer. This creates a bias against ChatGPT.
However, those issues aside, we feel we can draw a few specific conclusions from our experiments which are worth sharing.
Does ChatGPT Provide Better Answers Than Top Members?
It’s not always easy to be definitive, but here are some of our experiments.
Using this question on the Spotify Community as an example.
This is a very precise question – but its explanation is somewhat long-winded.
We compared the Accepted Solution with the Chat GPT response below.
This answer perhaps best showcases the challenge in comparing the quality of answers.
The accepted solution suggests using private mode to prevent influencing music recommendations. The author also suggests training the algorithm by liking/disliking songs and the Spotify Kids app. The use of a video is useful. However, the explicit content part of the question is ignored.
The ChatGPT answer suggests removing previously listened to songs from playlists – which seems like a lot more work. It also wouldn’t immediately affect future playlists as much as past playlists. However, it does include specific steps to filter explicit content from kids.
On balance, we judge the accepted solution better because the author understood the need to train the algorithm for the future.
ChatGPT Can Provide Better Answers Than Top Members
Let’s try this example from the SAP community.
We can compare the an answer from ChatGPT against an answer from the community below:
In this example, ChatGPT is giving a different answer to that provided by the solution. It’s not easy to independently check if it’s better, but at a glance, it seems to be a better answer than the one marked as Best Answer (also by the OP).
Thus it is possible for ChatGPT to provide better answers than top members.
ChatGPT Doesn’t Presume Expertise
Let’s use one more example from the Tableau community.
The question above is looking to create a shape based upon whether the table has been refreshed within the past four hours or not.
The question includes images (which ChatGPT can’t read) and a workbook (which it potentially can). I copied and pasted the workbook into a dropbox folder and linked to this. This was the result.
Again we have two slightly different answers here. The code for both seems to be accurate (although each used a slightly different approach). But the ChatGPT answer offers more information – especially explaining how to create shapes.
The difference here is a presumption of knowledge.
The community answer presumes the person asking the question has a certain level of expertise, the ChatGPT answer doesn’t. The former might be better for a member more familiar with Tableau, the latter might be better for someone that isn’t.
ChatGPT Can’t Share Experiences
Let’s try an example from Mayo Clinic.
The question is asking for opinions on two different medications.
There isn’t a ‘best answer’ on the platform – so we went with an answer near the top which had the most reactions (i.e. the one most people are likely to see if they visit the question).
The community answer is the clear-cut winner. Not only that but there are 48 pages(!) of responses to review the aggregated experiences of many, many, people. As noted in the ChatGPT answer, it’s only capable of giving a generic answer about the drugs. It can’t share the experiences of either of them.
ChatGPT Could Serve As A Good Assistant To Superusers
So, does ChatGPT provide better answers than top members?
Possibly!
It can certainly provide more complete answers and alternative approaches to solving problems. But it can’t understand anything related to the context.
The best use of ChatGPT right now is perhaps to encourage superusers to use it to generate an answer which they can then edit to the specifics of the question.
Does ChatGPT Provide Comparable Answers to Members?
For this test, we found 20 unanswered questions in the Tableau community over several weeks.
We put the questions into ChatGPT and copied/pasted the results without any editing.
Then we waited to see how members would respond and engage with the questions.
You can see the results in the table below:








