ChatGPT vs BARD: Explain "There are 10 types of people in this world, those who understand binary and who don't"
In recent years, large language models (#llms) have become increasingly popular. These models are trained on massive amounts of text data, and they can be used to generate text, translate languages, write different kinds of creative content, and answer your questions in an informative way.
Two of the most popular LLMs are #BARD and #ChatGPT (powered by #gpt ). BARD is an LLM from #Google AI, and ChatGPT is an LLM from #OpenAI.
Both models are very powerful and capable of intelligently answer a wide range of questions. However, there are some key differences between the two models. BARD is presumably trained on a dataset of text and code that is specifically curated for LLMs, while ChatGPT is trained on a huge dataset of text and code that is more general. This means that BARD is likely to be better at generating text that is relevant to specific topics, while ChatGPT is likely to be better at generating text that is creative and original.
Another key difference between the two models is that BARD is still under development and introduced as better at generating less offensive and hallucinating responses (#responsibleai), while ChatGPT is already available to the public for some months and in some come specific contexts shown to generate hallucinating responses. The extent of such characteristics is estimated to be below 20% but it varies from experiment to experiment.
I was also curious to try it out on the questions/prompt that I had in my mind and have the 1st hand experience. I divided the whole exercise into 4 sections
Subjective Ag marketing domain questions/prompts
Questions/Prompts on
It was difficult to come up with a clear choice but surely, ChatGPT seems to be more appealing with less sensitivity?to the choice of words and understanding the context of a question. One good example was the use of the word "invest", where BARD provided information on investing companies but ChatGPT could understand the word invest is not only referring to Venture Capitalists?(VCs). Once "investing" was changed to "working", BARD provided comparable answers to ChatGPT. Once identified this choice of word sensitivity, I played a bit more with words and my advice to work with BARD is to be a bit more precise with the choice of words for prompting.
Subjective domain unspecific questions/prompts
Questions/Prompts on
Again, both did a good job but BARD was good at reasoning and provided more "Expert Human" like responses. Logically, step-wise approach to explaining a catastrophic scenario and why it will happen and what impact it will create. ChatGPT was quicker in providing outcomes than explaining the logical steps before concluding. Explaining a funny prompt on the probability of spending -10 dollars in the next 10 days was another extreme but, I could validate my previous findings about both models. So, my advice for such questions is to go with BARD if you want to understand the step-wise reasoning behind a solution and if you just want an answer with a "good enough" explanation ChatGPT is the model to go.
领英推荐
Explaining a Joke or some teaser (light fun)
My previous finding also confirmed that BARD is more detail-oriented and provides step-wise reasoning while proving you the answer to your questions (given that it gets it right :) ) on the other hand ChatGPT is also providing you with the "correct" answer but with little less explanation.
Fact-based domain-specific questions/prompts
This part was the most interesting for me as an expert in the field, I was curious to see the accuracy of the responses. I started with some questions based on the fundamental understanding of the field and went towards questions on the well-established concept that are sometimes mixed by even experts in the field.
Some of the questions listed above were part of my question list but I had to stop the exercise after realizing that both models have limited capabilities of understanding and make "Expert-like" comments when it comes to a deep understanding of a specific field of study. Funny enough, the questions mentioned above are too basic for an expert in the Plant developmental Biology discipline. Both models frequently generated contradicting information in many cases due to the potential cases of misunderstanding?the references. Questions like "Who discovered FT in Arabidopsis thaliana and when" should be simple. I could sense that both models need a lot of domain-specific learning to be a good assistant to a researcher or SMEs. The responses were promising but not "good enough" as per "Research & Development Grade". I have attached screenshots of some of my interesting prompts below. Full session history will be provided on request.
In summary, It was a great fun exercise (2.5 hours nearly) to get some understanding of these models and while working with these models also realized the importance of using the right #prompts (#powerofquestions) to get the desired answers. I will keep exploring these models and will provide my take on these for various scenarios.
PS: Opinion expressed in this article belongs to me as an individual and do not represent opinion of any organization, I am part of.
Data Engineering and Science at BASF
1 年Dr. Vimal Rawat Thanks for this insightful post. Surely enough as you said, these LLMs are not yet ready to be a buddy of researchers because they lack domain specific learning. They are surely evolving but until then they need to be fine tuned to cater specific business needs by say expanding datasets of a pre trained model. Pre-trained model + Quality Datasets + meaningful prompts = Closer to accurate responses