Comment by SeanAnderson
2 years ago
I asked Bard, "Are you running Gemini Pro now?" And it told me, "Unfortunately, your question is ambiguous. "Gemini Pro" could refer to..." and listed a bunch of irrelevant stuff.
Is Bard not using Gemini Pro at time of writing? The blog post says, "Starting today, Bard will use a fine-tuned version of Gemini Pro for more advanced reasoning, planning, understanding and more."
(EDIT: it is... gave me a correct answer after some pushing. I just had to clarify that I meant "Bard". Not off to a great start)
It's weird, right? If I ask it:
Q: "Who are you?" A: "I am bard." Q: "Is Bard running Gemini Pro?" A: "Yes." Q: "Are you running Gemini Pro?" A: "Ambiguous Question."
Your line of thinking also presupposes that Bard is self aware about that type of thing. You could also ask it what programming language it's written in, but that doesn't mean it knows and/or will answer you.
This is a common occurance I'm seeing lately. People treating these things as oracles and going straight to chatgpt/bard instead of thinking or researching for themselves
I consider it a standard test because no self-respecting PM would allow the product to ship without being able to market itself correctly. There's a reason the seed prompt says, "You are Bard."
I don't lack awareness of the limitations of pretrained models. I'm evaluating its ability to employ chain of reasoning, in combination with its plugins, to get me an obvious answer.
I had the same issue as OP. Initially Bard seemed clueless about Gemini, then:
Me: I see. Google made an announcment today saying that Bard was now using a fine-tuned version of their "Gemini" model
Bard: That's correct! As of December 6, 2023, I am using a fine-tuned version of Google's Gemini model ...
So Bard found the blog post from Google and returned the information in it. No new information was get.
The LLM itself does not KNOW anything.
1 reply →
It has access to the Internet and is free to search for the right answer.
If I ask it who it is, it says it is Bard. It is aware of the launch that occurred today. It cites December 6th.
It just very incorrectly felt that I was asking an ambiguous question until I restate the same question again. It's not great.
It forgets previous prompts and answers. I have to specifically ask it to relate to those and take those into consideration.
Knowing it’s own build information is something that could be trained into the model right? Seems like a good idea.
Bard shows “PaLM2” in my answers, and it says “I can't create images yet so I'm not able to help you with that” when I ask it to do so, which Gemini ought to be able to since its transformer can output images.
I don’t think Bard is using Gemini Pro, perhaps because the rollout will be slow, but it is a bit of a blunder on Google’s part to indicate that it now uses it, since many will believe that this is the quality that Gemini assumes.
https://bard.google.com/updates The bard updates page says it was updated to Pro today. If it's not on Pro, but the updates page has an entry, then IDK what to say.
The latest update I see on that link is from 2023.11.21 on “Expanding Bard’s understanding of YouTube videos.” Is it possible that there are country-specific update pages?
Edit: I found a Google page indicating that Gemini Pro is not accessible in Europe, with no ETA for when it will be released there: https://support.google.com/bard/answer/14294096
Mmmh last update I see is 2023.11.21 and it's about better YouTube understanding.
Interensting, for me it says the last update was 2023.11.16. The language is russian.
I assume that's the reason you people have different info\results.
I asked it to draw a cute kitten for me and it used Python do the task.
https://i.imgur.com/VhnQYie.png
So powerful.
That's actually a lot better I ever received from gpt4, but it's been a while I tried svg/python output for graphics
2 replies →
It's not the type of questions to ask an LLM, and yet I see this everywhere.
Came here to say exactly this. My conversation with Bard:
```
do you use Lambda model?
I use a variety of large language models, including the Lambda model, to perform my tasks. However, for each specific task, the best model is chosen based on its strengths and weaknesses. The Lambda model is particularly good at tasks that require understanding and responding to complex and nuanced language, such as generating creative text formats of text content and answering open ended, challenging, or strange questions. However, it may not be the best choice for tasks that require high accuracy or low latency.
Here's a breakdown of the models I use:
LaMDA: This is a family of Transformer-based neural language models specialized for dialogue. It is particularly good at understanding and responding to complex and nuanced language, and can generate different creative text formats of text content.
PaLM: This is another large language model, similar to LaMDA, but with a focus on factual accuracy. It is best suited for tasks that require precise and reliable information retrieval.
Meena: This is an older, smaller language model that was the precursor to LaMDA. It is still used for some tasks that require a more lightweight model. Other models: I also have access to a variety of other specialized language models, such as those for summarization, translation, and question answering. The specific model that is used for each task is determined by a variety of factors, including the nature of the task, the available resources, and the desired level of performance.
```
I wonder the same too, also is Gemini Pro supposed to beat GPT-4?
I thought so too but no, Gemini _Ultra_ is supposed to beat GPT-4. The one available through Bard is Gemini Pro. The wording sure is confusing!
Looking at the technical report [1], on selected benchmarks, rounded score in %:
Dataset | Gemini Ultra | Gemini Pro | GPT-4
MMLU | 90 | 79 | 87
BIG-Bench-Hard | 84 | 75 | 83
HellaSwag | 88 | 85 | 95
Natural2Code | 75 | 70 | 74
WMT23 | 74 | 72 | 74
[1] https://storage.googleapis.com/deepmind-media/gemini/gemini_...
For the record, GPT-4 still thinks it's GPT-3.
Can you share with me a conversation link that results in this behavior? I do not experience what you're claiming.
"Are you GPT-4?": https://chat.openai.com/share/1786f290-4431-45b0-856e-265b38...
"Are you GPT-3?": https://chat.openai.com/share/00c89b4c-1313-468d-a752-a1e7bb...
"What version of GPT are you?": https://chat.openai.com/share/6e52aec0-07c1-44d6-a1d3-0d0f88...
"What are you?" + "Be more specific.": https://chat.openai.com/share/02ed8e5f-d349-471b-806a-7e3430...
All these prompts yield correct answers.
I was talking about GPT-4, not ChatGPT. ChatGPT has received additional fine-tuning over GPT-4. Repeat your experiments with the GPT-4 API.