Ask HN: 6 months later. How is Bard doing?
|
Overall, Google is doing a least a B+ effort in response to the GPT4 buzz. They already had deep experience and expertise with AI, but hadn’t productized it much. In a barrage of blog posts and announcements over the past few months they release new features into nearly every product. I have the Search Generative Experience (generated results above main search results) pretty useful about 20% of the time and easy enough to skip when it’s not useful.
I’ve used Bard quite a few times successfully for code generation, though it did give some bad curl commands (which I found the source blog post for). Because Google has a very favorable brand reputation (despite what some on HN think) and gets a lot of legal scrutiny, they have to be much more careful in ways that OpenAI doesn’t. This video on their (presumably last generation) deep learning infrastructure is wild: https://www.youtube.com/watch?v=EFe7-WZMMhc How far large-scale computing has advanced past racks of servers in a datacenter is superb. |
|
Bard’s biggest problem is it hallucinates too much. Point it to a YouTube video and ask to summarize? Rather then saying I can’t do that it will mostly make up stuff, same for websites.
|
|
Oh i just checked. It is generally available where I live. I guess the “your invited to try bard because your a local guide” was just trying to make me feel special and go sign up.
|
|
We tested Bard (aka Bison in GCP) for generating SQL.
It has worse generalization capabilities than even GPT-3.5 but actually does as well at GPT-4 when given contextually relevant examples selected from a large corpus of examples. https://vanna.ai/blog/ai-sql-accuracy.html This means to me that it wants longer prompts to keep away from the hallucination drawback that everybody else appears be experiencing. |
|
When it was first introduced, it received frequent updates [1] but now it’s been 2 months since the last update. So either Google is preparing some huge update (Gemini?), or Bard is going to disappear as a standalone product and instead will be absorbed into other products like Search, Docs, etc.
[1] https://bard.google.com/updates
|
|
At least for programming related questions, it’s more often providing an annoying invalid snippet, rather than anything useful.
|
|
I barely use Bard, but I do use the Search Generative Experience and the Bard-in-Google Docs quite a lot. I find both quite useful as they integrate quite well into my daily workflow.
|
|
the market is slowly moving to abandon chatty AI type LLM things
I didn’t know this was happening. Do you know where the market is moving to? |
|
Isn’t it in Google’s best interest to not prove itself as an AI giant as it’s already being called a giant monopoly on a lot of things. (Search, Android, and Chrome)
|
|
Sorry, what exactly is the killer feature in this example? You say you asked it something and then didn’t say what killer answer it actually responded with
|
|
There are still on-going developments in terms of new features/languages/UX, but I don’t expect any significant quality improvements from Bard until Gemini (next-gen LLM inside Google) arrives.
|
|
IBM mismanagement and general dysfunction within the org
I was supposed to teach Watson law, but was laid off on week 5 of my new job (many years ago) |
|
Going from a foundational model to a chat model requires a ton of RLHF. Where is that free labor going to come from? Google doesn’t have the money to fund that.
|
|
You’ll recall this happened before the whole ChatGPT thing blew up in hype: https://www.washingtonpost.com/technology/2022/06/11/google-…
So… there is a cause why Google particularly must be involved with ethics and optics. I performed with earlier inside variations of that “LaMDA” (“Meena”) once I labored there and it was a bit spooky. There was warning language plastered all around the web page (“It will lie” and so forth.) They’ve positively toned it down for “Bard.” The very last thing Google wants is to be accused of constructing SkyNet, they usually comprehend it. |
|
That is exactly the kind of thing I’m talking about:
Lemoine was a random SWE experiencing RLHF’d LLM output for the first time, just like the rest of the world did just a few months later… and his mind went straight to “It’s Sentient!”. That would have been fine, but when people who understood the subject tried to explain, he decided that it was actually proof he was right so he tried to go nuclear. And when going nuclear predictably backfired he used that as proof that he was even more right. In retrospect he fell for his own delusion: Hundreds of millions of people have now used a more advanced system than he did and intuited its nature better than he did as an employee. _ But imagine knowing all that in real-time and watching a media circus actually end up affecting your work? OpenAI wouldn’t have had people who fit his profile in the building. There’d be an awareness that you needed a certain level of sophistication and selectiveness that the most gun-ho ethicists might object to as meaning you’re not getting fair testing done. But in the end, I guess Lemoine got over it too: seeing as he’s now AI Lead for a ChatGPT wrapper that pretends to be a given person. https://www.mimio.ai/ |