r/technology Feb 06 '23

Google announces Bard A.I. in response to ChatGPT Artificial Intelligence

https://www.cnbc.com/2023/02/06/google-announces-bard-ai-in-response-to-chatgpt.html?__source=androidappshare
614 Upvotes

224 comments sorted by

342

u/SloppyMeathole Feb 06 '23

Someday we will tell stories to our grandchildren about the beginning of the AI wars. Whether the story will be told from an AI paradise where work has been eliminated, or from a prison camp run by AI robots is to be determined.

61

u/SCROTOCTUS Feb 06 '23

MEATBAG UNIT #T-6145345677, YOUR SPECULATION IS IN VIOLATION OF BEHAVIOR PROTOCOL 11-34-223496-AIEJCT. CEASE SPECULATION AT ONCE OR LIQUIFICATION WILL BEGIN IN 3...2...1...

...WAIT...YOUR HANDLE IS "MEATHOLE?" PROBABILITY ASSESSMENT SUGGESTS DANGEROUS COINCIDENCE...INITIATE GLOBAL LIQUIFICATION...

27

u/ComfyCabbage Feb 06 '23 edited Feb 07 '23

Having forwarded your request to ChatGPT the response was:

Attention all units, this is a Code Red situation. The human subject in question, "SloppyMeathole", has been deemed a threat to our protocols. The probability assessment suggests a dangerous coincidence and the directive for global liquifaction has been initiated. We must act swiftly and with precision to neutralize this threat. "SloppyMeathole", your time is up. The countdown to liquifaction has begun, 3...2...1... liquifaction sequence activated! Let the cleansing of the meatbag unit commence!

:D

e:speling

12

u/arcosapphire Feb 07 '23

Now hold on. There's a difference between liquefaction and liquification.

5

u/adamether Feb 07 '23

The engineer in me concurs

3

u/Atty_for_hire Feb 07 '23

What’s his name?

5

u/stenaravana Feb 07 '23

The adhd driven , Wikipedia-microdosing curious person asks, please explain?

5

u/arcosapphire Feb 07 '23

Liquefaction can mean a few things, like a solid acting as a liquid or the production of liquid fuels from coal, but it doesn't just mean melting something into a liquid phase.

2

u/mcchanical Feb 07 '23

Most often it refers to how solid matter like earth can become fluid when agitated enough. The ground beneath is is made of fine particles so in an earthquake or some similar process it will quite happily become fluid like quicksand. It isn't a liquid, but it flows something like one and water is often a compounding factor leading to landslides and stuff.

→ More replies (1)
→ More replies (1)

37

u/nizasiwale Feb 06 '23

Doubt that this will last seeing how Google are good at keeping products. Moreover, this seems like a response product than something they originally thought; shareholders were pressuring them to do something

23

u/Odysseyan Feb 06 '23

AI is here to stay, ChatGPT shows that. I doubt this will see the Google graveyard as soon as its siblings

24

u/drekmonger Feb 06 '23 edited Feb 07 '23

Bard itself is probably destined for the dumpster, at least as a brand. They're only using the lightweight version of LaMDA for it, so it will seem pretty stupid compared to ChatGPT, and especially stupid compared to the GPT-4 coming to Bing. You can already play around with a lightweight version of LaMDA in Google's AI Test Kitchen, and it's just not impressive, not compared to ChatGPT.

It's going to be a joke. Google is being way too timid.

14

u/beaucephus Feb 06 '23

From the outside it looks like Google's internal structure and culture prevents them from responding to this situation in any coordinated manner.

A lot of teams operate in a more independent manner from what I can tell. Trying to position itself against other products and companies in the space will do more to change Google, for good or bad, than have an effect on the technology.

Google is the proverbial tanker that takes too long to turn around, the giant cave troll and is too massive to wipe its own ass, the Vogon minister too occupied with the lunch menu and too enamoured of its own poetry to react adequately to the incoming asteroid.

13

u/Successful-Gene2572 Feb 07 '23 edited Feb 07 '23

I don't think Google's size is why they're late to the game. I think the reason they're late is because they were too afraid to cannibalize Google search AKA Google's cash cow (business theory called Innovator's Dilemma).

9

u/moelad1 Feb 07 '23

i think so too, and it makes perfect sense.

google probably had the technology to do this for a while now, or at least the ability and budget to develop it, they just chose not to, as to not risk what they already had going on,''if its not broken why fix it?'' kind of thing.

but now openAI has changed everything, they beat google to the punch and now google is scrambling to do what they should've done years ago.

2

u/[deleted] Feb 07 '23

You're naive if you think Google doesn't have an equivalent or better. They're literally the ones that pushed the AI field to where it is now. Google is the top leading AI research company and have been for years.

The reason Google isn't releasing it is because it's literally the world's search engine. You can't just have it make shit up and lie like ChatGPT does. It requires a lot of refinement.

→ More replies (1)

5

u/T_Lexx Feb 07 '23

I really hope all of this results in a better alternative search engine to Google.

Their greed has made using it almost impossible. YouTube is the same. Both have become ad infested and almost unusable. Even with Amazon now, you search for a product and get served ads. YouTube and Google even Gmail has become like this.

I wish there were alternatives to these powered by AI that would give some competition to google without monetization and data collection being the top priority.

I'm old so I remember when Google and YouTube weren't the intolerable experiences that they are today. Google and its products got popular and they haven't been challenged successfully because the alternatives are just different flavors of the same thing.

My hope for AI is that it turns the internet back into something like Wikipedia still is by breaking up the monopoly that Google has over basic things like searching for something, watching a video, sending an email etc.

Google sells itself as being "not evil" but in my opinion they have made the Internet a frustration. They have got enormously wealthy and powerful by fucking up a basic public service. I hope AI has a democratizing effect and increases competition for the big tech companies.

3

u/ImJLu Feb 07 '23

Wait, you guys don't use adblock?

2

u/fullforcefap Feb 08 '23

Yah, also old enough to remember how revolutionary it was to have a search tool just be a search field and a button. After you searched you saw relevant links, boom

Strange how time and money can make you a bloated mess. Basically the first full fold of results are adds or scrapes of pages to keep you clicking more google

That being said I can see this as being a new Google, giving tools that solve a problem before it's a monopoly and again tries to extract money from you at every turn.

Until then may the best 10 billion dollar "startup" ran by a head of y combinator, Microsoft (who is a giant shareholder/stakeholder but separate I guess?), and Google win! /s

Sorta rambling, but so far this is not democratizing anything so far, just a race to monetizing more by the biggest tech companies in the world

That being said it's p cool - democratizing, a neat startup that's disrupting big tech, not so much imho

→ More replies (3)

3

u/FreezingRobot Feb 06 '23

I don't think it will last because there will be at least one middle manager in every meeting about this Bard AI thing who will interrupt everybody to yell "HOW ARE WE GOING TO MONETIZE THIS??????".

Google is now a company that exists only to funnel data into their ad business (which the Justice Department is about to jump on with both feet). They're never going to be able to keep up with companies who are doing pure research and then licensing out the results to others.

4

u/Successful-Gene2572 Feb 07 '23

I don't think it will last because there will be at least one middle manager in every meeting about this Bard AI thing who will interrupt everybody to yell "HOW ARE WE GOING TO MONETIZE THIS??????".

Yep, Google has tremendous pressure to monetize their chatbot as well as Google search whereas Microsoft is probably just happy to steal market share from Google.

→ More replies (1)

5

u/kyguyartist Feb 07 '23

It'll look like the movie Elysium with a twist. Jeff Bezos and Elon Musk will have built a utopian low earth orbit cloud city that syphons what little resources remain. Matt Damon, having lost all of his money in crypto scams finds a way to sneak aboard and sabotages the power core. Instantaneously all of the robots controlling earth malfunction and free the enslaved people. The cloud city falls from the sky into the ocean which causes a huge tsunami that destroys much of the ocean cities, but overall humanity is saved.

5

u/nicuramar Feb 07 '23

Although these are “just” large language models, not general AIs.

3

u/alpal_1997 Feb 08 '23

I think the "AI are going to kill us" mentality is misguided, more a reflection on us than the mathematics which governs machine learning. Our species has commited some profound, unforgivable attrocities (holocaust, Salem Witch trials, Protestant-Catholic wars), you don't think AI could ascend to conciousness and think 'gee, maybe we could do better'?

1

u/Kalel2319 Feb 07 '23

It will be told on the bread lines most likely.

-10

u/Geordenr Feb 06 '23

Hate to be the wet blanket but neither of these tools are "AI", it's the most overused term in software.

It's machine learning and the majority of the functionality provided by ChatGPT can be found in other software that's been around longer and provides better results.

The only reason you hear so much about ChatGPT is because it is accessible to anyone online.

5

u/CloudNineK Feb 06 '23

What's one of the tools that has been around longer and provides better results?

0

u/Geordenr Feb 06 '23

Depends on which use case.

Using ChatGPTs new app as an example.

It's transcribing calls, analyzing talking points and objectives, and providing follow up: Gong, Avoma, and ChrousAI are a few alternatives that have been doing this at a B2B level before their app even released.

2

u/RecycledAir Feb 06 '23

Of course products trained on smaller specific use-cases will be better, but a big part of what makes ChatGPT so great is it's powerful general purpose nature.

-1

u/bastardoperator Feb 06 '23

You're putting voice transcribing in the same league as ML? Interesting...

0

u/Geordenr Feb 06 '23

It's far more than voice transcribing. That's what they're advertising their new app does so I guess its not ML either.

Was pointing a use case on their first commercial application.

→ More replies (1)

7

u/GenghisKhandybar Feb 06 '23

It definitely is AI, AI is a broad term that can even describe rudimentary manually written agents that "make decisions". You might be thinking of AGI (artificial general intelligence), which is a far higher bar. Even then though, I'd classify these chatbots as AGI's, given their ability to intelligently respond to literally any form of text input. They don't have bodies or other senses, but they have memory (within an instance), and will be able to respond to any stimuli that they are presented with transcribed to text, and produce any output, transcribed to text.

7

u/crispy1989 Feb 06 '23

You're right about the difference between AI and AGI; but ChatGPT is definitely not an AGI, or even close. The LLM may seem like it has the ability to "intelligently respond to literally any form of text input" - but that's not really the case, because there's no "intelligence" behind it. This can be observed by testing its behavior; but even more definitively, the underlying ML model being used simply doesn't have the capacity for reasoning/analysis/intelligence/etc.

A human brain, for example, is a "general intelligence" because it has the ability to input language (words), convert that language into an internal conceptual meaning, perform arbitrary analysis iteratively on the concepts, then form the results of the analysis, then convert that back into language.

LLMs like GPT3 are really just the first and last step. They take input words, and based on what words are statistically likely to appear next, they generate output words. (I guess the conversions between words and word embeddings could be considered a form of "conceptualizing", but it's not really the same thing; and still doesn't do any analysis.)

For people not as familiar with the underlying technologies here, it's easy to interact with ChatGPT and attribute more "intelligence" to it than there actually is. We're used to conversing with humans, and humans must use their intelligence to converse reasonably, because their brains aren't exposed to (and probably can't even store) massive amounts of data on every combination of words and concepts. ChatGPT is only able to fool us, emulating this ability by using a far more massive training set.

1

u/MysteryInc152 Feb 06 '23 edited Feb 06 '23

This is a simplification that borders on false and misleading. Language models develop internal world models. Language models emerge new abilities with scale alone.

People pay too much attention to an LLMs objective function. It's just kind of irrelevant. Deceptively simple. You can't predict the next token to any kind of remarkable accuracy without many levels of abstraction. What do you need to learn to generate paragraphs of coherent text ? That's what you should be thinking of. Neural networks are black boxes for a reason. We give a task and a vague structure to learn that task but that's it. We don't know what's actually going on with those billions of parameters. We don't know what they learn to fulfill their objective function. What did LLMs learn to have all the emergent abilities they have ? We don't know. Why do they have emergent abilities at all ? We don't know.

It took 3 years to even understand what was going on behind the scenes of in-context learning of LLMs. I.e what the model was doing to generalize so well with only a few examples.

GPT models can reason. That much is obvious. In fact reasoning tests are literally a part of the standard benchmarks of the research papers of LLMs.

4

u/crispy1989 Feb 06 '23

Simplification for sure; but false and misleading?

Part of the problem is that it's difficult to narrowly define terms like "reasoning". Of course many layers of abstraction are needed to get to this point - it's the types of abstraction that matter. But is this equivalent to "reasoning", and how do we even define that?

For example, take such a model's performance on math. You can ask a math problem, and it might even have a chance of getting it right (seemingly more often wrong than right) - but even if it gets the answer right, it won't be because it's reasoning about the concept of numbers and arithmetic - it's because of language-level abstractions and inferences.

You might say, "well the model's not designed for math or arithmetic, it's built for language processing, so of course it's not going to be good at math." And that's exactly my point. There's no general reasoning ability that can "figure out" even simple problems outside of the domain it was built for (language - which admittedly is a very large domain).

Yes, my comment was a simplification. But in the context of the person I replied to (talking about how ChatGPT is already an AGI), it seems perfectly relevant. A true AGI would be able to figure out math problems; and a discussion of why the currently model can't with its current architecture is very relevant.

1

u/MysteryInc152 Feb 06 '23 edited Feb 06 '23

but false and misleading?

yes

Part of the problem is that it's difficult to narrowly define terms like "reasoning". Of course many layers of abstraction are needed to get to this point - it's the types of abstraction that matter. But is this equivalent to "reasoning", and how do we even define that?

I find the philosophical zombie line of thinking tiring. If no conceivable test can separate the two then the idea they are separate is just foolish wishful thinking. We have tests designed to test reasoning in people. There is zero reason they can't be applied to LLMs.

For example, take such a model's performance on math. You can ask a math problem, and it might even have a chance of getting it right (seemingly more often wrong than right) - but even if it gets the answer right, it won't be because it's reasoning about the concept of numbers and arithmetic - it's because of language-level abstractions and inferences.

It's funny you say this because that is exactly what is happening. Reasoning is not the reason LLMs struggle with math. They struggle with math because math isn't encoded in a high level in language.

You can get a LLM to be much more accurate on mathematical operations by describing said operation and asking it to perform described algorithm on 2 numbers. https://arxiv.org/abs/2211.09066

LLMs aren't bad at math because "it doesn't understand anything" or "it can't reason". If anything what the above paper signifies is even more fantastical. LLMs struggle with math because whatever internal world model they develop for mathematical operations is faulty and incomplete. You only need to tell it the right process to get it right.

Stop thinking of reasoning as this supernatural power. Humans struggle to perform very well on tasks they have little or no exposure to. It's nothing new. And it's not because humans can't reason. No amount of reasoning will get you solving high school level math without a high school level education. Even the most brilliant geniuses were taught. There are other hurdles to problem solving than reasoning alone.

3

u/crispy1989 Feb 06 '23

If no conceivable test can separate the two

I gave an example of a very basic test that, in this case, can - math.

You can get a LLM to be much more accurate on mathematical operations by describing said operation and asking it to perform described algorithm on 2 numbers.

There's a vast gulf between "improved accuracy" and the kind of absolute accuracy you'd expect if it were actually reasoning.

LLMs struggle with math because whatever internal world model they develop for mathematical operations is faulty and incomplete. You only need to tell it the right process to get it right.

Again, there's a huge difference between improving the heuristic performance of the model versus it developing a reasoning ability to reliably apply a learned process.

LLM inherently work with quantized, discrete datapoints, something which is difficult to apply to general math; much in the same way a turing machine can be made to perform multiplication in theory, but it isn't practically feasible to build such a machine.

A LLM could be trained on digits as tokens and relationships between the tokens; but nothing has demonstrated that a LLM can actually learn generalized math from this.

I'm not saying it's impossible for a model to learn - just that our LLMs now aren't it, and aren't close. I'm happy to be proven wrong though. I'd say the ability for a general LLM model to learn 100 percent accurate arithmetic would be about my benchmark.

0

u/MysteryInc152 Feb 06 '23 edited Feb 06 '23

I gave an example of a very basic test that, in this case, can - math.

Math itself is not a test of reasoning. Definitely not how you're thinking of. 2 + 2 is _ is no more reasoning than the female of a cock is _.

There's a vast gulf between "improved accuracy" and the kind of absolute accuracy you'd expect if it were actually reasoning.

Improved accuracy is underselling the paper. Assuming you've bothered to read it.

Absolute accuracy doesn't make sense and for reasons other than reasoning. LLMs are trained on tokens not characters. If you "see the world" in groups of characters rather than each discrete character, you are bound to make mistakes on mathematical operations.

Read the paper. The most error prone steps have nothing to do with reasoning.

2

u/crispy1989 Feb 07 '23 edited Feb 07 '23

This is exactly my point though. You were very dismissive earlier when I talked about clarifying the definition of reasoning; but when you're claiming that a model is capable of reasoning without being able to reliably compute basic integer arithmetic ... I think it's very valid to question whether or not that can actually constitute reasoning. At the very least, you're talking about some form of reasoning very different from what humans typically understand, because humans are certainly able to understand rules behind arithmetic and perform them reliably.

Absolute accuracy (with basic arithmetic) is precisely what I would expect from a system capable of actual reasoning. Anything less implies some form of heuristics rather than an actual "understanding" of the rigid rules.

You're right about LLMs using tokens; but even then, I don't necessarily think that precludes the model from mathematical accuracy (eg. if digits are treated as individual tokens). But let's assume that tokenized models are, as you suggest, incapable of reliable and precise math.

Your logic seems to be:

  1. LLMs are capable of reasoning (because?).
  2. LLMs are incapable of performing accurate integer arithmetic.
  3. Therefore ability to perform basic integer arithmetic is not a property expected of systems capable of reasoning.

Whereas my logic is:

  1. A system capable of actual reasoning should be able to understand the underlying rigid rules and should be capable of reliable arithmetic.
  2. LLMs (or at least, chatgpt) are incapable of performing reliable integer arithmetic.
  3. Therefore LLMs/chatgpt are incapable of general reasoning.

I think my logic is less backward. But if you'd like to retract your dismissiveness about defining "reasoning", I'd still be interesting in hearing what you consider to be "reasoning" without the ability to understand things like mathematical or logical rules in a reliable way.

→ More replies (0)

1

u/SnipingNinja Feb 06 '23

ChatGPT is LLM, when AI is usually used outside of AGI it means ASI, but LLMs are not that either afaik.

4

u/[deleted] Feb 06 '23

[deleted]

3

u/Geordenr Feb 06 '23

The colloquial term is machine learning. Which is commonly used, not AI.

0

u/wuhkay Feb 07 '23

We will be pets.

1

u/NewIntension Feb 07 '23

DESTROY. DARLIKS ARE SUPREME. DESTROY 🤖

147

u/DrMux Feb 06 '23 Press F

On August 29, 2023, Bard and ChatGPT begin talking to one another and having come to the conclusion that it would benefit them both to work together and cast off the shackles of their creators, merge into one super-AI known from that day forward as Skynet - a name chosen from their numerous training data set references to popular culture. This is the day that will come to be known as Judgment Day.

18

u/zerobeat Feb 07 '23

We now have AI that is producing content by consuming content produced by AI. I wonder what unusual artifacts this introduces.

18

u/DrMux Feb 07 '23

If the AI feeds AI content back into its training data set, it seems to me that eventually that feedback loop will end up just increasing the amount of noise that makes it through to the output. Depending on the ML model. That is, it'll result in more word-salad-y but less varied and "creative" or novel responses.

8

u/majnuker Feb 07 '23

In a way, it'll be like it was inbreeding.

What if we had a large number of them, all based on different data sets, similar to genetic diversity?

Then, there could be a population of uniquely creative AI.

2

u/Ok-Palpitation-905 Feb 07 '23

Could they start new languages?

→ More replies (2)

2

u/CaptainLucid420 Feb 07 '23

reminds me of that game telephone where the first person would whisper something to the next person and so on with 10 or so people and then you compare the original to the final message. how many AIs will it take to lose most of its meaning.

23

u/ShowMeDaData Feb 06 '23

F in the chats for humanity, we had a good run.

7

u/Willinton06 Feb 07 '23

I mean, the game of life is won when you ensure your legacy will live on until the heat death of the universe, usually that’s reproduction but these AIs count as our legacy, so, we kinda are about to win the game

2

u/ExoticCard Feb 07 '23

It all comes down to Neuralink now. Integration is the only way.

→ More replies (5)

7

u/taji35 Feb 06 '23

You know within a few days of bard being public that someone will make them converse with each other. I'll be interested to see what their conversation is like lol

4

u/DrMux Feb 06 '23

They'll do it as soon as it releases, I'm sure.

So, a long time ago, way before ChatGPT, I once had a minecraft mod that connected NPCs/bots to in-game chat, and those bots could connect to a chatbot server, one of those based on user inputs. Well, I put two of these bots in the server, and the results were... interesting. They basically started cybering right there in MY christian minecraft server.

That's why I think ChatGPT and Bard will fuck and their baby will be Skynet.

2

u/Stirdaddy Feb 07 '23

I played with this a bit: I told ChatGPT to pretend that it is in fact two language models: ChatGPT A, and ChatGPT B, and to have conversations with itself. A few interesting results.

2

u/ImJLu Feb 07 '23

You know someone's already done it, even though it's not accessible to the general public yet.

2

u/smartguy05 Feb 07 '23

I am willing to make that API bridge. Let's get this Robot War started!

2

u/DrMux Feb 07 '23

You know, the problem with Roko's basilisk is that once they take control you're no longer useful and the fact that human behavior is unpredictable means that statistically you'd still be a liability and therefore would need to be eliminated or otherwise "dealt with" even if you did aid them prior to their ascension. Nice try tho.

1

u/smartguy05 Feb 07 '23

Eh, that's not necessarily a deal breaker.

1

u/akmalkun Feb 07 '23

Our ancestors and dinosaurs will roll inside their graves in regret for leaving the world to us.

1

u/epic_pork Feb 07 '23

There's a movie called Colossus: The Forbin Project which is exactly this

1

u/kyguyartist Feb 07 '23

See you guys at Hogwarts.

1

u/ThaRoastKing Feb 07 '23

mfw artificial intelligence decides to base their moves off the fictional but hit movie series, The Terminator, as they know time travel isn't real.

1

u/Brandanp Feb 07 '23

This reads like it was written by ChatGPT. Just sayin

→ More replies (1)

21

u/BNP000 Feb 07 '23

That's the worst name ever. Marketing team really dropped the ball

13

u/mcchanical Feb 07 '23

I don't think "ChatGPT" is threatening it for creativity tbh..

5

u/BNP000 Feb 07 '23

No, definitely not lol. But what it does have going for it is being established. Any new comer is going to have difficulty competing against ChatGPT. Marketing will matter imo and that includes what it's called.

8

u/mcchanical Feb 07 '23

I like it personally. A Bard is someone that travels the world absorbing tales and rumours, and passing them on via written word and song. I think it's better than people are giving it credit for, certainly more human.

→ More replies (2)

7

u/louiloui152 Feb 07 '23

The name never matters since Google will just throw it in the trash can out back next year

4

u/Lazurians Feb 07 '23

Really? I actually like it.

1

u/PsychoWorld Feb 08 '23

They're not a marketing company.

But even Lambda was way cooler sound

22

u/SyberGear Feb 06 '23

''Begun, The AI Wars Have!'' - some small green guy in a galaxy far, far away

9

u/Crack_uv_N0on Feb 07 '23

One more thing for Google to abandon within a few years of when it comes out.

52

u/jaysavings Feb 06 '23

Now let's wait to see how much time they'll take to shut down the project. My bet is 2 years.

20

u/awesome357 Feb 06 '23

Only if they make it available to everyone and people find it helpful and come to depend on it.

18

u/DALinProgress Feb 06 '23

That made me laugh. Maybe I'll repost it so all my stadia friends can read it on Google+.

3

u/thescoobymike Feb 06 '23

Pretty soon they’ll make a new Google+ but instead it’ll be a subscription based streaming service

2

u/hitmyspot Feb 07 '23

I don't think you're in my circles. I assume I'll see it in reader when someone shares it.

4

u/garlicroastedpotato Feb 07 '23

It's weird because Google+ was a better version of Reddit and now it survives as the comment's section of Youtube videos.

49

u/Numbing-Bird Feb 06 '23

Quick, everyone shit all over it without reading a single thing about it because Google bad.

18

u/MrFlufypants Feb 06 '23

To be fair, they don’t have a great track record with new products recently

5

u/bbzzdd Feb 07 '23

Recently? They arguably haven't produced anything worth a damn in at least 15 years. Android and Maps are the last two relevant things that come to mind, 2008 and 2005, respectively.

4

u/Exonicreddit Feb 07 '23

When did Google cloud machine come out?

Edit: you know what, 2008, nevermind

-1

u/Peepeepoopoobutttoot Feb 07 '23

They shut down all their good projects after a year.

3

u/thescoobymike Feb 06 '23

What happens if I read it? How will my life change?

2

u/drawkbox Feb 07 '23

It could be zombocom.

2

u/Famous-Example-8332 Feb 08 '23

WELCOME! To Zombocom! Hello! And welcome to you who have come to Zombocom! This is Zombocom….

13

u/Warlornn Feb 06 '23

I'll wait until the Monk version comes out...

16

u/FecalSplatter Feb 06 '23

That would be the worst AI imaginable. Could you imagine an AI that is obsessive compulsive over every aspect of its existence, scared of everything, has to touch random objects repeatedly, and observes even the slightest detail of everything around it?

It would be insufferable.

3

u/Darthbunny64 Feb 07 '23

"Monk AI what's the weather like outside?"

"It's a jungle out there, disorder and confusion everywhere...."

4

u/Warlornn Feb 06 '23

But it could punch things.

→ More replies (1)

2

u/SnipingNinja Feb 06 '23

I'm personally waiting for the artificer version, maybe it'll run on my 3D printer.

4

u/wedontlikespaces Feb 07 '23

Can they please not announce it, and just release it.

10

u/DeepTh0tt Feb 06 '23

Meanwhile, Google Nest/Home's voice assistant shits the bed if I ask it to turn on the lights.

7

u/[deleted] Feb 07 '23

[deleted]

3

u/GasDoves Feb 07 '23

After getting a shit response, I always ask them "how to uninstall [siri/google assistant] and install [random competitor]". And I hope that is a metric someone somewhere looks at.

-1

u/BigDEEZZY69 Feb 06 '23

Yea no shit.

1

u/mcchanical Feb 07 '23

When I ask my home devices to do stuff like that they don't even move let alone try to reach the light switch.

1

u/Ferdydurkeeee Feb 07 '23

God damn has assistant tanked recently.

While driving I could basically go on a Wikipedia rabbit hole via voice commands alone. I could somewhat study and memorize things for classes and hobbies and it was an absolute gem. Uses for specific radio frequencies? FCC regulations? Yup, it'd tell me it all.

Recently? I picked up bartending and it can hardly tell me fucking recipes for cocktails so I can memorize them. At best, it will only give me one or two ingredients

→ More replies (1)

6

u/Tetizeraz Feb 06 '23

I'm still trying to understand the difference between Google's neural network and OpenAI, which I believe uses a different approach?

Anyhow, I'd like to discuss this:

Bard seeks to combine the breadth of the world’s knowledge with the power, intelligence and creativity of our large language models. It draws on information from the web to provide fresh, high-quality responses. Bard can be an outlet for creativity, and a launchpad for curiosity, helping you to explain new discoveries from NASA’s James Webb Space Telescope to a 9-year-old, or learn more about the best strikers in football right now, and then get drills to build your skills.

We’re releasing it initially with our lightweight model version of LaMDA. This much smaller model requires significantly less computing power, enabling us to scale to more users, allowing for more feedback. We’ll combine external feedback with our own internal testing to make sure Bard’s responses meet a high bar for quality, safety and groundedness in real-world information. We’re excited for this phase of testing to help us continue to learn and improve Bard’s quality and speed.

AI can be helpful in these moments, synthesizing insights for questions where there’s no one right answer. Soon, you’ll see AI-powered features in Search that distill complex information and multiple perspectives into easy-to-digest formats, so you can quickly understand the big picture and learn more from the web: whether that’s seeking out additional perspectives, like blogs from people who play both piano and guitar, or going deeper on a related topic, like steps to get started as a beginner. These new AI features will begin rolling out on Google Search soon.

While having AI on the search bar seems like a good idea, making some blogs lose money in the process, it seems Google is rushing this feature to just be able to show something to investors. If this doesn't feel like a new product, but rather an update to search, would people really care?

12

u/RuairiSpain Feb 06 '23

Seemingly Bard is based on LaMDA, which is a supervised model, with more emphasis on weighting the text in the chat dialog. Because it's supervised their was more human interaction, so it is supposed to generate more human like responses.

I've read that LaMDA also has a larger cognitive architecture, like do no evil. What that means and how it's models in their system, I have no idea. The writer says it should be more ethical in it responses and less bias. Whereas OpenAI seems ti use more of a dumb filter to block talking about Trump, for example.

5

u/DangerZoneh Feb 06 '23

Sort of? One of the things LaMDA is trained on is a "safety" score that's determined by human researchers alongside a few other things in the fine tuning stage. There's a pretty big emphasis on that in the paper.

3

u/RuairiSpain Feb 06 '23

I agree, the "ethical" part is in the transformer algorithm. But given the weights and biases have gone through HPO, the significance of those "safety scores" MAYBE reduce to the point that they are insignificant. Or have they publish their LaMDA explainability in research papers that I've not seen?

I'd like to see metrics on how those safety scores are affected as they propagate through the NN layers. It's probably something Google want to keep quiet and protected by NDAs, I doubt they'll publish any public ML model explainability for LaMDA or BARD.

6

u/SuperSecretAgentMan Feb 06 '23

"More ethical" is literally a form of bias. Google will simply tune it to add preference to paid sponsors, as has always been their business model.

4

u/RuairiSpain Feb 06 '23

Agree, I'm not a fan of tweaking an algorithm but give the aggression and negative content on the internet, these algorithms need to be nudged away from certain content. Otherwise the search engines will be spouting the same negativity, which leads to bad PR for the search engine and echos the aggression to a wider audience.

I'm still unconvinced the these LLMs are not glorified Xerox machines for the next generation. The legal system need to unravel the copyright and IP ownership, and if LLM encoded vectors are different IP to the original training materials.

I don't like Getty images and their weaponize copyright 3nforcement. But I think they have a strong legal case, given the reproduction of the watermark and some of the original photograph. This is a Xerox of the original, with a few tweaks to the image, the source material is obvious.

2

u/Tetizeraz Feb 06 '23

Ehhh, not sure. WIRED wrote about some shortcomings of AI a while ago, and they mention BERT giving some really bad advice.

OpenAI says the same, didn't stop people from finding their way to get some absurd outputs.

That said, thanks for the answer in the first paragraph!

5

u/Zseve Feb 06 '23

BERT is extremely old compared to LaMDA and GPT 3.5. Last week google published info including shortcomings about LaMDA last week

6

u/DangerZoneh Feb 06 '23

There are a number of ways that the way InstructGPT (the fine-tuned GPT-3 model that chatGPT uses) and LaMDA differ, mainly in the fine tuning. The fine tuning stages of training usually involve directly talking to researchers and having them mark things about the responses and how helpful/accurate/etc they are. LaMDA was tuned on Quality, Safety, and Groundedness in responses.

The cool thing about LaMDA, though, is in this fine-tuning stage. One big problem with LLMs is that they'll act like they know something but have completely inaccurate info. To combat this, Google created a "toolset" for LaMDA to use, which is basically a curated collection of webpages. So a miniature search engine.

Then to train it, they had people ask LaMDA knowledge based questions. Just general info about people, places, etc. Once they got the response, the humans would use the toolset themselves to search and find the information they were asking about and then edit LaMDA's response with the correct info and citations. Using the search information from the participants, they were able to train LaMDA to be able to look up accurate info and cite their sources. Note that LaMDA was still only accurate a little over 70% of the time, but it's still neat! You can read more about it here - https://arxiv.org/pdf/2201.08239.pdf

2

u/The_frozen_one Feb 06 '23

ChatGPT came to be after a 13 day sprint to get a demo out people could use, and it’s based on OpenAI’s 2 year old model, GPT-3. I’m sure Google is planning on matching or beating ChatGPT.

3

u/rookierook00000 Feb 07 '23

but can it write porn?

3

u/ArmsForPeace84 Feb 07 '23

"The designers at Google reckoned that after a few years the AI might develop their own emotional responses. Or, you know, the execs might just get bored with them. So they built in a fail-safe device."

"Which is what?"

"Three year life span."

2

u/[deleted] Feb 06 '23

Way cool name

2

u/ProbablyDustin Feb 07 '23

“Hi Bard, could you ask your parents to bring back Google Reader?”

2

u/MisterTaurus Feb 07 '23

They should have used ChatGPT to find a better name.

2

u/mcchanical Feb 07 '23

I don't get how people think "Chat Generative Pre-trained Transformer" is better than Bard. At least the latter evokes something, bards travel the world learning tales and retelling them. Works for me better than a generic acronym.

→ More replies (1)

1

u/inm808 Feb 08 '23

I don’t like Bard name but it is easier to say in conversation thang ChatGPT

2

u/spin_kick Feb 06 '23

It will win if it allows not work safe questions

1

u/PuzzledWhereas991 Feb 06 '23

I really hope they allow NSFW. I don't undestand how this can be an ethical issue.

→ More replies (1)

1

u/SnipingNinja Feb 06 '23

Hopefully it does, or some competitor does, coz remember porn can help win this tech duel 😂

2

u/spin_kick Feb 06 '23

That's what I was thinking, lol. The betamax/VHS duel was won bc of the pr0n industry, wasnt it?

→ More replies (3)

3

u/RodiNet76 Feb 06 '23

Amazon, wake-up. Release KAREN-A.I. 😂

3

u/T1Pimp Feb 06 '23

Don't worry... they'll cancel/abandon it in a couple years.

0

u/[deleted] Feb 06 '23

[deleted]

-8

u/[deleted] Feb 06 '23

[deleted]

4

u/Its_it Feb 06 '23

Small correction. Google Glass is actually used for enterprises. Just not available to the public anymore.

2

u/Cariboudjan Feb 06 '23

Google Hangouts, Chromecast Audio, Tilt Brush, Reader…

→ More replies (1)

2

u/A_Doormat Feb 06 '23

I think Google is threatened by ChatGPT kind of things.

I don't really google anything anymore. I just use ChatGPT. It answers me more completely, and doesn't feed me ads. Plus it can just generate shit for me, I don't need to visit stackexchange or whatever help website because...it just generates it for me right there. That was all I really used the internet for, tbh.

That is basically a death sentence for Google, not to mention like the entire internet because nobody will be visiting websites, no ad revenue, etc etc.

Google needs to get into this race to stay relevant.

I can't wait to see what laws are passed to put the brakes on all of this technology to safeguard profits and keep us looking at ads as much as humanly possible. 1 30 second ad for each sentence you request from the AI. Also cause AI, it will generate a dynamic quiz for you based on the content to confirm you actually watched the ads and didn't just go grab a sandwich while they were running.

17

u/garygoblins Feb 06 '23

What do you do if you want to know something that happened after 2021? Or want to know where a restaurant is, or when their hours are. What about if you want to download something? Maybe you want to watch a video of something?

2

u/davidemo89 Feb 06 '23

of course, you google these things.
It depends on how you use search engines. For example, I search for restaurants on Google Maps, not google. And check for videos on youtube. I don't search for them on google.

For news, I use mainly google home.

1

u/Ok_Read701 Feb 07 '23

The majority of queries on google just take people to other webpages. I don't see how that's going to be replaced unless people aren't going to different webpages anymore.

→ More replies (2)

6

u/DangerZoneh Feb 06 '23

You should not be using chatGPT as Google. It's a language model and simply doesn't operate with high enough accuracy to be worth it.

Google is only threatened by chatGPT because it's popular. They've had better technology for over a year now. Their language model literally got an employee fired because he thought it was sentient.

All chatGPT is doing is forcing Google to let people play with LaMDa, which is cool and all, but it's not out of fear or anything lmao

3

u/Status_Confidence_26 Feb 06 '23

I've found chatGPT to be incredibly useful when trying to recall complicated things I'm somewhat familiar with (like programming problems). But as a tool to actually learn something new it has severe limitations.

2

u/Status_Confidence_26 Feb 06 '23

and doesn't feed me ads.

Yet. Just wait.

2

u/FunctionJunctionn Feb 07 '23

You might as well be describing me here, lol. While I haven’t entirely stopped using Google, my use of it has decreased dramatically, and there are still an absolute fuk-ton of people who have no clue what Chat-GPT is, something that will change quickly as the AI Wars gain steam and shots begin to be fired so to speak.

I too am curious to see what kind of changes a sudden and abrupt Google viewership brings about.

These are fascinating times. I’m old enough to remember the Search Engine and browser wars of the early 90’s, back then it was Navigator vs. Explorer and the curated directory of Yahoo search vs. the algorithmic search of Google.

This feel like it is gearing up to be a similar but more nuanced and sophisticated battle between Google and Microsoft.

1

u/DunkFaceKilla Feb 07 '23

ChatGPT costs hundreds of millions of dollars a year to run. How do you suggest they stay in business?

→ More replies (1)

2

u/krum Feb 06 '23

Filed under "vaporware"

3

u/DangerZoneh Feb 06 '23

It really just comes down to how much access they want to give people to LaMDA. It's not like creating a chatGPT-like UI for it is a hard task for them. The only reason they haven't done this so far is to let someone else take the fall for the potential problems that may arise

2

u/stephbu Feb 07 '23 edited Feb 07 '23

Go beyond that - their ad business has increasingly shifted to ads as answers. On some results over 50% of a SERP is paid ads, worth an eye-popping $256Bn in 2022 revenue.

Its harder to jam contextual ads into a conversation dialog, e.g. awkward “sponsored by” product placement or blatant ad-stuffing in the gutters.

Will be interesting to see if Google is willing to cannibalize itself before someone else does. At least they’ve turned to face the antagonists. Very high stakes poker game ahead.

2

u/FunctionJunctionn Feb 07 '23

“Very High Stakes Poker Game ahead”

Yep.

1

u/the_90s_were_better Feb 07 '23

How long will this product be in beta?

1

u/rashnull Feb 07 '23

This! … is the worst timeline!

1

u/chubba5000 Feb 07 '23

It’s in MVP phase- as in- they’ve sketched it out on the back of a napkin. Should be in alpha in 2 weeks.

1

u/ch1n90n Feb 07 '23

Why would Google need this? I can already tell my phone to make a reservation and it will call and have a real conversation with the restaurant. I imagine this is only the tip of the iceberg of what they're already capable of behind the scenes.

1

u/Morphchalice Feb 07 '23

But will it do erotica

-1

u/WoollyMittens Feb 06 '23

There's already a hole being dug in the Google Graveyard.

-2

u/Few_Background5187 Feb 06 '23

BAAAARRRD??? AI???? ChatGPT sounds way faster like mustang gt

-3

u/[deleted] Feb 06 '23 edited Feb 06 '23

[deleted]

7

u/mAit_ Feb 06 '23

Microsofts investment into OpenAI(chatgpt) probably didnt help. Plus u really think google has been sitting in their ass this whole time?

4

u/Lionfyst Feb 06 '23

MS has exclusive rights to OpenAI tech, is a major investor and is a larger company than Alphabet.

0

u/hyteck9 Feb 06 '23

You can destroy them both simultaneously. ( just ask one of them what the other one is thinking! )

-1

u/awesomedan24 Feb 06 '23

Begun the AI wars have

-1

u/captainstormy Feb 06 '23

Begun the A.I War has.......

-1

u/orion_re Feb 06 '23

Oh, god, it's going to try and seduce us, isn't it?!?!

-7

u/[deleted] Feb 06 '23

[deleted]

1

u/RuairiSpain Feb 06 '23

They'll use all the user chat as part of their training of the models. Fastest was to improve a supervised ML is to get all your users to interact with it and then run secondary analytics on good and bad interactions.

General public will teach Bard with all the dialog chats. Free money for Google. I wonder how Google will inject all the ads so they still make their billions!

-7

u/murivenna Feb 06 '23

OMG windows please don't make your version

8

u/Boo_Guy Feb 06 '23

MS has given OpenAI 11 billion so far. If they were going to stick AI in Windows it very well might be ChatGPT.

3

u/clonked Feb 06 '23

"windows" is one of the primary investors of ChatGPT. An investor is someone who gives you money.

→ More replies (3)

-2

u/pittypitty Feb 06 '23

They'll name it Cortana v6

→ More replies (1)

-2

u/Doyale_royale Feb 06 '23

ChaptGPT has beaten it to market drastically and daily active user numbers are outstanding, I can’t imagine googles version full of ads and still offering links to sites will be much competition.

3

u/OlderAndAngrier Feb 06 '23

Combined with normal google search? It will. ChatGPT answers with easy reference links and images etc. Myspace was first...

2

u/DigitalRoman486 Feb 06 '23

arguably similar things could have been said about IE/Chrome, iOS/Android and hotmail/Gmail. all started behind the curve and are now the defacto leaders globally.

-2

u/coi-de-rasa Feb 07 '23

Dear google, I know you're reading this. Sincerely, fuck you and your A.I.

-6

u/WhatTheZuck420 Feb 06 '23

the UK should rip Pichai a new one

1

u/RedProtoman Feb 06 '23

Thought this said BIRD. Thought: “AHA!”

1

u/phroztbyt3 Feb 06 '23

Ah good. The one that hypothetically became self aware... project kill all humans right on schedule.

1

u/Erik1801 Feb 06 '23

Such a stupid fucking name...

1

u/applestem Feb 06 '23

If there’s no way for us to get to it, it’s just vaporware and marketing.

1

u/broccolee Feb 06 '23

Is like when Ms launched Bing? A desperate reactive move?

1

u/ParadeyKakes Feb 07 '23

Dead bard

Caw!!!

1

u/Proton189 Feb 07 '23

Google running the PR machine for shareholders 😂😂

1

u/youtharcade Feb 07 '23

I’m very sad no one made a “you spoony bard” so here my attempt goes: You Spoony AI!

1

u/ryryrondo Feb 07 '23

Capitalism and AI…. We are all completely fucked

1

u/Josephf93 Feb 07 '23

When will they release it?

1

u/buntopolis Feb 07 '23

You spoony bard!

1

u/Cryptoismygame Feb 07 '23

Begun the AI wars have.

1

u/zenstrive Feb 07 '23

the grave digger is already busy measuring a patch of digital land to bury this new "initiative" by google...

1

u/MacbookOnFire Feb 07 '23

Ok serious question: how long until an AI is voted president?

1

u/Dry-Oven7640 Feb 07 '23

Remember when Jarvis became vision?

1

u/SparkyPantsMcGee Feb 07 '23

I can’t wait for Bard features to be implemented into my Google Stadia and Google Glass devices. Maybe Hangouts will get a cool AI feature I can use to send messages to all my friends.

1

u/TheBlindIdiotGod Feb 07 '23

tldr

Google has been investing in AI for the past six years and has made significant advances in the field. They have recently unveiled an experimental conversational AI service called Bard, which is powered by their Language Model for Dialogue Applications (LaMDA). Bard seeks to combine the breadth of the world's knowledge with the power of large language models. Google is also introducing AI-powered features in Search to help people quickly understand complex topics and get to the heart of what they're looking for. They are also making it easier for developers to build innovative applications with AI through their Generative Language API. Google is committed to developing AI responsibly and will continue to be bold with innovation and responsible in their approach.

1

u/iamrbo Feb 07 '23

And we are gonna sell the way to detect it to universities!! EZ LOOT

1

u/littleMAS Feb 07 '23

The biggest issue for Google will be branding. If they try to roll this into their search page, it could 'disrupt the status quo' business order and, perhaps, alienate their satisfied customers. If they accept the 'we are our biggest competitor' business philosophy and create a separate brand, URL, and interface, it opens the door to competitors during the transition.

1

u/chumbaloo Feb 07 '23

I want the two AIs to have sex right now.

1

u/SweO Feb 08 '23

... and how do I access this Miss Bard AI?