As long as they don’t fuck it up in a similar fashion to seemingly every other thing they have tried for a couple decades.
As long as they don’t fuck it up in a similar fashion to seemingly every other thing they have tried for a couple decades.
Assuming it takes its answer from search results, and the search results are all affiliate marketing sites that just want you to click on a link and buy something, this makes perfect sense.
Is language conscious?
Are atoms?
I don’t know if LLMs of a large enough size can achieve (or sufficiently emulate) consciousness, but I do know that we barely know anything about consciousness, let alone it’s limits.
The thing is, LLMs can be used for something like this, but just like if you asked a stranger to write a letter for your loved one and only gave them the vaguest amount of information about them or yourself you’re going to end up with a really generic letter.
…but to give me amount of info and detail you would need to provide it with, you would probably end up already writing 3/4 of the letter yourself which defeats the purpose of being able to completely ignore and write off those you care about!
I keep forgetting that that’s an option
No clue? Somewhere between a few years (assuming some unexpected breakthrough) or many decades? The consensus from experts (of which I am not) seems to be somewhere in the 2030s/40s for AGI. I’m guessing accuracy probably will be more on a topic by topic basis, LLMs might never even get there, or only related to things they’ve been heavily trained on. If predictive text doesn’t do it then I would be betting on whatever Yann LeCun is working on.
Perhaps there is some line between assuming infinite growth and declaring that this technology that is not quite good enough right now will therefore never be good enough?
Blindly assuming no further technological advancements seems equally as foolish to me as assuming perpetual exponential growth. Ironically, our ability to extrapolate from limited information is a huge part of human intelligence that AI hasn’t solved yet.
GPT-2 came out a little more than 5 years ago, it answered 0% of questions accurately and couldn’t string a sentence together.
GPT-3 came out a little less than 4 years ago and was kind of a neat party trick, but I’m pretty sure answered ~0% of programming questions correctly.
GPT-4 came out a little less than 2 years ago and can answer 48% of programming questions accurately.
I’m not talking about mortality, or creativity, or good/bad for humanity, but if you don’t see a trajectory here, I don’t know what to tell you.
This is very much the type of case that settles out of court for an undisclosed amount of money.
The company is offering affected users a 30 percent discount on a new Ecobee thermostat, valid for up to 15 thermostats.
…
whatever the hell ‘X’ is supposed to be
It’s the social media company he founded, obviously.
Or, if you’re trying to make it more directly comparable to 17 million (because humans aren’t great at implicitly comparing that many zeros), that would be 56,000 millions. It’s not how we normally say it, but 17 vs 56,000 feels different than 17 million vs 56 billion.
From what I understand the bonus requires shareholder approval…so the people making that choice are the same ones watching the stock tank and continuing to hold (or buy more)…
I don’t know why one would do that without thinking Elon is playing some super-advanced 5D Checkers.
If I purchase a TV, that I now own, and after I own it the company “updates” my TV that I now have to watch ads in order to use the TV I purchased without that condition?
At minimum it’s a breach of contract
Increasing profit margins…which is the thing our society puts above all else.
I’m not making any moral judgements one way or the other, but I have a strong feeling kids today are just going to grow up with this stuff and it will be normalized and we are going to be the weird old prudes who have a weird sense of personal identity connected to our physical appearance and voice while they’re going around looking like SpongeBob and talking like The Fonz.
The hype around AI language models has companies scrambling to hire prompt engineers to improve their AI queries and create new products.
Who is hiring all these prompt engineers? Who is ‘scrambling’ to find people for this? The jobs I do see have basically replaced “developer” with “prompt engineer” with the same job requirements.
Emergent behavior is pretty much anything an old model couldn’t do that a new model can. Simple reasoning, creating coherent sentences, “theory of mind”, basic math, translation, I think are a few examples.
They aren’t “amazing” in the sense that a human can’t do them, but they are in the sense that a computer is doing it.
Do “printers” have an approval rating? Because I bet it’s low.
This is what Ilya saw…