Hacker Newsnew | past | comments | ask | show | jobs | submit | esotericimpl's commentslogin

This is definitely false, see Facebook's pivot to video.


Wasn't that famously based on advertising network lies?


Not really baseless when he was caught with classified documents after he was selling them to the highest bidder.

Yes americas right wing courts blocked justice from being applied in this case but this would hardly be baseless.


I love how this slop is something you think we should be looking forward to.


The thing about slop, is if it is continuously improving slop, maybe one day it won’t be slop any more


quote:

> not claiming that this will necessarily be a good thing


Why are you using full self driving as part of your driving?

Weird ask imo.


The cops lied about where the suspect was identified?

So the warrant was granted from false pretense.


This is always the answer for anyone who thinks LLMs are capable of "intelligence".

It's good at answering questions that its trained on, I would suggest general intelligence are things you didnt want/train the AI to be good at answering.


Are you good at answering questions you are not trained to answer?

How about a middle school test in a language you don’t speak?


For a while I was into a trivia program on my phone. It was kind of easy, so I decided to set the language to Catalan, a language which I never studied. I was still able to do well, because I could figure out the questions more or less from languages I do know and could generalize from them. It would be interesting to know if you could say, train an LLM on examples from Romance languages but specifically exclude Catalan and see if it could do the same.


  > Are you good at answering questions you are not trained to answer?
Yes. Most schooling is designed around this.

Pick a random math textbook. Any will do. Read a chapter. Then move to the homework problems. The typical fashion is that the first few problems are quite similar to the examples in the chapter. Often solvable by substitution and repetition. Middle problems generally require a bit of extrapolation. To connect concepts from previous chapters or courses in ways that likely were not explicitly discussed. This has many forms and frequently includes taking the abstract form to practical (i.e. a word problem). Challenge problems are those that require you to extrapolate the information into new domains. Requiring the connection of many ideas and having to filter information for what is useful and not.

  > How about a middle school test in a language you don’t speak?
A language course often makes this explicitly clear. You are trained to learn the rules of the language. Conjugation is a good example. By learning the structure you can hear new words that you've never heard before and extract information about it even if not exactly. There's a reason you don't just learn vocabulary. It's also assumed that by learning vocabulary you'll naturally learn rules.

Language is a great example in general. We constantly invent new words. It really is not uncommon for someone you know to be be talking to you and in that discussion drop a word they made up on the spot or just make a sound or a gesture. An entirely novel thing yet you will likely understand. Often this is zero-shot (sometimes it might just appear to be zero-shot but actually isn't)


Well ... https://puzzling.stackexchange.com/questions/94326/a-cryptic...

(Someone made a cryptic crossword[1] whose clues and solutions were in the Bahasa Indonesia language, and it was solved by a couple of people who don't speak that language at all.)

[1] These are mostly a UK thing; the crosswords in US newspapers are generally of a different type. In a cryptic crossword, each word is given a clue that typically consists of a definition and some wordplay; there are a whole lot of conventions governing the wordplay. So e.g. the clue "Chooses to smash pots (4)" would lead to the answer OPTS; "chooses" is the definition, "smash pots" is the wordplay, wherein "smash" indicates that what follows should be anagrammed (smashed up).

Disclaimer #1: it took those people a lot more work than it would have taken them to solve an English-language cryptic crossword of similar difficulty, and they needed a bunch of external resources.

(Dis)claimer #2: one of those people was me.

Disclaimer #3: I do not claim that something needs to be able to do this sort of thing in order to be called intelligent. Plenty of intelligent people (including plenty of people more intelligent than me) would also be unable to do it.


Yes — reasonably so, anyway. I don't have to have seen millions of prior examples of exactly the same kind in order to tackle a novel problem in mathematics, say.


Well, LLMs are also remarkably good at generalizing. Look at the datasets, they don't literally train on every conceivable type of question the user might ask, the LLM can adapt just as you can.

The actual challenge towards general intelligence is that LLMs struggle with certain types of questions even if you *do* train it on millions of examples of that type of question. Mostly questions that require complex logical reasoning, although consistent progress is being done in this direction.


  > Well, LLMs are also remarkably good at generalizing. Look at the datasets, they don't literally train on every conceivable type of question the user might ask, the LLM can adapt just as you can.
Proof needed.

I'm serious. We don't have the datasets. But we do know the size of the datasets. And the sizes suggest incredible amounts of information.

Take an estimate of 100 tokens ~= 75 words[0]. What is a trillion tokens? Well, that's 750bn words. There are approximately 450 words on a page[1]. So that's 1.66... bn pages! If we put that in 500 page books, that would come out to 3.33... million books!

Llama 3 has a pretraining size of 15T tokens[2] (this does not include training, so more info added later). So that comes to ~50m books. Then, keep in mind that this data is filtered and deduplicated. Even considering a high failure rate in deduplication, this an unimaginable amount of information.

[0] https://help.openai.com/en/articles/4936856-what-are-tokens-...

[1] https://wordcounter.net/words-per-page

[2] https://ai.meta.com/blog/meta-llama-3/


That’s a very good point. I just speak from my experience of fine-tuning pre-trained models. At least at that stage they can memorize new knowledge, that couldn’t have been in the training data, just by seeing it once during fine-tuning (one epoch), which seems magical. Most instruction-tuning datasets are also remarkably small (very roughly <100K samples). This is only possible if the model has internalized the knowledge quite deeply and generally, such that new knowledge is a tiny gradient update on top of existing expectations.

But yes I see what you mean, they are dumping practically the whole internet at it, it’s not unreasonable to think that it has memorized a massive proportion of common question types the user might come up with, such that minimal generalization is needed.


  > that couldn’t have been in the training data
I'm curious, how do you know this? I'm not doubting, but is it falsifiable?

I also am not going to claim that LLMs only perform recall. They fit functions in a continuous manner. Even if the data is discrete. So they can do more. The question is more about how much more.

Another important point is that out of distribution doesn't mean "not in training". This is sometimes conflated, but if it were true then that's a test set lol. OOD means not belonging to the same distribution. Though that's a bit complicated, especially when dealing with high dimensional data


I agree. It is surprising the degree to which they seem to be able to generalise, though I'd say in my experience the generalisation is very much at the syntax level and doesn't really reflect an underlying 'understanding' of what's being represented by the text — just a very, very good model of what text that represents reality tends to look like.

The commenter below is right that the amount of data involved is ridiculously massive, so I don't think human intuition is well equipped to have a sense of how much these models have seen before.


That's called innovation, something the current AIs aren't capable of.


Amazing that it all actually accelerates in the early 80s with Reagan who kicks it into overdrive with the collapse of unions.

But yeah, it was the "gold standard" and not the collapse of the labor movement, there's nothing stopping us from taking back the value of labor except it comes out of the pockets of our oligarchs.


777 as well, except for the fact that MH370 where the pilot deliberately brought down the plane and the other one was shot down by Russian Separatists in Ukraine.


Except the research doesn't show this at all.

However, having wealth provides tutors, better schooling and teachers which lead to better outcomes from having better mentors, opportunities and education which leads to...

and so on and so on.


Right, that is step 6.

If you go over the research, you realize that they all assume the premise "everyone is born with an identical blank slate". Which leads to conclusions (either said outright or implied) like; "If you give enough resources to any given child, they can grow up to be anything."

So we end up with schools that serve low income areas getting enormous amounts of funding to teach admin assistants and tile layers to be heart surgeons. And the data on that end shows unequivocally that it is not working.


> And the data on that end shows unequivocally that it is not working.

Of course throwing money at (sometimes corrupt) school boards isn’t going to be effective. There isn’t much 1/20th of a school teacher can do 8am-2pm M-F to counter a bad environment. It takes a village after all.

Not exactly data that’s useful to your hypothesis.


> If you go over the research, you realize that they all assume the premise "everyone is born with an identical blank slate".

From my experience this is a large exaggeration to not true at all. For example the Colorado adoption project has been around has been around for a longer time and was well regarded last I checked.

Are there piles of hard science papers out there that are concluding "everyone is born with an identical blank slate" that I am missing?


So your solution is boots on the ground and regime change? it's gone so well in the past it definitely will in the future. /s

Maybe instead the US should honor their commitments which would include the Iran Nuclear deal that was scrapped for no reason 5 years ago.


I don't believe it to be a good idea. Biden has flirted with Iran with deadly consequences.

I'm not going to speculate how good it would to attempt regime change right now but there was a very good opportunity for it when Iran had unrest and they came to a genius idea to murder around 15000 arrested people.

The best opportunity would have been of course when US had sizable military formations on both sides of Iran border. On top of it it would have very likely limited large number of civilian causalities in Iraq due to reduced activity of Iran backed terrorist cells.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: