Hacker News
new
|
past
|
comments
|
ask
|
show
|
jobs
|
submit
login
sn41
on May 18, 2023
|
parent
|
context
|
favorite
| on:
Ask HN: Can someone ELI5 transformers and the “Att...
complete newbie here: what is the intuition behind the conclusion that "cat" is highly related to "black" as opposed to, say, "mat"?
nborwankar
on May 18, 2023
|
next
[–]
Attention and the Transformer make it possible to recognize that the probability of “black” applying to the cat is much much higher than to the mat due to the phrasing “which is” in between the cat and black.
sn41
on May 18, 2023
|
parent
|
next
[–]
Thank you. So this is based on the training data, I assume.
aGHz
on May 18, 2023
|
prev
[–]
It is a lot harder to take the black out of the cat than it is to take the mat out from under it.
djbusby
on May 18, 2023
|
parent
[–]
Humans know that, how does transform know that? Based on training data?
Accujack
on May 18, 2023
|
root
|
parent
[–]
Sort of. Part of the training for a model includes telling it which parts of a sentence are important... a human points and clicks.
testrun
on May 18, 2023
|
root
|
parent
|
next
[–]
This is extremely important to know. That the relationships between words in the sentence are actually trained by human evaluation.
breezeTrowel
on May 18, 2023
|
root
|
parent
|
next
[–]
They are not.
mollerhoj
on May 18, 2023
|
root
|
parent
|
prev
[–]
No, thats incorrect. The connections are automatically deduced from the training data (which is just vast amounts of raw text).
Consider applying for YC's Spring batch! Applications are open till Feb 11.
Guidelines
|
FAQ
|
Lists
|
API
|
Security
|
Legal
|
Apply to YC
|
Contact
Search: