katie on Nostr: Yeah, it’s super interesting about the RLHF! Overfit is a very real problem and ...
Yeah, it’s super interesting about the RLHF! Overfit is a very real problem and adjusting weights on models this large can be kind of like a butterfly effect. I think there is a TON of value it its generalization. But I’m of the opinion that it can’t or maybe shouldn’t do all tasks for itself - to me it’s just not necessarily efficient, like using a hammer on a screw. Bigger doesn’t always mean better - it will start to underperform at a certain size. TBD what that is. But let it do what it does best! Language and conceptual derivations and awesome encoding, let other models do what they’re better suited at. Kind of like how our brains work… we have separate specialized areas that we delegate tasks to when necessary. We’re building awesome components, but I’d like us to acknowledge their limitations, not to discourage the work that has been done, but to figure out the next problem that needs to be solved.
Published at
2023-04-19 02:11:28Event JSON
{
"id": "48a636bc3dfae85f9e68a41a8d79f513dc52be69a4603345f1db86519214923c",
"pubkey": "07eced8b63b883cedbd8520bdb3303bf9c2b37c2c7921ca5c59f64e0f79ad2a6",
"created_at": 1681870288,
"kind": 1,
"tags": [
[
"e",
"494bc5c6d8516f1bb712b0dcbe5a43a11724ff0f4ce44ba45fa87bd33ed8192d"
],
[
"e",
"db1c5b78ad8f185905bfcb8ddaf979b253d4757b95628240f1bf6b185930cbb3"
],
[
"p",
"1bc70a0148b3f316da33fe3c89f23e3e71ac4ff998027ec712b905cd24f6a411"
],
[
"p",
"9baed03137d214b3e833059a93eb71cf4e5c6b3225ff7cd1057595f606088434"
]
],
"content": "Yeah, it’s super interesting about the RLHF! Overfit is a very real problem and adjusting weights on models this large can be kind of like a butterfly effect. I think there is a TON of value it its generalization. But I’m of the opinion that it can’t or maybe shouldn’t do all tasks for itself - to me it’s just not necessarily efficient, like using a hammer on a screw. Bigger doesn’t always mean better - it will start to underperform at a certain size. TBD what that is. But let it do what it does best! Language and conceptual derivations and awesome encoding, let other models do what they’re better suited at. Kind of like how our brains work… we have separate specialized areas that we delegate tasks to when necessary. We’re building awesome components, but I’d like us to acknowledge their limitations, not to discourage the work that has been done, but to figure out the next problem that needs to be solved. ",
"sig": "a02973a5681f6066119f1695585613f358d9cd166e2d8aa35ea74f631451984bc87155c5813609daaecf4cb8d43777e84946bf809f0d0103806e82285fbf7d8e"
}