pull down to refresh
related posts
10 sats \ 0 replies \ @mudbloodvonfrei 31 May 2023
It's an interesting article. I didn't understand most of it, but still. It seems that the author wants believes that interpretability can lead to better alignment of superintelligent AIs but my question is this: if you can interpret the behavior of a system, wouldn't that mean that the system is not superintelligent compared to humans? We have a lot of experts that try to interpret human behavior, or to manipulate it (i.e. align it) with certain goals, but we still don't know all that much about the human brain and people can't always interpret their own behavior much less someone else's.
reply
0 sats \ 0 replies \ @hn OP 31 May 2023
This link was posted by og_kalu 1 hour ago on HN. It received 76 points and 4 comments.
reply