zlacker

[return to "Chess-GPT's Internal World Model"]
1. triyam+r51[view] [source] 2024-01-07 02:01:24
>>homarp+(OP)
Is a linear probe part of observability/interpretability?
◧◩
2. canjob+M51[view] [source] 2024-01-07 02:04:43
>>triyam+r51
Yes, a pretty fundamental technique and one of the earliest. It lets you determine which layers contain what information among other things.
◧◩◪
3. Legend+G71[view] [source] 2024-01-07 02:23:39
>>canjob+M51
The downside is that it's a supervised technique, so you need to already know what you're looking for. It would be nice to have an unsupervised tool that could list out all the things the network has learned.
◧◩◪◨
4. Joshua+xe1[view] [source] 2024-01-07 03:40:46
>>Legend+G71
Anthropic has published some cool stuff in that direction: https://transformer-circuits.pub/2023/monosemantic-features
◧◩◪◨⬒
5. fireja+Ll1[view] [source] 2024-01-07 05:09:20
>>Joshua+xe1
Whoa, this is super cool! I can imagine if we had something like this for ChatGPT, we could use it to do some serious prompt engineering. Imagine seeing what specific neurons you were activating with your prompt, and being able to identify which word in your prompt was triggering an undesired behavior. Super cool stuff, excited to see if it scales
[go to top]