zlacker

[parent] [thread] 1 comments
1. joegib+(OP)[view] [source] 2026-01-01 01:40:34
Would it make more sense to instead train a model and tokenise the syntax of languages differently so that white space isn’t counted, keywords are all a single token each and so on?
replies(1): >>__Matr+D
2. __Matr+D[view] [source] 2026-01-01 01:46:38
>>joegib+(OP)
After watching models struggle with string replacement in files I've started to wonder if they'd be better off in making those alterations in a lisp: where it's normal to manipulate code not as a string but as a syntax tree.
[go to top]