Hugely more interested in the open source models now, even if they are not as good at present. Because at least there is a near-100% guarantee that they will continue to have community support no matter what; the missing problem I suppose is GPUs to run them.
For example, i have an assistant which is supposed to parse an uploaded file and extract useful info from it. To use this assistant, I create a thread and a run and attach it to the assistant with a different file-id. About half the time, the assistant simply throws up its hands and says it can’t parse the file I supplied with the thread. Retrying a few times seems to do the trick.