if a model architecture is supported by transformers, you can expect it to be supported in the rest of the ecosystem.
It's interesting, but it seems regardless of the TLDR, Llama.CPP maintains its own implementation and so won't benefit from any "day-0 support" for new architectures in Transfomers. I was kind of hoping Llama would get straight Transformer compatibility for a moment.
It's not clear what, if anything, is changing there.