Hi everyone!
I’m part of the team behind xThought, an upcoming AI accelerator hardware for LLM inference. We’ve already built a working Transformers extension (xthought_transformers) that ships:
- XThoughtConfig + XThoughtModel classes
- Basic safetensor export + runtime bindings
- Examples and tests
We’d like to integrate xThought with the HFT ecosystem and open-source the stack so a wider community can use it—hence my question about the correct path to upstream our backend into the main Hugging Face GitHub repositories.
I’ve studied the Optimum ecosystem and noticed most vendor backends live in separate repos like optimum-nvidia
, optimum-intel
, optimum-executorch
, etc. That looks like the right pattern for us, so the plan is to open-source an optimum-xthought
repo that mirrors those structures and then PR any Transformers-side hooks if required.
What’s the very first step? Should I open an RFC-style issue in the main Optimum
repository, or is there another channel where I can submit my own optimum-xthought
GitHub repo for an initial review?
If anyone has gone through this journey (e.g. Optimum-NVIDIA/-Executorch/-TPU maintainers ), I’d really appreciate your war stories and any gotchas you hit along the way.
Thanks a ton!