Skip to content

Does it support a TPU? #2497

Answered by hiro-v
fwartner asked this question in Get Help
Discussion options

You must be logged in to vote

Hi @mawoka-myblock and @Disane87

Currently we do not have plan for supporting TPU, as our current local inference engines are llama.cpp and nvidia tensorrt llm do not support those. We might support Python runtime and we can discover the possibility of running on TPU

There are several reference here:

  • NVIDIA Tensorrt-llm of course only supports NVIDIA GPU
  • For llama.cpp
    • https://github.com/ggerganov/llama.cpp/issues/1052#issuecomment-1515339426
    • https://github.com/ggerganov/llama.cpp/issues/3253

Replies: 6 comments

Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Answer selected by imtuyethan
Comment options

You must be logged in to vote
0 replies
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
6 participants
Converted from issue

This discussion was converted from issue #2492 on March 26, 2024 03:23.