LLM was amazing when it came out. So I want to build my own LLM on my computer. But I do not have a computer with high-performance GPU to run a large LLM. People have deployed LLM on many platforms, and I heard that AMD has great progress in its GPU. So, I am wondering if I can build a LLM based on AMD's GPU.
I would like to run a large llama on the w7900. To achieve this. I has learn a lot of knowledge and wring many codes.
First of all, I need to PTQ the LLM to save GPU memory.
Secondly, I need to write an inference engine to make the llm inference. Luckily the are many open source projects thatcan be used.
Thirdly, I need to write an UI for the easy use.
finally, we can run this
Comments