I was creating a project where I wanted to interface LLM (Tiny LLAMA). I was thinking if I could do it locally in the android app itself without calling server for the same. Is there any way out to implement it locally? Or I need to interface it on server
I have implemented it at Server but it need alot of RAM. So just wanted to minimize the cost by interfacing it offline on device