LLM MEMORY FINETUNNING: Data preparation for Chat. I only have long chunks of proprietary text data
I’m planning to do memory fine-tune on an instruction-based model to replace a RAG and utilize proprietary data.
Exploring Large Models’ Knowledge Understanding Output and Academic Research on Specific Content
I am currently exploring the performance of large models in understanding knowledge in specific domains, and attempting to construct a knowledge framework similar to what humans establish when learning a subject. This understanding does not need to be flawless, but it should provide a comprehensive grasp of the core concepts and structure of the subject. […]
Python environment Tensorflow cuda version different from global cuda version
First why do we need to install tensorflow with cuda does it mean we dont need to install cuda globally. I installed cuda(i.e for os) globally using nvidia documentation and tensorflow cuda using pip install tensorflow[and -cuda].
Finetuning LLM with SQL Metadata
I am currently working on a text2sql project. To avoid sending a large text of metadata as a part of the prompt as context, I created a set of question and answer pairs as fine tuning training data. I included all SQL metadata along with every question and the correct SQL as the answer in the training data
ImageBind LLM checkpoint
i want to use imagebind llm model for my task, but i can not import llama and find out the checkpoints for ImageBind-LLM.enter image description here
Function to extract sub-statements from overall statement with LLMs
I have the following situation: I have a statement, e.g.
LLM to convert binary to decimal
So, I’m a complete beginner on how to use any LLMs, given below is the code which I was using on my jupyter notebook to run a prompt to convert binary to decimal. However I was getting completely incorrect responses on Llama 3.
Using LLM to create playlist
I am creating a system that creates playlist through text description and have decided to use LLM by fine-tuning it
How to pretrain a mini LLM by 8 * A800 within 1~2 weeks
I am doing an internship in a llm company, want to try to go throught the pretraining of a mini llm.
how can i intergare my self-developed large-language-model to openai(OneApi),thank you
I’m new to large-language-model. My manager wants me to integrate our large-language-model to OneApi, similar to Llama 3 in OneApi, but I can’t find the documentation or sample code for the interface. Thank you very much for your reply.