Relative Content

Tag Archive for large-language-model

Two phase pre-training Qwen 2.5?

I have read the Qwen 2.5 Technical Report and other papers, but I am still curious about how they train models in two phases, with each phase using a different context length.

Two phase pre-training Qwen 2.5?

I have read the Qwen 2.5 Technical Report and other papers, but I am still curious about how they train models in two phases, with each phase using a different context length.

Two phase pre-training Qwen 2.5?

I have read the Qwen 2.5 Technical Report and other papers, but I am still curious about how they train models in two phases, with each phase using a different context length.

Finetuning LLM with SQL Metadata

I am currently working on a text2sql project. To avoid sending a large text of metadata as a part of the prompt as context, I created a set of question and answer pairs as fine tuning training data. I included all SQL metadata along with every question and the correct SQL as the answer in the training data

ImageBind LLM checkpoint

i want to use imagebind llm model for my task, but i can not import llama and find out the checkpoints for ImageBind-LLM.enter image description here

LLM to convert binary to decimal

So, I’m a complete beginner on how to use any LLMs, given below is the code which I was using on my jupyter notebook to run a prompt to convert binary to decimal. However I was getting completely incorrect responses on Llama 3.