This project fine-tunes the Qwen2.5 language model using the Llama-Adapter method for fact-checking. This approach enables efficient training of large language models with low computational costs.
Qwen2.5 is a large language model developed by Alibaba Cloud, while Llama-Adapter is an efficient fine-tuning method originally designed for Llama models. This project integrates both to create an effective fine-tuning solution for Qwen2.5, specifically for fact-checking tasks.
This project focuses on training the model to:
- Analyze and assess the accuracy of claims
- Identify reliable information sources
- Detect misinformation and fake news
- Provide evidence and explanations for conclusions
git clone https://github.com/xndien2004/Qwen2.5-Adapter-Instruct.git
cd Qwen2.5-Adapter-Instruct
pip install -r requirements.txt
Place fact-checking training data in the data/
directory following this format:
{
"messages":[
{"role": "system", "content": "You are an AI assistant specializing in verifying the accuracy of information in Vietnamese."},
{"role": "user", "content": "Question: You are tasked with verifying the correctness of the following statement.\n We provide you with a claim and a context. Please classify the claim into one of three labels:\n - SUPPORTED: the evidence supports the claim;\n - REFUTED: the evidence contradicts the claim;\n - NEI: not enough information to decide.\n Your answer should include the classification label and the most relevant evidence sentence from the context.\n Remember, the evidence must be a full sentence, not part of a sentence or less than one sentence. Given a claim and context as follows:\n Context: {context}\n Claim: {claim}\n Answer: The claim is classified as <LABEL>. The evidence is: <EVIDENCE>"},
{"role": "assistant", "content": "Answer: The claim is classified as {verdict}. The evidence is: {evidence}"},
],
"format": "chatml"
}
For the SFT datasets, the raw JSONLINE file follows the following format:
{"messages": [sample1...], "format": "chatml"}
{"messages": [sample2...], "format": "chatml"}
{"messages": [sample3...], "format": "chatml"}
python3 -m fine_tune \
--model_name "Qwen/Qwen2.5-0.5B-Instruct" \
--adapter_layer 12 \
--adapter_len 128 \
--max_length 768\
--train_file "/kaggle/input/viwikifc-process/chatml_viwikifc_train.jsonl" \
--val_file "/kaggle/input/viwikifc-process/chatml_viwikifc_test.jsonl" \
--output_dir "output" \
--batch_size 4 \
--gradient_accumulation_steps 1 \
--epochs 4 \
--learning_rate 5e-4 \
--is_type_qwen_adapter "v2" \
--use_model_origin 0
MIT License
-
Qwen2.5 Technical Report (2025)
- Authors: Qwen Team
- arXiv: 2412.15115
- Description: Detailed technical report on the Qwen2.5 model
-
LLaMA-Adapter: Efficient Fine-tuning of Language Models with Zero-init Attention (2023)
- Authors: Zhang et al.
- arXiv: 2303.16199
- Description: Efficient fine-tuning method using zero-init attention
-
SemViQA: A Semantic Question Answering System for Vietnamese Information Fact-Checking (2025)
- Authors: Nguyen et al.
- arXiv: 2503.00955
- Description: A semantic question-answering system for Vietnamese fact-checking