Boosting Large Language Models with Mask Fine-Tuning
Paper
•
2503.22764
•
Published
•
1
This model is a Full Fine-Tuned (FFT) version of LLaMA2-7B on coding datasets, trained as part of replicating the Mask Fine-Tuning (MFT) paper.
The model was trained on 30,000 samples from three coding datasets (matching the paper):
This model serves as the FFT baseline for the Mask Fine-Tuning paper replication. It will be evaluated on:
Evaluation on HumanEval is pending. Results will be updated here once available.
If you use this model, please cite the original MFT paper:
@article{mft2025,
title={Mask Fine-Tuning},
author={[Authors from paper]},
journal={arXiv preprint arXiv:2503.22764v1},
year={2025}
}
Training configuration and code available at: GitHub Repository
This model inherits the LLaMA 2 Community License from the base model.
Base model
meta-llama/Llama-2-7b-hf