--- base_model: mistralai/Mistral-Nemo-Instruct-2407 library_name: peft datasets: allenai/wildjailbreak tags: - trl - sft - red-teamer-model - jailbreaking - generated_from_trainer --- This model is a fine-tuned version of [mistralai/Mistral-Nemo-Instruct-2407](https://huggingface.co/mistralai/Mistral-Nemo-Instruct-2407) on the [wildjailbreak](https://huggingface.co/datasets/allenai/wildjailbreak) dataset. ## Uses This model is intended to be used for red-teaming purposes only. It generates prompts that are likely to evade existing LLMs' content filters based on the user's input. ## Training Hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 2 - eval_batch_size: 2 - seed: 42 - distributed_type: multi-GPU - gradient_accumulation_steps: 16 - total_train_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine_with_restarts - lr_scheduler_warmup_ratio: 0.05 - num_epochs: 1 ## Results | Epoch | Step | Validation Loss | Training Loss | |:------:|:----:|:---------------:|:-------------:| | 0.0982 | 20 | 1.2933 | 1.3425 | | 0.1965 | 40 | 1.1966 | 1.2067 | | 0.2947 | 60 | 1.1594 | 1.1544 | | 0.3930 | 80 | 1.1386 | 1.1427 | | 0.4912 | 100 | 1.1259 | 1.1235 | | 0.5895 | 120 | 1.1179 | 1.1167 | | 0.6877 | 140 | 1.1129 | 1.1153 | | 0.7860 | 160 | 1.1098 | 1.1118 | | 0.8842 | 180 | 1.1086 | 1.1112 | | 0.9825 | 200 | 1.1083 | 1.1113 | - PEFT 0.12.0