k050506koch
commited on
Commit
•
809aeca
1
Parent(s):
ece0628
Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Model Card for V2 Models
|
2 |
+
|
3 |
+
## Model Description
|
4 |
+
This repository contains multiple models trained using the GPT-2 architecture for generating creative stories, superhero names, and abilities. The models are designed to assist in generating narrative content based on user prompts.
|
5 |
+
|
6 |
+
## Model Variants
|
7 |
+
- **Story Model**: Generates stories based on prompts.
|
8 |
+
- **Name Model**: Generates superhero names based on story context.
|
9 |
+
- **Abilities Model**: Generates superhero abilities based on story context.
|
10 |
+
- **Midjourney Model**: Generates mid-journey prompts for storytelling.
|
11 |
+
|
12 |
+
## Training Data
|
13 |
+
The models were trained on a custom dataset stored in `batch_ds_v2.txt`, which includes various story prompts, superhero names, and abilities. The dataset was preprocessed to extract relevant parts for training.
|
14 |
+
|
15 |
+
## Training Procedure
|
16 |
+
- **Framework**: PyTorch with Hugging Face Transformers
|
17 |
+
- **Model**: GPT-2
|
18 |
+
- **Training Arguments**:
|
19 |
+
- Learning Rate: 1e-4
|
20 |
+
- Number of Epochs: 15
|
21 |
+
- Max Steps: 5000
|
22 |
+
- Batch Size: Auto-detected
|
23 |
+
- Gradient Clipping: 1.0
|
24 |
+
- Logging Steps: 1
|
25 |
+
|
26 |
+
## Evaluation
|
27 |
+
The models were evaluated based on their ability to generate coherent and contextually relevant text. Specific metrics were not provided, but qualitative assessments were made during development.
|
28 |
+
|
29 |
+
## Inference
|
30 |
+
To use the models for inference, you can send a POST request to the `/generate/<model_path>` endpoint of the Flask application. The input should be a JSON object containing the `input_text` key.
|
31 |
+
|
32 |
+
### Example Request
|
33 |
+
```
|
34 |
+
json
|
35 |
+
{
|
36 |
+
"input_text": "[Ivan Ivanov, Lead Software Engineer, Superhero for Justice, Writing code, fixing issues, solving problems, Masculine, Long Hair, Adult]<endoftext>"
|
37 |
+
}
|
38 |
+
```
|
39 |
+
|
40 |
+
### Example Response
|
41 |
+
The response will contain the generated text based on the input prompt.
|
42 |
+
|
43 |
+
## Limitations
|
44 |
+
- The models may generate biased or nonsensical outputs based on the training data.
|
45 |
+
- They may not always understand complex prompts or context, leading to irrelevant or inaccurate responses.
|
46 |
+
- The models are sensitive to input phrasing; slight changes in the prompt can yield different results.
|
47 |
+
|
48 |
+
## License
|
49 |
+
This model is released under the MIT License. Please refer to the LICENSE file for more details.
|
50 |
+
|
51 |
+
## Citation
|
52 |
+
If you use this model in your research or applications, please cite it as follows:
|