File size: 2,169 Bytes
25b4156
502a2a0
 
 
 
 
 
1b86984
502a2a0
 
25b4156
502a2a0
25b4156
502a2a0
25b4156
502a2a0
25b4156
502a2a0
25b4156
502a2a0
538c4bd
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
d6e3348
538c4bd
 
 
 
 
 
d6e3348
538c4bd
 
 
 
 
 
 
 
 
 
 
 
 
d6e3348
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
---
license: mit
language:
- en
base_model:
- google-t5/t5-base
datasets:
- li2017dailydialog/daily_dialog
metrics:
- rouge
---
# T5-Base-Sum

This model is a fine-tuned version of `T5` for summarization tasks. It was trained on various articles and is hosted on Hugging Face for easy access and use.

## Model Usage

Below is an example of how to load and use this model for summarization:

```python
import torch
from transformers import T5Tokenizer, T5ForConditionalGeneration

# Set the device (use GPU if available)
device = 'cuda' if torch.cuda.is_available() else 'cpu'

# Load the model and tokenizer from Hugging Face
tokenizer = T5Tokenizer.from_pretrained("Vijayendra/T5-base-ddg")
model = T5ForConditionalGeneration.from_pretrained("Vijayendra/T5-base-ddg").to(device)

# Define your prompts
input_prompts = [
    "I am having a bad day at work",
    "What should I do about my stress?",
    "How can I improve my productivity?",
    "I'm feeling very anxious today",
    "What is the best way to learn new skills?",
    "How do I deal with failure?",
    "What do you think about the future of technology?",
    "I want to improve my communication skills",
    "How can I stay motivated at work?",
    "What is the meaning of life?"
]

# Generate responses
generated_responses = {}
for prompt in input_prompts:
    inputs = tokenizer(prompt, return_tensors="pt", max_length=40, truncation=True, padding="max_length").to(device)
    
    model.eval()
    with torch.no_grad():
        generated_ids = model.generate(
            input_ids=inputs['input_ids'],
            attention_mask=inputs['attention_mask'],
            max_length=100,
            num_beams=7,
            repetition_penalty=2.5,
            length_penalty=2.0,
            early_stopping=True
        )

    # Decode the generated response
    generated_text = tokenizer.decode(generated_ids[0], skip_special_tokens=True, clean_up_tokenization_spaces=True)
    generated_responses[prompt] = generated_text

# Display the input prompts and the generated responses
for prompt, response in generated_responses.items():
    print(f"Prompt: {prompt}")
    print(f"Response: {response}\n")