Spaces:
Sleeping
Sleeping
File size: 1,194 Bytes
dc480fb 732d546 dc480fb 043a657 dc480fb da70a87 043a657 dc480fb 732d546 dc480fb 043a657 dc480fb 732d546 dc480fb 043a657 da70a87 dc480fb da70a87 dc480fb 732d546 dc480fb |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 |
import os
from transformers import TFBertForSequenceClassification, BertTokenizerFast
def load_model(model_name):
try:
# Load TensorFlow model from Hugging Face
model = TFBertForSequenceClassification.from_pretrained(model_name, use_auth_token=os.getenv('API_KEY'))
except OSError:
# Fallback to PyTorch model if TensorFlow fails
model = TFBertForSequenceClassification.from_pretrained(model_name, use_auth_token=os.getenv('API_KEY'), from_pt=True)
return model
def load_tokenizer(model_name):
tokenizer = BertTokenizerFast.from_pretrained(model_name, use_auth_token=os.getenv('API_KEY'))
return tokenizer
def predict(text, model, tokenizer):
inputs = tokenizer(text, return_tensors="tf")
outputs = model(**inputs)
return outputs
def main():
model_name = os.getenv('MODEL_PATH')
if model_name is None:
raise ValueError("MODEL_PATH environment variable not set or is None")
model = load_model(model_name)
tokenizer = load_tokenizer(model_name)
# Example prediction
text = "Sample input text"
result = predict(text, model, tokenizer)
print(result)
if __name__ == "__main__":
main() |