moooji commited on
Commit
afc4db9
1 Parent(s): 07d2dac

Update handler.py

Browse files
Files changed (1) hide show
  1. handler.py +6 -7
handler.py CHANGED
@@ -3,24 +3,23 @@ from PIL import Image
3
  import torch
4
  import base64
5
  from io import BytesIO
6
- from transformers import AutoProcessor, BlipForQuestionAnswering
7
 
8
  device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
9
 
10
  class EndpointHandler():
11
  def __init__(self, path=""):
12
- self.processor = AutoProcessor.from_pretrained("Salesforce/blip-vqa-capfilt-large")
13
- self.model = BlipForQuestionAnswering.from_pretrained("Salesforce/blip-vqa-capfilt-large").to(device)
14
 
15
  def __call__(self, data: Any) -> List[float]:
16
  inputs = data.pop("inputs", data)
17
 
18
  image = Image.open(BytesIO(base64.b64decode(inputs['image'])))
19
- inputs = self.processor(image, inputs['question'], return_tensors="pt").to(device)
20
 
21
  with torch.no_grad():
22
- outputs = self.model.generate(**inputs)
23
 
24
- pooler_output = outputs.pooler_output
25
- return processor.decode(out[0], skip_special_tokens=True)
26
 
 
3
  import torch
4
  import base64
5
  from io import BytesIO
6
+ from transformers import AutoProcessor, BlipForConditionalGeneration
7
 
8
  device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
9
 
10
  class EndpointHandler():
11
  def __init__(self, path=""):
12
+ self.processor = AutoProcessor.from_pretrained("Salesforce/blip-image-captioning-large")
13
+ self.model = BlipForConditionalGeneration.from_pretrained("Salesforce/blip-image-captioning-large").to(device)
14
 
15
  def __call__(self, data: Any) -> List[float]:
16
  inputs = data.pop("inputs", data)
17
 
18
  image = Image.open(BytesIO(base64.b64decode(inputs['image'])))
19
+ inputs = self.processor(image, return_tensors="pt").to(device)
20
 
21
  with torch.no_grad():
22
+ outputs = self.model(**inputs)
23
 
24
+ return outputs
 
25