Spaces:
Running
on
T4
Running
on
T4
Merge branch 'main' of hf.co:spaces/Amiel/terra-cognita-frugal-ai
Browse files
.gitattributes
CHANGED
@@ -35,4 +35,7 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
models/best_YOLOv8n_full_batch_1.engine filter=lfs diff=lfs merge=lfs -text
|
37 |
models/best_YOLOv8n_half_batch_64.engine filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
38 |
*.engine filter=lfs diff=lfs merge=lfs -text
|
|
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
models/best_YOLOv8n_full_batch_1.engine filter=lfs diff=lfs merge=lfs -text
|
37 |
models/best_YOLOv8n_half_batch_64.engine filter=lfs diff=lfs merge=lfs -text
|
38 |
+
models/best_YOLOv8n_full_batch_64.engine filter=lfs diff=lfs merge=lfs -text
|
39 |
+
models/best_YOLOv8n_half_batch_128.engine filter=lfs diff=lfs merge=lfs -text
|
40 |
+
models/best_YOLOv8n_full_batch_128.engine filter=lfs diff=lfs merge=lfs -text
|
41 |
*.engine filter=lfs diff=lfs merge=lfs -text
|
models/best_YOLOv8n_full_batch_128.engine
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fd4efa91a2416356443ff9ef3e7e7eff3c633e85a32794f7258d59cf70808b97
|
3 |
+
size 14028776
|
models/best_YOLOv8n_full_batch_64.engine
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:eed8580f8dfb46c683550881ddfa189830612bb261444d1f7fb654ba45bdaec0
|
3 |
+
size 14351822
|
models/best_YOLOv8n_half_batch_128.engine
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:682594caad53e9be74bdbd8b5767cd6151cb45d20407a3ec64fbdbeeddba3e60
|
3 |
+
size 7867111
|
tasks/image.py
CHANGED
@@ -37,7 +37,7 @@ print("CUDA version:", torch.version.cuda)
|
|
37 |
print("cuDNN version:", torch.backends.cudnn.version())
|
38 |
|
39 |
#MODEL_TYPE = "YOLOv11n"
|
40 |
-
DESCRIPTION = f"YOLOv8n model with batch 64 inference on TensorRT"
|
41 |
print(14)
|
42 |
ROUTE = "/image"
|
43 |
print(15)
|
@@ -141,11 +141,11 @@ async def evaluate_image(request: ImageEvaluationRequest):
|
|
141 |
#PATH_TO_MODEL = 'models/best_YOLOv11n_1280.onnx'
|
142 |
#PATH_TO_MODEL = 'models/best_yolov6n_1280.pt'
|
143 |
#PATH_TO_MODEL = 'models/best_YOLOv11n_1280_real_half.onnx'
|
144 |
-
PATH_TO_MODEL = 'models/
|
145 |
INFERENCE_ENGINE_TYPE = 'pt'
|
146 |
INPUT_SIZE = 640
|
147 |
N_TEST_BATCHES = 2
|
148 |
-
BATCH_SIZE =
|
149 |
print("PATH_TO_MODEL", PATH_TO_MODEL)
|
150 |
def preprocessor(frame):
|
151 |
#frame = cv2.cvtColor(frame, cv2.COLOR_BGR2RGB) # Only when read from file
|
@@ -309,7 +309,7 @@ async def evaluate_image(request: ImageEvaluationRequest):
|
|
309 |
print(n_box_distr)
|
310 |
#break
|
311 |
pass
|
312 |
-
|
313 |
|
314 |
for idx in range(len(all_preds)):
|
315 |
if has_smoke_list[idx]:
|
|
|
37 |
print("cuDNN version:", torch.backends.cudnn.version())
|
38 |
|
39 |
#MODEL_TYPE = "YOLOv11n"
|
40 |
+
DESCRIPTION = f"YOLOv8n quantisized model with batch 64 inference on TensorRT"
|
41 |
print(14)
|
42 |
ROUTE = "/image"
|
43 |
print(15)
|
|
|
141 |
#PATH_TO_MODEL = 'models/best_YOLOv11n_1280.onnx'
|
142 |
#PATH_TO_MODEL = 'models/best_yolov6n_1280.pt'
|
143 |
#PATH_TO_MODEL = 'models/best_YOLOv11n_1280_real_half.onnx'
|
144 |
+
PATH_TO_MODEL = 'models/best_YOLOv8n_full_batch_128.engine'
|
145 |
INFERENCE_ENGINE_TYPE = 'pt'
|
146 |
INPUT_SIZE = 640
|
147 |
N_TEST_BATCHES = 2
|
148 |
+
BATCH_SIZE = 128 # Can be adjusted as needed
|
149 |
print("PATH_TO_MODEL", PATH_TO_MODEL)
|
150 |
def preprocessor(frame):
|
151 |
#frame = cv2.cvtColor(frame, cv2.COLOR_BGR2RGB) # Only when read from file
|
|
|
309 |
print(n_box_distr)
|
310 |
#break
|
311 |
pass
|
312 |
+
pred_boxes = []
|
313 |
|
314 |
for idx in range(len(all_preds)):
|
315 |
if has_smoke_list[idx]:
|