Skip to content
This repository has been archived by the owner on Sep 7, 2024. It is now read-only.
/ brainhack-24 Public archive
generated from TIL-24/til-24-base

Commit

Permalink
Add VLM models for Finals
Browse files Browse the repository at this point in the history
  • Loading branch information
neosouwchuan authored and qitianshi committed Jun 13, 2024
1 parent 2e6dc91 commit 6fee80f
Showing 1 changed file with 2 additions and 2 deletions.
4 changes: 2 additions & 2 deletions vlm/src/VLMManager.py
Original file line number Diff line number Diff line change
Expand Up @@ -24,9 +24,9 @@ def __init__(self):
])
print(os.getcwd())
print([f for f in os.listdir('.') if os.path.isfile(f)])
self.clipmodel= torch.load(path.join(path.dirname(path.abspath(__file__)), "clip_ft_2.pt"))
self.clipmodel= torch.load(path.join(path.dirname(path.abspath(__file__)), "clip_ft_final.pt"))
self.objects = ["cargo aircraft","light aircraft","commercial aircraft","drone","missile","helicopter","fighter jet","fighter plane"]
self.model = YOLOWorld(path.join(path.dirname(path.abspath(__file__)), "augmentedbestv2.pt")).to(self.device)
self.model = YOLOWorld(path.join(path.dirname(path.abspath(__file__)), "highdegree60.pt")).to(self.device)
for i in self.clipmodel.parameters():
i.requires_grad=False
for i in self.model.parameters():
Expand Down

0 comments on commit 6fee80f

Please sign in to comment.