add README from github.com/warplab/megafishdetector
Browse files
README.md
CHANGED
@@ -1,3 +1,45 @@
|
|
1 |
---
|
2 |
license: mit
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: mit
|
3 |
---
|
4 |
+
|
5 |
+
# megafishdetector
|
6 |
+
|
7 |
+
Detector for generic "fish" trained on publicly available datasets, currently supporting YOLO-style bounding boxes prediction and training. Can also be used as pre-trained networks for further fine-tuning.
|
8 |
+
|
9 |
+
Initial experiments to train a generic MegaFishDetector modelled off of the MegaDetector for land animals (https://github.com/microsoft/CameraTraps/blob/main/megadetector.md)
|
10 |
+
|
11 |
+
Currently based on YOLOv5 (https://github.com/ultralytics/yolov5).
|
12 |
+
|
13 |
+
This repo contains links to public datasets, code to parse datasets into a common format (currently YOLO darknet only), and a model zoo for people to start with. For instructions to run, see the link above.
|
14 |
+
|
15 |
+
## Instructions
|
16 |
+
1. Install [Yolov5](https://github.com/ultralytics/yolov5)
|
17 |
+
2. Download desired network [weights](https://github.com/warplab/megafishdetector/blob/main/MODEL_ZOO.md)
|
18 |
+
3. Usage (from yolov5 root): python detect.py --imgsz 1280 --conf-thres 0.1 --weights [path/to/megafishdetector_v0_yolov5m_1280p] --source [path/to/video/image folder]
|
19 |
+
|
20 |
+
## Public Datasets Used in v0:
|
21 |
+
|
22 |
+
- [AIMs Ozfish](https://github.com/open-AIMS/ozfish)
|
23 |
+
- [FathomNet](https://www.fathomnet.org/)
|
24 |
+
- [VIAME FishTrack](https://viame.kitware.com/#/collection/62afcb66dddafb68c8442126)
|
25 |
+
- [NOAA Puget Sound Nearshore Fish (2017-2018)](https://lila.science/datasets/noaa-puget-sound-nearshore-fish)
|
26 |
+
- [DeepFish](https://alzayats.github.io/DeepFish/)
|
27 |
+
- [NOAA Labelled Fishes in the Wild](https://www.st.nmfs.noaa.gov/aiasi/DataSets.html)
|
28 |
+
|
29 |
+
## To Cite:
|
30 |
+
|
31 |
+
[paper](https://arxiv.org/abs/2305.02330)
|
32 |
+
```
|
33 |
+
@misc{yang2023biological,
|
34 |
+
title={Biological Hotspot Mapping in Coral Reefs with Robotic Visual Surveys},
|
35 |
+
author={Daniel Yang and Levi Cai and Stewart Jamieson and Yogesh Girdhar},
|
36 |
+
year={2023},
|
37 |
+
eprint={2305.02330},
|
38 |
+
archivePrefix={arXiv},
|
39 |
+
primaryClass={cs.RO}
|
40 |
+
}
|
41 |
+
```
|
42 |
+
## TODO:
|
43 |
+
- Train larger models
|
44 |
+
- requirements.txt for things like fathomnet environment
|
45 |
+
- COCO format output
|