Replies: 4 comments 4 replies
-
Great, i'm glad the data can be of use. Let's break these down into small questions. I'll need to review the code to make sure I remember correctly.
I believe the data was sampled at random. Looking here https://github.com/weecology/BirdDetector/blob/b2c50d0c840e4a589056d3cfdc923f02969a6672/generalization.py#L181 each fine tune dataset draws from a train or test csv. Those csv are generated in utils/prepare.py And that function splits them randomly. I just check out our HPC and I still have those train test files. Would you like them? One for each dataset. Which datasets are you using? I think the larger issue here is that we need better reproducibility. I can put the train test splits up on zenodo with the rest of the records. https://zenodo.org/records/5033174
|
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
I'd need alot more information to comment here. Can I see your models? Your learning rates? Any preprocessing steps? Have you compared your models to the ones in the BirdDetector repo? Let's start by getting those train/test splits up on zenodo. As is shown in Figure 4, each fine-tuned model, the ones released on zenodo, has the maximum number of train annotations in it, they are all different sizes. You can see this in Figure 4, the one pasted above, notice for example the 'Seabirds - Indian Ocean' dataset doesn't go beyond 5,000, because there are not more than 10,000 annotations in the dataset.
Is there something in the manuscript that made this confusing? Good to know for the others and in the future. I'm re-reading the paper, and I'm guessing because in Table 3, there is a F1 score from a finetuned model of 1000 annotations? This was just for comparison to the Local only models of similar sizes. Perhaps we should have named this model something like 'Finetune - small' to make it clearer than Finetuned in other parts of the paper that refer to models with all available training data. |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
We have been rigorously studying your BirdDetector model developed using the Deep Forest package and are quite impressed with its performance. We do, however, have some specific questions about the model's implementation and training results. We hope you could shed some light on these aspects.
Sampling Process: Upon reviewing the code repository on GitHub (BirdDetector Repo), it seems the model relies on randomly selected image samples for training. Could you confirm whether the fine-tuned model available online was trained on optimally sampled images or if the sampling was done randomly? Also are the finetuned model provided based on training with 1000 samples
Fine-Tuned Model Performance: We noticed that the online fine-tuned models perform exceptionally well on various datasets. However, our attempts to train a local-only model have yielded results that are only closely comparable. Could you provide details on the number of training samples used in these fine-tuned models? In the paper (Figure 2), it's mentioned that fine-tuning with just 1000 annotations results in 50% performance compared to the local-only model. This seems contradictory to the high performance of the online models.
We would greatly appreciate any clarification or insights you could offer on these matters. If there is something we have misunderstood, please do let us know. We value your contributions to this field and look forward to your response.
Beta Was this translation helpful? Give feedback.
All reactions