We package useful components developed in the Camera Traps project into APIs and host them using the AI for Earth API Framework on Azure. This folder contains the source code of the APIs and documentation on how to set them up. To see the APIs in action or start using them for your own applications, visit our APIs homepage to check out our demo apps and request a key.
Our animal detection model (MegaDetector) trained on camera trap images from a variety of ecosystems is exposed through two APIs, one for real-time applications or small batches of test images (synchronous API), and one for processing large collections of images (batch processing API). These APIs can be adapted to deploy any algorithms or models – see our tutorial in the AI for Earth API Framework repo.
This API’s /detect
endpoint processes up to 8 images at a time, and optionally returns copies of the input images annotated with the detection bounding boxes. This API powers the demo web app.
To build the API, first download the MegaDetector model file to detector_synchronous/api/animal_detection_api/model
.
This API runs the detector on up to 2 million images in one request using Azure Machine Learning Service’s Managed Compute functionality, formerly known as Batch AI. To use this API the input images need to be copied to Azure Blob Storage. Please see the user guide and get in touch with us if you’re interested in processing camera trap images this way.
The batch_processing folder includes the API itself, tools for working with the results the API generates, and support for integrating our API output with other tools.