Animal Welfare: AI tracks emotional state of ...
Animal Welfare

AI tracks emotional state of cows and pigs

biorxiv
Sample of images from the data set. Facial features of pigs and cows expressing varying emotions.
Sample of images from the data set. Facial features of pigs and cows expressing varying emotions.

THE NETHERLANDS, Wageningen. An animal scientist with Wageningen University & Research in the Netherlands has created an artificial-intelligence-based application that can gauge the emotional state of farm animals based on photographs taken with a smartphone. In his paper uploaded to the bioRxiv preprint server, Suresh Neethirajan describes his app and how well it worked when tested.

Emotions play an indicative and informative role in the investigation of farm animal behaviors. Systems that respond and can measure emotions provide a natural user interface in enabling the digitalization of animal welfare platforms. The faces of farm animals can be one of the richest channels for expressing emotions. Emotions sensing offers a vast amount of potential for improving animal welfare and animal-human interactions.

How does it work?

The system is called WUR Wolf and is based on several pieces of technology: the YOLO Object Detection System, the YOLOv4 that works with a convolution neural network and Faster R-CNN, which also allows for detection of objects, but does so with different feature sets. For training, he used the Nvidia GeForece GTX 1080 Ti GRP running on a CUDA 9.0 computer. The data consisted of thousands of images of cows and pigs taken with a smartphone from six farms located in several countries with associated classification labels indicating which physical features could be associated with which mood.

Using Python-based algorithms, the system detects and tracks the facial features of cows and pigs, analyzes the appearance, ear postures, and eye white regions, and correlates with the mental and emotional states of the farm animals. The software detects 13 facial actions and 9 emotional states, including whether the animal is aggressive, calm, or neutral. Detecting expressions of farm animals simultaneously in real time makes many new interfaces for automated decision-making tools possible for livestock farmers.

App supports better human-animal-interaction

Prior research and anecdotal evidence has shown that farm animals are more productive when they are not living under stressful conditions. This has led to changes in farming practices, such as shielding cows' eyes from the spike that is used to kill them prior to slaughter to prevent stress hormones from entering the meat.

More recent research has suggested that it may not be enough to shield farm animals from stressful situations—adapting their environment to promote peacefulness or even playfulness can produce desired results, as well. Happy cows or goats, for example, are likely to produce more milk than those that are bored. But as Neethirajan notes, the emotional state of an animal can be quite subjective, leading to incorrect conclusions. To address this problem, he adapted human face recognition software for use in detecting emotions in cows and pigs.

Testing showed the system to be approximately 85% accurate. Neethirajan suggests more tweaks are needed, but notes that once it is finished, it could serve as a useful tool for improving the quality of products produced via continuous monitoring of livestock.

Source: Phys.org
stats