Toward collaborative artificial intelligence development for animal well-being

Jennifer J. Sun Department of Computer Science, College of Computing and Information Science, Cornell University, Ithaca, NY

Search for other papers by Jennifer J. Sun in
Current site
Google Scholar
PubMed
Close
 PhD

Abstract

This review focuses on opportunities and challenges of future AI developments in veterinary medicine, from the perspective of computer science researchers in developing AI systems for animal behavior analysis. We examine the paradigms of supervised learning, self-supervised learning, and foundation models, highlighting their applications and limitations in automating animal behavior analysis. These emerging technologies present future challenges in data, modeling, and evaluation in veterinary medicine. To address this, we advocate for a collaborative approach that integrates the expertise of AI researchers, veterinary professionals, and other stakeholders to navigate the evolving landscape of AI in veterinary medicine. Through cross-domain dialogue and an emphasis on human and animal well-being, we can shape AI development to advance veterinary practice for the benefit of all.

Abstract

This review focuses on opportunities and challenges of future AI developments in veterinary medicine, from the perspective of computer science researchers in developing AI systems for animal behavior analysis. We examine the paradigms of supervised learning, self-supervised learning, and foundation models, highlighting their applications and limitations in automating animal behavior analysis. These emerging technologies present future challenges in data, modeling, and evaluation in veterinary medicine. To address this, we advocate for a collaborative approach that integrates the expertise of AI researchers, veterinary professionals, and other stakeholders to navigate the evolving landscape of AI in veterinary medicine. Through cross-domain dialogue and an emphasis on human and animal well-being, we can shape AI development to advance veterinary practice for the benefit of all.

Introduction

Artificial intelligence is rapidly reshaping our world, transforming how we process information, make decisions, and interact with our environment. In veterinary medicine, this transformative power holds tremendous potential for improving animal well-being1,2; for example, computational systems that track behavioral, genomic, and diagnostic data throughout an animal’s life to deliver precise, personalized medicine; unified data repositories; and AI models that not only predict disease outbreaks but also suggest targeted interventions to safeguard population health. The application of AI systems for animal behavior analysis in fields such as behavioral neuroscience35 has demonstrated the potential of this technology to automate data analysis, discover new patterns, and unify data workflows across research groups. With the perspective of a computer science researcher in this area, this review will discuss opportunities and challenges that different AI paradigms present for veterinary medicine, encompassing aspects of animal health, veterinary practice, and the human-animal bond.

Veterinary medicine is certainly positioned to benefit from the AI revolution, but its development and deployment demands careful consideration. As a discipline deeply intertwined with human, animal, and environmental health,6,7 decisions in veterinary medicine can have far-reaching consequences. This interconnectedness, recognized since the 1800s with the observation of similar disease processes in humans and animals,8 is now amplified by AI. For instance, imagine AI-powered tools that allow pet owners to submit saliva samples to generate genomic profiles for their pets, which can then be linked and analyzed through social media platforms, facilitating connections between pet owners with similar pets or concerns. These systems offer potential benefits for forming new connections between humans and animals (eg, family trees) as well as improving disease prediction and personalized medicine. However, this also raises critical questions about insurance practices, breeding decisions, and even pet adoption trends. Furthermore, consider the use of AI systems that can track wildlife and predict disease outbreaks (eg, avian flu). These models could integrate data from various sources, such as animal tracking collars, environmental sensors, and public health records, to predict and mitigate zoonotic disease outbreaks. While this has the potential to safeguard both human and animal health, it also raises ethical considerations about data privacy, surveillance, and potential interventions that could negatively disrupt ecosystems. Navigating these complexities and ensuring responsible AI development requires a collaborative approach that prioritizes the well-being of all stakeholders.

Veterinary medicine presents unique challenges and opportunities for AI. Unlike human medicine, in which regulations for the use of new technologies are more developed, the regulatory landscape in veterinary medicine is still evolving.9,10 Deployment of AI for veterinary practices in the current landscape has been reviewed in prior work such as Basran and Appleby.1 New systems in veterinary medicine must contend with diverse species, varying clinical contexts, and potential biases in datasets. To ensure responsible development, it is crucial to prioritize the well-being of all involved: animals, veterinarians, pet owners, the environment, and society as a whole. This requires close collaboration between AI researchers, veterinary professionals, public health experts, and policymakers to navigate the ethical and practical considerations unique to this domain.

This review will first discuss the perspective of model developers working with scientists on AI for animal behavior analysis, providing insights into the practical realities of developing and deploying these technologies. It will then outline future opportunities and challenges for AI in veterinary medicine more broadly, emphasizing the importance of collaboration across fields and an interconnected approach to ensure these powerful tools are used responsibly and effectively to improve the well-being of all.

Current AI Paradigms

Before exploring the future of AI in veterinary medicine, we direct the reader’s attention to the area of animal behavior analysis in scientific research where AI is making a significant impact.

Animal behavior analysis is crucial across diverse scientific disciplines, including ecology, ethology, and behavioral neuroscience.1113 For example, when neuroscientists study the brain circuitry that regulates animal behavior, they need to define and categorize actions from large amounts of recorded behavioral videos. Manual analysis is prohibitively expensive with rapidly increasing amounts of data. However, automating this process is challenging because different laboratories study different behaviors, with different definitions, using different experimental designs.11,14 We will discuss research in the direction of scientist-in-the-loop AI systems toward tackling these challenges.5,15,16 These systems are developed in collaboration with neuroscientists and computer scientists, in order to transform diverse and complex experimental data into interpretable representations.

A comprehensive review of tools for behavior quantification can be found in a few other works.4,11,14 This section will focus specifically on the impact of 3 different AI paradigms on method development and deployment: supervised learning, self-supervised learning, and foundation models (Figure 1).

Figure 1
Figure 1

We discuss three AI paradigms for animal behavior analysis. Supervised learning trains models on labeled datasets to identify human-defined behaviors like ‘sniff’ or ‘attack.’ Self-supervised learning uses ‘pretext tasks’ to learn from unlabeled data, such as predicting future movements to learn behavioral representations. Foundation models, trained on massive datasets, learn general-purpose representations adaptable to various tasks.

Citation: Journal of the American Veterinary Medical Association 2025; 10.2460/javma.24.10.0650

Supervised learning

Supervised learning is a widely used paradigm in animal behavior analysis. These methods have enabled researchers to automate the process of identifying and categorizing behaviors from vast amounts of data. This approach relies on large, human-annotated datasets to train models. During training, the model is exposed to many examples (like a student studying many cases) and learns to perform a specific task or group of tasks, such as identifying different behaviors from video recordings. The model learns the task by adjusting its internal parameters to improve its performance.

Once trained, these models can be applied to new, unseen data, similar to how a veterinarian applies their knowledge to diagnose new patients. However, the model’s performance on new data depends on its ability to generalize: to apply what it has learned to different situations. For example, it would be difficult for an AI model trained only on videos of mice to accurately analyze the behavior of birds, or even for models trained only on videos of mice from one laboratory to generalize to mice from a different laboratory. In this case, the AI model may require fine-tuning to adjust to data that differ from its original training set. This fine-tuning involves further refining the model’s internal parameters to better match the characteristics of the new data.

The ability to generalize effectively often depends on the size and diversity of the original training dataset. A model trained on a wide range of behaviors, species, and experimental setups is more likely to generalize better to new and varied data. Therefore, before using an AI model in your workflow, it’s crucial to understand the type of data on which it was trained and evaluate it on your own data. This will help you assess whether the model is suitable for your specific needs or requires further adaptation.

One example of supervised learning in animal behavior analysis is the Mouse Action Recognition System (MARS),3 a collaborative effort between behavioral neuroscientists and computer scientists. The MARS tackles the challenge of automating behavior analysis by first requiring human experts to annotate video data. This process involves labeling anatomical landmarks (also called keypoints) on the animal’s body, such as the nose, tail, and paws. A model is then trained with the use of human annotations to detect these keypoints on new images; this task is called pose estimation, in which pose refers to the overall configuration of the body defined by keypoints. The next step requires users to assign behavioral labels to different video segments, like sniffing, grooming, or mounting. The model then learns to recognize these behaviors by matching the visual information (keypoints and movement patterns) to the human-provided labels. This workflow of first performing pose estimation, then behavior classification is common across supervised learning setups and has also been reviewed in multiple works.4,11,17 Notably, tools for pose estimation include DeepLabCut18 and SLEAP,19 which has been applied to label keypoints on species including flies and gerbils, as well as to track multiple animals in videos. Other tools for behavior classification include SimBA20 and DeepEthogram,21 in which the goal is to map each input video frame to a human-defined behavior label.

While supervised learning has significantly advanced the field of animal behavior analysis, several key challenges remain. (1) Annotation bottleneck: these models require a labeled training set to learn new tasks. Creating large, accurately annotated datasets is time-consuming and expensive, often requiring significant manual effort from experts, which limits the scalability of supervised learning approaches. For example, MARS is developed with approximately 14 hours of frame-level behavior annotations, and 15,000 frames of pose annotations across training and evaluation data. Researchers are exploring various techniques to address this challenge, including active learning, transfer learning, and semi-supervised learning. (2) Generalization to new data: supervised learning models trained on a single dataset may not generalize well to new datasets with different experimental setups, animal strains, or behavioral definitions. This is because the model might have learned features that are specific to the training dataset that don’t transfer well to new situations. For example, the MARS pose estimators were trained with images from a top-view camera, and retraining with new data was needed for camera angles from the side. Additionally, the model was initially trained with black and white mice and might not generalize well to other colors or species. This highlights a common challenge in supervised learning: models often need to be adapted to new settings or types of data. Techniques to tackle this include transfer learning via fine-tuning, as well as data augmentation, which artificially increases the diversity of training data. (3) Interpretability: many powerful models are complex ‘black boxes’ due to their architectures and large numbers of parameters.17,22 This can make it difficult for researchers to trust the model’s output and gain insights into the underlying factors driving animal behavior. Researchers are actively developing techniques to make AI models more interpretable, such as highlighting important features that influence decisions or identifying representative examples that activate specific model components.

Self-supervised learning

While supervised learning has been impactful,4,11,17 it often faces limitations due to the reliance on labeled data. Self-supervised learning23 offers an alternative approach by leveraging the inherent structure of the data itself to generate supervisory signals, eliminating the need for explicit labels. In self-supervised learning,24 pretext tasks are designed to encourage the model to learn relevant representations from unlabeled data without relying on explicit labels. These pretext tasks include predicting missing parts of the data or reconstructing the original data from a scrambled version. These tasks enable the model to learn important features and relationships within the data without labels (eg, training the model to predict future trajectories5), which can then be used for other downstream tasks (eg, using the learned representation as input for behavior classification). The performance of a self-supervised model is often judged by how well it performs on a separate task, like accurately classifying different behaviors, after it has learned from the unlabeled data.

One example of a self-supervised framework for studying behavior is task programming,5 which uses pretext tasks designed by experts to learn representations of behavior in a self-supervised way, in addition to predicting future trajectories. These tasks represent attributes important for studying behavior, such as the distance between animals or speed of each animal. This encourages the model to learn about spatial and temporal relationships in trajectory data, as well as important features of animal behavior defined by experts. Once trained, these representations can be applied to learn human-defined behaviors with significant reduction in annotation requirements. Other frameworks for self-supervised learning of behavior are reviewed in Luxem et al4 and notably includes VAME,25 which uses a variational autoencoder to map trajectory data to a lower dimensional space and perform clustering for behavior discovery. Instead of behavior, another line of work15,26 uses self-supervised learning to discover keypoints from behavioral videos. B-KinD trains a model to reconstruct the movements in the original video (eg, difference across 2 video frames) in order to enable the model to learn meaningful representations of the animal’s body and movement, even without explicit keypoint annotations. Finally, there are also frameworks that use unsupervised learning,27 which learns structure from data without a pretext task.

One of the key advantages of self-supervised learning is its ability to leverage large amounts of unlabeled data,24 which is more readily available compared to expert-labeled data. This makes these techniques especially valuable in veterinary medicine, in which labeled data for diverse animal species, complex patient records, and varied imaging modalities (eg, radiographs, ultrasounds, histopathology) can be scarce or costly to obtain. By leveraging unlabeled datasets, training with self-supervised learning compared to only supervised data can lead to more robust and generalizable models. However, as with supervised learning, if a self-supervised model is only trained on a limited set of unlabeled data (e.g., videos all from a single strain of mice), it is also unlikely to generalize to new settings. This can be seen with B-KinD, which has been successfully used to discover keypoints on animals like mice and fruit flies. While a model trained on black and white mice might not directly transfer to videos with two black mice, it demonstrates efficient performance when applied to videos with similar characteristics, requiring only a few minutes of training video for such cases. Furthermore, designing effective pretext tasks is crucial in self-supervised learning.24 The task should be challenging enough to encourage the model to learn meaningful representations but not so difficult that it becomes impossible to solve. Collaborations between AI researchers and veterinarians could ensure that the learned representations are relevant to clinically relevant downstream tasks.

Foundation models

Building on advances in self-supervised learning, foundation models represent a recent paradigm shift in AI.28 These models are trained on massive, diverse datasets, often encompassing internet-scale data. During training, foundation models use self-supervised objectives that enable them to learn representations without explicit labels across massive datasets. This allows them to be adapted to a wide range of downstream tasks with minimal or no fine-tuning. A prominent example29 is large language models like GPT-4, which has demonstrated a wide range of capabilities in natural language understanding and generation after being trained on vast amounts of text data. This represents a departure from self-supervised methods discussed in the previous section in terms of training data scale, general-purpose nature of the learned model, and emergent capabilities.

In the domain of video analysis, foundation models are also emerging as a promising tool.3032 One example, which has been applied to animal behavior analysis, is VideoPrism,33 a video foundation model trained on a massive dataset of approximately 600 million videos. VideoPrism is a video encoder that maps input videos to semantically meaningful representations, which then empower a range of downstream tasks in computer vision, including captioning, video question answering, and video-text retrieval. In addition, this model, with minimal adaptation, is applicable to a variety of animal behavior analysis tasks,34 including localization, behavior classification, and retrieval of videos with similar behaviors.

Training and developing foundation models is a computationally intensive process that requires significant resources and expertise (eg, estimation for a single training run of a 1.5 billion–parameter model is at least $80,000, depending on factors such as hardware, dataset size, and training time).35 For example, these models often contain a large number of parameters (eg, a 70 billion–parameter open-sourced language model36) to fit the massive training datasets. However, once trained, these models can be readily used by researchers and practitioners without the need for extensive model training or fine-tuning.28 This provides researchers in animal behavior with access to powerful AI systems that could potentially work out of the box on their data. These general-purpose capabilities also hold promise for applications in veterinary medicine, such as analyzing patient videos, predicting disease risk, and assisting with clinical decision-making.

Furthermore, foundation models provide a strong starting point for specialization. Domain-specific models can be derived from foundation models through fine-tuning on smaller, targeted datasets.28 In the future, a foundation model trained on diverse animal videos could be further fine-tuned on a dataset of videos specifically focused on canine behavior to improve its performance on canine-specific tasks. Alternatively, a decoder module, which translates the general representations learned by the foundation model into task-specific outputs, can be added to adapt it to specific tasks or datasets. VideoPrism, originally trained on internet videos, learns a general-purpose video representation that is compatible with various decoders (eg, classification or localization) for animal behavior analysis.34 In a similar fashion, it is also possible to train decoders to identify specific diseases or conditions from general-purpose video foundation model representations.

Future Opportunities and Challenges

The AI paradigms discussed in the previous section—supervised, self-supervised, and foundation models—each present unique opportunities and challenges for animal behavior analysis as well as veterinary medicine. While the current focus in the field is on leveraging large-scale datasets to train generalizable models, several challenges remain. Here, we discuss key challenges related to data, model development, and evaluation, focusing on their implications for users of AI systems in veterinary practice. As the field of AI rapidly evolves, it is crucial to proactively anticipate and address these emerging opportunities and challenges to ensure the responsible and effective integration of new technologies into veterinary medicine. Readers interested in a more comprehensive discussion of the evolving regulatory landscape and ethical frameworks for AI in veterinary medicine are referred to recent reviews on this topic.1,9,10

Data

For modern AI models to generalize to unseen settings (eg, new clinics, new patients, new tasks), they require large, diverse training datasets. There is an opportunity to pool datasets containing heterogeneous data modalities, such as electronic health records, medical images (x-rays, ultrasounds, etc), pathology reports, genomic sequences, and even sensor data from wearable devices, across various sources, such as veterinary hospitals, research institutions, and diagnostic laboratories. Assuming a sufficiently large and representative dataset based on current trends in foundation model research,28,37 this aggregated dataset has the potential to enable the development of more robust models that work out of the box without fine-tuning to tackle tasks such as more accurate diagnoses, personalized treatments, and improved disease prediction.

However, data sharing and aggregation raise important questions about data ownership, privacy, security, and usage rights. Clear agreements and ethical guidelines are needed to ensure responsible data handling and to address potential biases in the data, such as underrepresentation of certain breeds or species, overrepresentation of specific clinical settings, or variations in data collection practices. Such biases could lead to inaccurate predictions or limited generalizability to diverse populations. Furthermore, the ownership, control, and potential liability of the trained AI models, especially when developed with the use of collaborative efforts or sensitive data, raise complex questions about intellectual property rights, data usage agreements, and potential conflicts of interest. Artificial intelligence developers also need to ensure the privacy of clinical data, as training data can be extracted from trained generative models such as large language models.38 Transparent agreements and ethical guidelines are needed to ensure fair and responsible use of AI models and data.7

Finally, because the performance of AI systems is highly dependent on their training data, practitioners and users should be aware of the data on which the model they are using is trained (or have a way to evaluate the model carefully in their setting) before deploying it in clinical practice.

Model

The diverse range of data modalities and tasks in veterinary medicine—from radiographs and bloodwork to clinical notes and genomic data—presents a unique opportunity to push the capabilities of modern AI systems. However, this also presents a challenge: effectively integrating these diverse data types into a cohesive, interpretable system. For example, imagine a computational system that can analyze a radiograph; cross-reference it with the patient’s breed, age, and clinical history; and then suggest a diagnosis while also providing the reasoning behind its conclusion, drawing on established veterinary knowledge and best practices. To achieve this, we need models that not only leverage the power of deep learning on large datasets but also incorporate decades of accumulated veterinary expertise. This human-in-the-loop approach, in which AI learns from and collaborates with veterinarians, is crucial not only for improving animal care but also for building trust in these systems.

One step toward this approach is to develop models that can explain their reasoning in a way that is meaningful to veterinarians. Veterinary medicine encompasses a wide range of specialties, each with its own unique knowledge base and clinical focus. For instance, an explanation of patient conditions and diagnosis might vary for a radiologist, oncologist, or cardiologist. To achieve this, close collaboration between AI researchers and veterinary professionals is crucial. By working together, we can ensure that these models are developed and deployed in a way that addresses the specific needs, concerns, and ethical considerations of the veterinary community. This collaboration can help bridge the gap between computational expertise and domain-specific knowledge, leading to more effective and trustworthy AI solutions for veterinary medicine.

Finally, there is an opportunity to enhance the human-animal bond by providing tools to continuously monitor and analyze animal behavior, communication, and emotional states from videos, wearables, or other devices. This can lead to improved response to the needs of animals (eg, improving livestock health39,40) as well as a strengthened connection between humans and their animal companions. For example, AI-powered apps could provide personalized recommendations for pet care and training, tailored to the specific needs and personality of the animal. However, it’s important to consider how the technology is used in a way that complements, not replaces, human-animal interactions.

Evaluation

Evaluating the impact of AI on animal well-being is crucial but complex, requiring a multifaceted approach that goes beyond traditional metrics like accuracy.41 While AI offers promising applications in continuous monitoring, early disease detection, and personalized interventions, assessing its true impact on animal welfare can be challenging. To tackle this challenge, we need to develop robust evaluation metrics that capture the multifaceted aspects of animal well-being, including the following: (1) physical health indicators such as recovery time, longevity, and disease rate and (2) behavioral and emotional state indicators such as social interactions, vocalizations, and signs of stress. In addition, we also need metrics to assess how AI impacts the veterinary practice itself, such as client and veterinarian satisfaction, diagnostic accuracy, and efficiency.

Furthermore, establishing standardized protocols for data collection and evaluation facilitates comparability across different studies, enabling us to better measure the progress of model development in the field. This is particularly crucial as the field moves toward more collaborative and open research. Notable examples of standardization effort from other fields include data sheets for documenting data collection and usage from machine learning42; the Neurodata Without Borders format, introduced for organizing and sharing neurophysiology data43; and guidelines for using and developing open-source tools for behavioral video analysis (such as the use of large-scale standardized benchmarks).4 In human medicine, large-scale datasets with standardized formats have enabled the development of powerful models for language (eg, Med-PaLM44) and medical image segmentation (eg, Med-SAM45).

Ethical considerations must be at the forefront of any evaluation process. Collaborations between AI developers and veterinarians are needed to ensure that these models prioritize the best interests of stakeholders and avoid any unintended consequences. Ideally, new systems or modifications to existing ones should undergo comprehensive evaluations on existing datasets or simulated environments before deployment. This allows for controlled testing and refinement, minimizing risks to real animals. By combining robust evaluation metrics, ethical considerations, and simulated testing, we can pave the way for the responsible and effective use of emerging technologies to improve animal well-being.

Call for Collaborations

As AI’s impact on veterinary medicine grows, we stand at a crossroads. Will AI-powered tools work with veterinarians to improve animal health worldwide and enable new ways for humans and animals to connect? Or will the rise of commercial AI systems exacerbate existing challenges, potentially influencing shortages in the veterinary workforce, impacting human expertise, increasing stress, and disrupting work-life balance? Will these algorithms ultimately help or hinder the practice of veterinary medicine and the connection between humans and animals? The choices we make today will shape the future of veterinary medicine for years to come.

To ensure a future in which AI benefits all, we must actively forge a path of responsible development and deployment. This requires a collaborative effort, bringing together diverse perspectives and expertise (Figure 2):

  • Datasets and benchmarks: The trajectory of AI is shaped not only by researchers but also by the availability of high-quality data and benchmarks. To develop AI for all, we need diverse and representative datasets that encompass a wide range of animal species, breeds, ages, health conditions, and geographic locations. This enables us to develop models that generalize well across populations and clinical settings and help mitigate bias (eg, overrepresentation of certain species such as cats and dogs). Furthermore, establishing clear data ownership and sharing agreements is crucial when building these datasets. To encourage wider participation, open data initiatives could be considered, but we need to carefully balance the benefits of data sharing with the need to protect patient privacy and confidentiality. Finally, developing standardized benchmarks and evaluation metrics will allow for more fair comparisons between different AI models and facilitate progress in the field.

  • Domain expertise: Decades of accumulated human expertise already exist in veterinary medicine. Integrating AI systems with this knowledge requires a multifaceted approach. First, human-in-the-loop systems can empower veterinarians to actively participate in the entire model lifecycle, from model development and deployment to ongoing evaluation. This can involve providing feedback on model predictions, improving existing algorithms, and contributing training data. Second, we need human-interpretable models to enable bidirectional knowledge transfer. This is an active area with works in explainable AI and machine learning,46 mechanistic interpretability,47 and neurosymbolic modeling.16,48 Further research in these areas enables veterinarians to both integrate their domain expertise into the model and extract new insights and knowledge from the model’s outputs. By designing AI systems that actively collaborate with veterinarians, we can develop technology that prioritizes the well-being of both animals and the veterinary profession.

  • Accelerate discovery: To fully realize the potential of AI in veterinary medicine, we need new approaches that leverage the unique data and domain knowledge in the field. For example, research projects that leverage AI to analyze complex veterinary datasets can lead to breakthroughs in areas such as disease diagnostics, drug discovery, and personalized medicine. Funding agencies can play a crucial role in facilitating these efforts by prioritizing grants that support collaborative projects between veterinary researchers and AI experts. Additionally, developing open-source AI tools and platforms specifically designed for veterinary research can promote accessibility and collaboration. These tools should be tailored to the specific needs of veterinary researchers, providing user-friendly interfaces and functionalities that facilitate data analysis, model training, and knowledge sharing. Furthermore, supporting initiatives that encourage data sharing and collaboration between veterinary researchers and computer scientists, such as workshops, conferences, and shared data repositories, can maximize the potential of AI to accelerate research and improve animal health.

  • Next generation: Integrating AI education into veterinary curricula is crucial to prepare the next generation of veterinarians for future technologies; for example, a curriculum that covers topics such as AI ethics, data literacy, and the responsible use of new technologies. To do so, we need to overcome challenges related to curriculum development, faculty training, and access to resources. This may involve updating existing curricula, incorporating computational modules into relevant courses, and providing faculty with the necessary training and resources to effectively teach AI concepts. Creating AI-focused internships and training programs for veterinary students can provide valuable hands-on experience with these tools and technologies. These programs can offer students opportunities to work on real-world projects, collaborate with AI researchers, and gain practical skills in applying these models to veterinary problems, ensuring a smooth transition for the next generation of veterinarians fluent in these emerging tools.

  • Cross-domain conversations: Fostering opportunities for increased dialogue between AI researchers and veterinarians is equally important. Initiatives like the Symposium on Artificial Intelligence in Veterinary Medicine, which facilitates interdisciplinary exchanges between these communities, are invaluable for bridging the gap. These conversations are essential for fostering mutual understanding, identifying shared goals, and driving collaborative innovation. Funding agencies, academic institutions, and industry should prioritize efforts to create more platforms for dialogue, such as workshops, conferences, and online forums, and actively foster interdisciplinary research collaborations between AI researchers and veterinary professionals.

Figure 2
Figure 2

Toward collaborative effort across domains to build AI systems to improve well-being. Integrating data and expertise from veterinary medicine, computer science, and other fields enables the creation of robust AI tools, benchmarks, and feedback mechanisms to evaluate the impact of these new models across diverse settings.

Citation: Journal of the American Veterinary Medical Association 2025; 10.2460/javma.24.10.0650

We hope this review, alongside others,1,2,7,9,10 serves as a catalyst for these vital discussions. By fostering a deeper understanding of the unique challenges and opportunities in veterinary medicine among AI researchers, and by equipping veterinary professionals with the knowledge to critically evaluate the potential and limitations of AI, we can collaboratively build this transformative technology to improve the health and well-being of all.

None reported.

Disclosures

The author discloses that Gemini was used to proofread this manuscript.

Funding

The author has nothing to disclose.

References

  • 1.

    Basran PS, Appleby RB. What’s in the box? A toolbox for safe deployment of artificial intelligence in veterinary medicine. J Am Vet Med Assoc. 2024;262(8):1090-1098. doi:10.2460/javma.24.01.0027

    • Search Google Scholar
    • Export Citation
  • 2.

    Akinsulie OC, Idris I, Aliyu VA, et al. The potential application of artificial intelligence in veterinary clinical practice and biomedical research. Front Vet Sci. 2024;11:1347550. doi:10.3389/fvets.2024.1347550

    • Search Google Scholar
    • Export Citation
  • 3.

    Segalin C, Williams J, Karigo T, et al. The Mouse Action Recognition System (MARS) software pipeline for automated analysis of social behaviors in mice. eLife. 2021;10:e63720. doi:10.7554/eLife.63720

    • Search Google Scholar
    • Export Citation
  • 4.

    Luxem K, Sun JJ, Bradley SP, et al. Open-source tools for behavioral video analysis: setup, methods, and best practices. eLife. 2023;12:e79305. doi:10.7554/eLife.79305

    • Search Google Scholar
    • Export Citation
  • 5.

    Sun JJ, Kennedy A, Zhan E, Anderson DJ, Yue Y, Perona P. Task programming: learning data efficient behavior representations. Proc IEEE Comput Soc Conf Comput Vis Pattern Recognit. 2021;2021:2875-2884. doi:10.1109/CVPR46437.2021.00290

    • Search Google Scholar
    • Export Citation
  • 6.

    Binot A, Duboz R, Promburom P, et al. A framework to promote collective action within the One Health community of practice: using participatory modelling to enable interdisciplinary, cross-sectoral and multi-level integration. One Health. 2015;1:44-48. doi:10.1016/j.onehlt.2015.09.001

    • Search Google Scholar
    • Export Citation
  • 7.

    Destoumieux-Garzón D, Mavingui P, Boetsch G, et al. The One Health concept: 10 years old and a long road ahead. Front Vet Sci. 2018;5:14. doi:10.3389/fvets.2018.00014

    • Search Google Scholar
    • Export Citation
  • 8.

    Gyles C. One medicine, one health, one world. Can Vet J. 2016;57(4):345-346.

  • 9.

    Cohen EB, Gordon IK. First, do no harm. Ethical and legal issues of artificial intelligence and machine learning in veterinary radiology and radiation oncology. Vet Radiol Ultrasound. 2022;63(suppl 1):840-850. doi:10.1111/vru.13171

    • Search Google Scholar
    • Export Citation
  • 10.

    Bellamy JEC. Artificial intelligence in veterinary medicine requires regulation. Can Vet J. 2023;64(10):968-970.

  • 11.

    Pereira TD, Shaevitz JW, Murthy M. Quantifying behavior to understand the brain. Nat Neurosci. 2020;23(12):1537-1549. doi:10.1038/s41593-020-00734-z

    • Search Google Scholar
    • Export Citation
  • 12.

    Gordon DM. The fusion of behavioral ecology and ecology. Behav Ecol. 2011;22(2):225-230. doi:10.1093/beheco/arq172

  • 13.

    Belzung C, Griebel G. Measuring normal and pathological anxiety-like behaviour in mice: a review. Behav Brain Res. 2001;125(1-2):141-149. doi:10.1016/S0166-4328(01)00291-1

    • Search Google Scholar
    • Export Citation
  • 14.

    Anderson DJ, Perona P. Toward a science of computational ethology. Neuron. 2014;84(1):18-31. doi:10.1016/j.neuron.2014.09.005

  • 15.

    Sun JJ, Ryou S, Goldshmid RH, et al. Self-supervised keypoint discovery in behavioral videos. Proc IEEE Comput Soc Conf Comput Vis Pattern Recognit. 2022;2022;2161-2170. doi:10.1109/CVPR52688.2022.00221

    • Search Google Scholar
    • Export Citation
  • 16.

    Sun JJ, Tjandrasuwita M, Sehgal A, et al. Neurosymbolic programming for science. ArXiv. Preprint posted online November 7, 2022. doi:10.48550/arXiv.2210.05050

    • Search Google Scholar
    • Export Citation
  • 17.

    Goodwin NL, Nilsson SRO, Choong JJ, Golden SA. Toward the explainability, transparency, and universality of machine learning for behavioral classification in neuroscience. Curr Opin Neurobiol. 2022;73:102544. doi:10.1016/j.conb.2022.102544

    • Search Google Scholar
    • Export Citation
  • 18.

    Mathis A, Mamidanna P, Cury KM, et al. DeepLabCut: markerless pose estimation of user-defined body parts with deep learning. Nat Neurosci. 2018;21(9):1281-1289. doi:10.1038/s41593-018-0209-y

    • Search Google Scholar
    • Export Citation
  • 19.

    Pereira TD, Tabris N, Matsliah A, et al. SLEAP: a deep learning system for multi-animal pose tracking. Nat Methods. 2022;19(4):486-495. doi:10.1038/s41592-022-01426-1

    • Search Google Scholar
    • Export Citation
  • 20.

    Goodwin NL, Choong JJ, Hwang S, et al. Simple Behavioral Analysis (SimBA)-an open source toolkit for computer classification of complex social behaviors in experimental animals. Nat Neurosci. 2024;27(7):1411-1424. doi:10.1038/s41593-024-01649-9

    • Search Google Scholar
    • Export Citation
  • 21.

    Bohnslav JP, Wimalasena NK, Clausing KJ, et al. DeepEthogram, a machine learning pipeline for supervised behavior classification from raw pixels. Elife. 2021;10:e63377. doi:10.7554/eLife.63377

    • Search Google Scholar
    • Export Citation
  • 22.

    Rudin C, Chen C, Chen Z, Huang H, Semenova L, Zhong C. Interpretable machine learning: fundamental principles and 10 grand challenges. Statist Surv. 2022;16:1-85. doi:10.1214/21-SS133

    • Search Google Scholar
    • Export Citation
  • 23.

    Gui J, Chen T, Zhang J, et al. A survey on self-supervised learning: algorithms, applications, and future trends. IEEE Trans Pattern Anal Mach Intell. 2024;46(12):9052-9071. doi:10.1109/TPAMI.2024.3415112

    • Search Google Scholar
    • Export Citation
  • 24.

    Balestriero R, Ibrahim M, Sobal V, et al. A cookbook of self-supervised learning. ArXiv. Preprint posted online April 24, 2023. doi:10.48550/arXiv.2304.12210

    • Search Google Scholar
    • Export Citation
  • 25.

    Luxem K, Mocellin P, Fuhrmann F, et al. Identifying behavioral structure from deep variational embeddings of animal motion. Commun Biol. 2022;5(1):1267. doi:10.1038/s42003-022-04080-7

    • Search Google Scholar
    • Export Citation
  • 26.

    Sun JJ, Karashchuk L, Dravid A, et al. BKinD-3D: self-supervised 3D keypoint discovery from multi-view videos. ArXiv. Preprint posted online December 14, 2022. doi:10.48550/arXiv.2212.07401

    • Search Google Scholar
    • Export Citation
  • 27.

    Wiltschko AB, Johnson MJ, Iurilli G, et al. Mapping sub-second structure in mouse behavior. Neuron. 2015;88(6):1121-1135. doi:10.1016/j.neuron.2015.11.031

    • Search Google Scholar
    • Export Citation
  • 28.

    Bommasani R, Hudson DA, Adeli E, et al. On the opportunities and risks of foundation models. ArXiv. Preprint posted online July 22, 2022. doi:10.48550/arXiv.2108.07258

    • Search Google Scholar
    • Export Citation
  • 29.

    Achiam J, Adler S, Agarwal S, et al. Gpt-4 technical report. ArXiv. Preprint posted online March 4, 2024. doi:10.48550/arXiv.2303.08774

  • 30.

    Yan S, Zhu T, Wang Z, et al. VideoCoCa: video-text modeling with zero-shot transfer from contrastive captioners. ArXiv. Preprint posted online December 9, 2022. doi:10.48550/arXiv.2212.04979

    • Search Google Scholar
    • Export Citation
  • 31.

    Lin B, Ye Y, Zhu B, et al. Video-llava: Learning united visual representation by alignment before projection. ArXiv. Preprint posted online November 16, 2023. doi:10.48550/arXiv.2311.10122

    • Search Google Scholar
    • Export Citation
  • 32.

    Wang Y, Li K, Li X, et al. InternVideo2: scaling foundation models for multimodal video understanding. ArXiv. Preprint posted online March 22, 2024. doi:10.48550/arXiv.2403.15377

    • Search Google Scholar
    • Export Citation
  • 33.

    Zhao L, Gundavarapu NB, Yuan L, et al. Videoprism: a foundational visual encoder for video understanding. In: Proceedings of the 41st International Conference on Machine Learning. International Conference on Machine Learning; 2024.

    • Search Google Scholar
    • Export Citation
  • 34.

    Sun JJ, Zhou H, Zhao L, et al. Video foundation models for animal behavior analysis. bioRxiv. Preprint posted online July 31, 2024. doi:10.1101/2024.07.30.605655

    • Search Google Scholar
    • Export Citation
  • 35.

    Sharir O, Peleg B, Shoham Y. The cost of training NLP models: a concise overview. ArXiv. Preprint posted online April 19, 2020. doi:10.48550/arXiv.2004.08900

    • Search Google Scholar
    • Export Citation
  • 36.

    Meta-Llama-3-70B. Accessed December 4, 2024. https://huggingface.co/meta-llama/Meta-Llama-3-70B

  • 37.

    Moor M, Banerjee O, Abad ZSH, et al. Foundation models for generalist medical artificial intelligence. Nature. 2023;616(7956):259-265. doi:10.1038/s41586-023-05881-4

    • Search Google Scholar
    • Export Citation
  • 38.

    Nasr M, Carlini N, Hayase J, et al. Scalable extraction of training data from (production) language models. ArXiv. Preprint posted online November 28, 2023. doi:10.48550/arXiv.2311.17035

    • Search Google Scholar
    • Export Citation
  • 39.

    Cerri RLA, Burnett TA, Madureira AML, et al. Symposium review: linking activity-sensor data and physiology to improve dairy cow fertility. J Dairy Sci. 2021;104(1):1220-1231. doi:10.3168/jds.2019-17893

    • Search Google Scholar
    • Export Citation
  • 40.

    Siachos N, Neary JM, Smith RF, Oikonomou G. Automated dairy cattle lameness detection utilizing the power of artificial intelligence: current status quo and future research opportunities. Vet J. 2024;304:106091. doi:10.1016/j.tvjl.2024.106091

    • Search Google Scholar
    • Export Citation
  • 41.

    Fraser D. Assessing animal well-being: common sense, uncommon science. Purdue University Office of Agricultural Research Programs. Accessed November 17, 2024. https://www.wellbeingintlstudiesrepository.org/cgi/viewcontent.cgi?article=1005&context=assawel

    • Search Google Scholar
    • Export Citation
  • 42.

    Gebru T, Morgenstern J, Vecchione B, et al. Datasheets for datasets. Commun Assoc Comput Machinery. 2021;64(12):86-92. doi:10.1145/3458723

  • 43.

    Rübel O, Tritt A, Ly R, et al. The Neurodata Without Borders ecosystem for neurophysiological data science. Elife. 2022;11:e78362. doi:10.7554/eLife.78362

    • Search Google Scholar
    • Export Citation
  • 44.

    Singhal K, Tu T, Gottweis J, et al. Towards expert-level medical question answering with large language models. ArXiv. Preprint posted online May 16, 2023. doi:10.48550/arXiv.2305.09617

    • Search Google Scholar
    • Export Citation
  • 45.

    Ma J, He Y, Li F, Han L, You C, Wang B. Segment anything in medical images. Nat Commun. 2024;15(1):654. doi:10.1038/s41467-024-44824-z

  • 46.

    Adadi A, Berrada M. Explainable AI for healthcare: from black box to interpretable models. In: Bhateja V, Satapathy SC, Satori H, eds. Embedded Systems and Artificial Intelligence. Springer; 2020:327-337. doi:10.1007/978-981-15-0947-6_31

    • Search Google Scholar
    • Export Citation
  • 47.

    Bereska L, Gavves E. Mechanistic interpretability for AI safety: a review. ArXiv. Preprint posted online April 22, 2024. doi:10.48550/arXiv.2404.14082

    • Search Google Scholar
    • Export Citation
  • 48.

    Chaudhuri S, Ellis K, Polozov O, Singh R, Solar-Lezama A, Yue Y. Neurosymbolic programming. Found Trends Program Languages. 2021;7(3):158-243. doi:10.1561/2500000049

    • Search Google Scholar
    • Export Citation
All Time Past Year Past 30 Days
Abstract Views 0 0 0
Full Text Views 1895 1895 689
PDF Downloads 496 496 181
Advertisement