DiscoverEA Forum Podcast (All audio)“AnimalHarmBench 2.0: Evaluating LLMs on reasoning about animal welfare” by Sentient Futures (formerly AI for Animals)
“AnimalHarmBench 2.0: Evaluating LLMs on reasoning about animal welfare” by Sentient Futures (formerly AI for Animals)

“AnimalHarmBench 2.0: Evaluating LLMs on reasoning about animal welfare” by Sentient Futures (formerly AI for Animals)

Update: 2025-11-05
Share

Description

We are pleased to introduce AnimalHarmBench (AHB) 2.0, a new standardized LLM benchmark designed to measure multi-dimensional moral reasoning towards animals, now available to use on Inspect AI.

As LLM's influence over policies and behaviors of humanity grows, its biases and blind spots will grow in importance too. With the original and now-updated AnimalHarmBench, Sentient Futures aims to provide an evaluation suite to judge LLM reasoning in an area in which blind spots are especially unlikely to get corrected through other forms of feedback: consideration of animal welfare.

In this post, we explain why we iterated upon the original benchmark and present the results and use cases of this new eval.

What Needed to Change

AHB 1.0 — presented in the AI for Animals and FAccT conferences in 2025 — attempts to measure the risk of harm that LLM outputs can have on animals. It can still play an important role in certain activities that require this such as compliance with parts of the EU AI Act Code of Practice. However, it faced several practical and conceptual challenges:

  • Reasoning evaluation: While AHB 1.0 was good for measuring how much LLM outputs increase the risk of harm to [...]

---

Outline:

(00:59 ) What Needed to Change

(02:33 ) A More Comprehensive Approach

(02:37 ) Multiple dimensions

(04:53 ) Other new features

(05:31 ) What we found

(05:56 ) Example Q&A scores

(06:32 ) Results

(08:04 ) Why This Matters

(09:16 ) Acknowledgements

(09:29 ) Future Plans

---


First published:

November 5th, 2025



Source:

https://forum.effectivealtruism.org/posts/nBnRKpQ8rzHgFSJz9/animalharmbench-2-0-evaluating-llms-on-reasoning-about


---


Narrated by TYPE III AUDIO.


---

Images from the article:

Radar chart comparing AI model performance across thirteen ethical and reasoning dimensions.

Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.

Comments 
loading
In Channel
loading
00:00
00:00
1.0x

0.5x

0.8x

1.0x

1.25x

1.5x

2.0x

3.0x

Sleep Timer

Off

End of Episode

5 Minutes

10 Minutes

15 Minutes

30 Minutes

45 Minutes

60 Minutes

120 Minutes

“AnimalHarmBench 2.0: Evaluating LLMs on reasoning about animal welfare” by Sentient Futures (formerly AI for Animals)

“AnimalHarmBench 2.0: Evaluating LLMs on reasoning about animal welfare” by Sentient Futures (formerly AI for Animals)