DiscoverLessWrong (30+ Karma)“Base64Bench: How good are LLMs at base64, and why care about it?” by richbc
“Base64Bench: How good are LLMs at base64, and why care about it?” by richbc

“Base64Bench: How good are LLMs at base64, and why care about it?” by richbc

Update: 2025-10-06
Share

Description

This was a quick, short side-project produced during the MATS Research 8.1 extension. It's related to my group's main thread of work on black-box scheming monitoring through the connections to monitoring I explore below, but was time-boxed and pursued independently because I thought it was interesting!

Executive Summary

Figure 1. Accuracy vs. similarity threshold (0.95+) across 1700 pairs of encoding/decoding examples across a variety of datatypes and lengths. The accuracy is the proportion of the 3400 examples each model translated successfully (directly, with no reasoning or tools). Success for each task is defined by the normalised Levenshtein similarity of the answer/target pair hitting a given threshold, with a scoring requirement that model-encoded strings are decodable. Legend ordered by accuracy@1.0.
  • Introducing Base64Bench: a simple new benchmark for evaluating models on their ability to encode and decode base64.
    • Base64 encoding and decoding are reasonably complex computational tasks to do perfectly [...]

---

Outline:

(00:31 ) Executive Summary

(03:07 ) An accidental (and surprising) discovery

(08:03 ) Have LLMs actually learned the algorithm?

(09:39 ) Introducing

(13:11 ) Accuracy vs. similarity threshold

(16:02 ) Encoding vs. decoding by model

(17:00 ) Task-level breakdown

(19:37 ) Why should we care?

(21:26 ) Monitoring implications

(23:51 ) Conclusion

(25:23 ) Appendix

(25:26 ) Zoomed-in threshold sweeps

The original text contained 8 footnotes which were omitted from this narration.

---


First published:

October 5th, 2025



Source:

https://www.lesswrong.com/posts/5F6ncBfjh2Bxnm6CJ/base64bench-how-good-are-llms-at-base64-and-why-care-about


---


Narrated by TYPE III AUDIO.


---

Images from the article:

Figure 1. Accuracy vs. similarity threshold (0.95+) across 1700 pairs of encoding/decoding examples across a variety of datatypes and lengths. The accuracy is the proportion of the 3400 examples each model translated successfully (directly, with no reasoning or tools). Success for each task is defined by the normalised Levenshtein similarity of the answer/target pair hitting a given threshold, with a scoring requirement that model-encoded strings are decodable. Legend ordered by accuracy@1.0.
Figure 2. Encoding accuracy vs. similarity threshold for all models tested. The normalised Levenshtein similarity of each answer/target is compared to the similarity threshold. The accuracy is the proportion of the eval sample scores which meet a given threshold. The curves saturate at accuracy <1.0 due to receiving 0 for providing a non-decodable string. A zoomed-in version is included in the Appendix.
Figure 3. Decoding accuracy vs. similarity threshold for all models tested. The normalised Levenshtein similarity of each answer/target is compared to the similarity threshold. The accuracy is the proportion of sample scores which meet a given threshold. A zoomed-in version is included in the Appendix.
Figure 4. Encoding accuracy broken down by model and text type. Models ordered by overall encoding performance, and text type ordered by average difficulty across models.
Figure 4. Decoding accuracy broken down by model and text type. Models ordered by overall decoding performance, and text type ordered by average difficulty across models.
Bar graph comparing
Line graph comparing accuracy vs similarity threshold for multiple language models (0.95-1.0 range).
Graph showing

Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.

Comments 
In Channel
loading
00:00
00:00
x

0.5x

0.8x

1.0x

1.25x

1.5x

2.0x

3.0x

Sleep Timer

Off

End of Episode

5 Minutes

10 Minutes

15 Minutes

30 Minutes

45 Minutes

60 Minutes

120 Minutes

“Base64Bench: How good are LLMs at base64, and why care about it?” by richbc

“Base64Bench: How good are LLMs at base64, and why care about it?” by richbc