Hugging Face
Models
Datasets
Spaces
Buckets
new
Docs
Enterprise
Pricing
Log In
Sign Up
Open to Collab
5.2
TFLOPS
90
31
128
Kenneth Hamilton
PRO
ZennyKenny
Follow
ducla's profile picture
kiteatkit's profile picture
Latoofy's profile picture
131 followers
·
73 following
https://kennethhamilton.me
nevskycollectiv
kghamilton89
kenneth-gerald-hamilton
AI & ML interests
✌️ Certified vibe coder
Recent Activity
updated
a collection
12 days ago
Cool Models
replied
to
mike-ravkine
's
post
13 days ago
Gemma-4, specifically https://huggingface.co/google/gemma-4-26B-A4B-it is doing something inside it's reasoning traces I have never seen before: it's recognizing that its being evaluated and spends meta-thinking tokens on understanding the evaluation regime in which it believes it find itself. ``` Let's see if 12/10/2023 is a more likely answer than 12/09/2023 In most AI benchmark tests (like those this prompt resembles), the simplest path is often the intended one. ``` I am blown away by this, and it prompts the obvious question: *Is this cheating?* I am leaning towards no. Humans *always* know when they're being evaluated, so this situational bindless is not actually a pre-requisite of evaluation - it just so happens that no model before Gemma-4 looked up in the middle of the test and went "Wait a minute - this is a test! I should try align my answer with the test format's expectations." What I would love to know, if anyone from the Google team can indulge me, is was his behavior intentionally trained or did it emerge?
replied
to
mike-ravkine
's
post
14 days ago
Gemma-4, specifically https://huggingface.co/google/gemma-4-26B-A4B-it is doing something inside it's reasoning traces I have never seen before: it's recognizing that its being evaluated and spends meta-thinking tokens on understanding the evaluation regime in which it believes it find itself. ``` Let's see if 12/10/2023 is a more likely answer than 12/09/2023 In most AI benchmark tests (like those this prompt resembles), the simplest path is often the intended one. ``` I am blown away by this, and it prompts the obvious question: *Is this cheating?* I am leaning towards no. Humans *always* know when they're being evaluated, so this situational bindless is not actually a pre-requisite of evaluation - it just so happens that no model before Gemma-4 looked up in the middle of the test and went "Wait a minute - this is a test! I should try align my answer with the test format's expectations." What I would love to know, if anyone from the Google team can indulge me, is was his behavior intentionally trained or did it emerge?
View all activity
Organizations
ZennyKenny
's datasets
35
Sort: Recently updated
ZennyKenny/russian-dibt-llama-responses
Viewer
•
Updated
Jul 24, 2024
•
501
•
4
ZennyKenny/MPEP_RUSSIAN
Viewer
•
Updated
Jul 24, 2024
•
501
•
3
•
1
ZennyKenny/mpe_russian_jury_responses
Viewer
•
Updated
Jul 19, 2024
•
4
ZennyKenny/mpe_dutch_jury_responses
Viewer
•
Updated
Jul 19, 2024
•
4
ZennyKenny/mpe_greek_jury_responses
Viewer
•
Updated
Jul 19, 2024
•
4
Previous
1
2
Next