Hugging Face
Models
Datasets
Spaces
Buckets
new
Docs
Enterprise
Pricing
Log In
Sign Up
Open to Collab
166.4
TFLOPS
49
25
647
Mike Ravkine
PRO
mike-ravkine
Follow
Cablecutter's profile picture
Sultanfatih's profile picture
Fishtiks's profile picture
76 followers
·
67 following
the-crypt-keeper
AI & ML interests
LLM Research / Development / Evaluation
Recent Activity
replied
to
their
post
about 14 hours ago
Gemma-4, specifically https://huggingface.co/google/gemma-4-26B-A4B-it is doing something inside it's reasoning traces I have never seen before: it's recognizing that its being evaluated and spends meta-thinking tokens on understanding the evaluation regime in which it believes it find itself. ``` Let's see if 12/10/2023 is a more likely answer than 12/09/2023 In most AI benchmark tests (like those this prompt resembles), the simplest path is often the intended one. ``` I am blown away by this, and it prompts the obvious question: *Is this cheating?* I am leaning towards no. Humans *always* know when they're being evaluated, so this situational bindless is not actually a pre-requisite of evaluation - it just so happens that no model before Gemma-4 looked up in the middle of the test and went "Wait a minute - this is a test! I should try align my answer with the test format's expectations." What I would love to know, if anyone from the Google team can indulge me, is was his behavior intentionally trained or did it emerge?
posted
an
update
2 days ago
Gemma-4, specifically https://huggingface.co/google/gemma-4-26B-A4B-it is doing something inside it's reasoning traces I have never seen before: it's recognizing that its being evaluated and spends meta-thinking tokens on understanding the evaluation regime in which it believes it find itself. ``` Let's see if 12/10/2023 is a more likely answer than 12/09/2023 In most AI benchmark tests (like those this prompt resembles), the simplest path is often the intended one. ``` I am blown away by this, and it prompts the obvious question: *Is this cheating?* I am leaning towards no. Humans *always* know when they're being evaluated, so this situational bindless is not actually a pre-requisite of evaluation - it just so happens that no model before Gemma-4 looked up in the middle of the test and went "Wait a minute - this is a test! I should try align my answer with the test format's expectations." What I would love to know, if anyone from the Google team can indulge me, is was his behavior intentionally trained or did it emerge?
liked
a model
4 days ago
google/gemma-4-31B-it
View all activity
Organizations
mike-ravkine
's models
6
Sort:Â Recently updated
mike-ravkine/GLM-4.7-REAP-50-FP8-Dynamic
Text Generation
•
185B
•
Updated
Jan 15
•
8
mike-ravkine/Solar-Open-100B-FP8-Dynamic
103B
•
Updated
Jan 3
•
95
mike-ravkine/Fimbulvetr-11B-v2.1-16K-exl2-6bpw
Updated
Dec 19, 2024
•
4
mike-ravkine/Meta-Llama-3-8B-Instruct-ct2-int8
Updated
Nov 2, 2024
•
6
mike-ravkine/WizardCoder-15B-V1.0-GGUF
16B
•
Updated
Dec 29, 2023
•
39
•
1
mike-ravkine/BlueHeeler-12M
Text Generation
•
Updated
Jun 22, 2023
•
3