Gemma 2 2b Crosscoder L13 Mu4.1e 02 Lr1e 04
G
Gemma 2 2b Crosscoder L13 Mu4.1e 02 Lr1e 04
Developed by science-of-finetuning
Cross-encoder trained on parallel activations from layer 13 of Gemma 2 2B and Gemma 2 2B IT models
Downloads 51
Release Time : 11/22/2024
Model Overview
This cross-encoder was trained on subsets of fineweb and lsmsy-chat-1m datasets, primarily for feature extraction tasks.
Model Features
Parallel activation training
Trained on parallel activations from layer 13 of Gemma 2 2B and Gemma 2 2B IT models
Efficient feature extraction
Focuses on extracting meaningful feature representations from intermediate model layers
Sparse feature learning
Supports L1 and L0 sparsity metrics to generate sparse feature representations
Model Capabilities
Intermediate model layer feature extraction
Cross-model feature fusion
Sparse feature generation
Use Cases
Model analysis
Model internal representation research
Analyze differences in internal representations of different models under identical inputs
Quantifiable comparison of feature representation similarity across models
Feature engineering
Downstream task feature extraction
Extract intermediate layer features from pre-trained models for downstream tasks
Provides richer feature representations
Featured Recommended AI Models
Qwen2.5 VL 7B Abliterated Caption It I1 GGUF
Apache-2.0
Quantized version of Qwen2.5-VL-7B-Abliterated-Caption-it, supporting multilingual image description tasks.
Image-to-Text
Transformers Supports Multiple Languages

Q
mradermacher
167
1
Nunchaku Flux.1 Dev Colossus
Other
The Nunchaku quantized version of the Colossus Project Flux, designed to generate high-quality images based on text prompts. This model minimizes performance loss while optimizing inference efficiency.
Image Generation English
N
nunchaku-tech
235
3
Qwen2.5 VL 7B Abliterated Caption It GGUF
Apache-2.0
This is a static quantized version based on the Qwen2.5-VL-7B model, focusing on image captioning generation tasks and supporting multiple languages.
Image-to-Text
Transformers Supports Multiple Languages

Q
mradermacher
133
1
Olmocr 7B 0725 FP8
Apache-2.0
olmOCR-7B-0725-FP8 is a document OCR model based on the Qwen2.5-VL-7B-Instruct model. It is fine-tuned using the olmOCR-mix-0225 dataset and then quantized to the FP8 version.
Image-to-Text
Transformers English

O
allenai
881
3
Lucy 128k GGUF
Apache-2.0
Lucy-128k is a model developed based on Qwen3-1.7B, focusing on proxy-based web search and lightweight browsing, and can run efficiently on mobile devices.
Large Language Model
Transformers English

L
Mungert
263
2
Š 2025AIbase