diff --git a/tests/integration/model_bridge/test_deepseek_adapter.py b/tests/integration/model_bridge/test_deepseek_adapter.py new file mode 100644 index 000000000..9409299fd --- /dev/null +++ b/tests/integration/model_bridge/test_deepseek_adapter.py @@ -0,0 +1,125 @@ +"""Integration tests for DeepSeek V3 architecture adapter.""" + +import tempfile + +import pytest +import torch +from transformers import AutoTokenizer, DeepseekV3Config, DeepseekV3ForCausalLM + +from transformer_lens.model_bridge.bridge import TransformerBridge + + +@pytest.fixture(scope="module") +def tiny_deepseek_bridge(): + tiny_config = DeepseekV3Config( + hidden_size=256, + intermediate_size=512, + num_hidden_layers=4, + num_attention_heads=8, + q_lora_rank=64, + kv_lora_rank=32, + qk_nope_head_dim=16, + qk_rope_head_dim=8, + v_head_dim=16, + vocab_size=1000, + first_k_dense_replace=1, + n_routed_experts=8, + n_shared_experts=1, + num_experts_per_tok=2, + n_group=2, + topk_group=1, + max_position_embeddings=128, + moe_intermediate_size=256, + ) + hf_model = DeepseekV3ForCausalLM(tiny_config) + + with tempfile.TemporaryDirectory() as tmpdir: + hf_model.save_pretrained(tmpdir) + tok = AutoTokenizer.from_pretrained("gpt2") + tok.save_pretrained(tmpdir) + bridge = TransformerBridge.boot_transformers(tmpdir, device="cpu") + yield bridge + + +class TestDeepSeekBridgeCreation: + def test_bridge_has_correct_block_count(self, tiny_deepseek_bridge): + assert len(tiny_deepseek_bridge.blocks) == 4 + + def test_bridge_has_embed_and_unembed(self, tiny_deepseek_bridge): + assert hasattr(tiny_deepseek_bridge, "embed") + assert hasattr(tiny_deepseek_bridge, "unembed") + assert hasattr(tiny_deepseek_bridge, "ln_final") + + def test_attention_is_mla(self, tiny_deepseek_bridge): + from transformer_lens.model_bridge.generalized_components.mla_attention import ( + MLAAttentionBridge, + ) + + assert isinstance(tiny_deepseek_bridge.blocks[0].attn, MLAAttentionBridge) + + +class TestDeepSeekForwardPass: + def test_forward_returns_logits(self, tiny_deepseek_bridge): + tokens = torch.tensor([[1, 2, 3, 4]]) + with torch.no_grad(): + output = tiny_deepseek_bridge(tokens) + assert output.shape == (1, 4, 1000) + assert not torch.isnan(output).any() + assert not torch.isinf(output).any() + + def test_forward_matches_hf(self, tiny_deepseek_bridge): + """SDPA vs manual matmul — small float32 differences expected.""" + tokens = torch.tensor([[1, 2, 3, 4]]) + hf_model = tiny_deepseek_bridge.original_model + with torch.no_grad(): + bridge_out = tiny_deepseek_bridge(tokens) + hf_out = hf_model(tokens).logits + max_diff = (bridge_out - hf_out).abs().max().item() + assert max_diff < 0.15, f"Bridge vs HF max diff = {max_diff}" + + +class TestDeepSeekDenseVsMoELayers: + def test_dense_layer_has_no_moe_hooks(self, tiny_deepseek_bridge): + tokens = torch.tensor([[1, 2, 3, 4]]) + _, cache = tiny_deepseek_bridge.run_with_cache(tokens) + cache_keys = set(cache.keys()) + assert not any("blocks.0.mlp.gate" in k for k in cache_keys) + assert not any("blocks.0.mlp.shared_experts" in k for k in cache_keys) + + def test_moe_layer_has_gate_hooks(self, tiny_deepseek_bridge): + tokens = torch.tensor([[1, 2, 3, 4]]) + _, cache = tiny_deepseek_bridge.run_with_cache(tokens) + assert any("blocks.1.mlp.gate" in k for k in cache.keys()) + + def test_moe_layer_has_shared_experts_hooks(self, tiny_deepseek_bridge): + tokens = torch.tensor([[1, 2, 3, 4]]) + _, cache = tiny_deepseek_bridge.run_with_cache(tokens) + assert any("blocks.1.mlp.shared_experts" in k for k in cache.keys()) + + def test_both_layers_have_mlp_hooks(self, tiny_deepseek_bridge): + tokens = torch.tensor([[1, 2, 3, 4]]) + _, cache = tiny_deepseek_bridge.run_with_cache(tokens) + for i in [0, 1]: + assert f"blocks.{i}.mlp.hook_in" in cache + assert f"blocks.{i}.mlp.hook_out" in cache + + def test_both_layers_produce_non_nan(self, tiny_deepseek_bridge): + tokens = torch.tensor([[1, 2, 3, 4]]) + _, cache = tiny_deepseek_bridge.run_with_cache(tokens) + for i in [0, 1]: + assert not torch.isnan(cache[f"blocks.{i}.mlp.hook_out"]).any() + + +class TestDeepSeekAttentionHooks: + def test_attention_hooks_fire_all_layers(self, tiny_deepseek_bridge): + tokens = torch.tensor([[1, 2, 3, 4]]) + _, cache = tiny_deepseek_bridge.run_with_cache(tokens) + for i in range(4): + assert f"blocks.{i}.attn.hook_in" in cache + assert f"blocks.{i}.attn.hook_out" in cache + + def test_mla_latent_hooks_fire(self, tiny_deepseek_bridge): + tokens = torch.tensor([[1, 2, 3, 4]]) + _, cache = tiny_deepseek_bridge.run_with_cache(tokens) + assert any("hook_q_latent" in k for k in cache.keys()) + assert any("hook_kv_latent" in k for k in cache.keys()) diff --git a/tests/unit/model_bridge/generalized_components/test_mla_attention_bridge.py b/tests/unit/model_bridge/generalized_components/test_mla_attention_bridge.py new file mode 100644 index 000000000..d9824be0a --- /dev/null +++ b/tests/unit/model_bridge/generalized_components/test_mla_attention_bridge.py @@ -0,0 +1,185 @@ +"""Unit tests for MLAAttentionBridge (DeepSeek Multi-Head Latent Attention).""" + +import pytest +import torch +from transformers import DeepseekV3Config, DeepseekV3ForCausalLM + +from transformer_lens.model_bridge.generalized_components.mla_attention import ( + MLAAttentionBridge, +) + + +@pytest.fixture(scope="module") +def tiny_config(): + return DeepseekV3Config( + hidden_size=256, + intermediate_size=512, + num_hidden_layers=4, + num_attention_heads=8, + num_key_value_heads=1, + q_lora_rank=64, + kv_lora_rank=32, + qk_nope_head_dim=16, + qk_rope_head_dim=8, + v_head_dim=16, + vocab_size=1000, + first_k_dense_replace=1, + n_routed_experts=8, + n_shared_experts=1, + num_experts_per_tok=2, + n_group=2, + topk_group=1, + max_position_embeddings=128, + moe_intermediate_size=256, + ) + + +@pytest.fixture(scope="module") +def tiny_model(tiny_config): + return DeepseekV3ForCausalLM(tiny_config) + + +@pytest.fixture(scope="module") +def hf_attn(tiny_model): + return tiny_model.model.layers[0].self_attn + + +@pytest.fixture(scope="module") +def mla_bridge(tiny_config, hf_attn, tiny_model): + bridge = MLAAttentionBridge(name="self_attn", config=tiny_config, submodules={}) + bridge.set_original_component(hf_attn) + bridge.set_rotary_emb(tiny_model.model.rotary_emb) + return bridge + + +class TestMLAAttentionBridgeHooks: + def test_all_expected_hooks_exist(self, mla_bridge): + for hook_name in [ + "hook_in", + "hook_out", + "hook_q_latent", + "hook_kv_latent", + "hook_q", + "hook_k", + "hook_v", + "hook_rot_q", + "hook_rot_k", + "hook_attn_scores", + "hook_pattern", + "hook_cos", + "hook_sin", + ]: + assert hasattr(mla_bridge, hook_name), f"Missing hook: {hook_name}" + + def test_W_Q_raises_not_implemented(self, mla_bridge): + with pytest.raises(NotImplementedError, match="not available on MLA"): + _ = mla_bridge.W_Q + + def test_W_K_raises_not_implemented(self, mla_bridge): + with pytest.raises(NotImplementedError, match="not available on MLA"): + _ = mla_bridge.W_K + + +class TestMLAAttentionBridgeForward: + @pytest.fixture + def sample_inputs(self, tiny_config, tiny_model): + batch, seq = 2, 8 + hidden_states = torch.randn(batch, seq, tiny_config.hidden_size) + position_ids = torch.arange(seq).unsqueeze(0).expand(batch, -1) + cos, sin = tiny_model.model.rotary_emb(hidden_states, position_ids) + return hidden_states, (cos, sin) + + def test_output_matches_hf(self, mla_bridge, hf_attn, sample_inputs, tiny_model): + """HF uses SDPA, bridge uses manual matmul — small float32 differences expected.""" + hidden_states, position_embeddings = sample_inputs + + with torch.no_grad(): + hf_attn_out = hf_attn( + hidden_states, + position_embeddings=position_embeddings, + attention_mask=None, + )[0] + bridge_attn_out = mla_bridge( + hidden_states, + position_embeddings=position_embeddings, + attention_mask=None, + )[0] + + max_diff = (hf_attn_out - bridge_attn_out).abs().max().item() + mean_diff = (hf_attn_out - bridge_attn_out).abs().mean().item() + assert max_diff < 0.15, f"Output too different: max diff = {max_diff}" + assert mean_diff < 0.02, f"Output too different: mean diff = {mean_diff}" + + def test_hooks_fire_and_have_correct_shapes(self, mla_bridge, sample_inputs, tiny_config): + hidden_states, position_embeddings = sample_inputs + batch, seq = hidden_states.shape[:2] + captured = {} + + hooks_to_check = [ + "hook_q_latent", + "hook_kv_latent", + "hook_q", + "hook_k", + "hook_v", + "hook_rot_q", + "hook_rot_k", + "hook_attn_scores", + "hook_pattern", + ] + + handles = [] + for name in hooks_to_check: + + def make_capture(n): + def hook_fn(module, input, output): + captured[n] = output.shape + + return hook_fn + + handles.append(getattr(mla_bridge, name).register_forward_hook(make_capture(name))) + + try: + with torch.no_grad(): + mla_bridge( + hidden_states, position_embeddings=position_embeddings, attention_mask=None + ) + finally: + for h in handles: + h.remove() + + n_heads = tiny_config.num_attention_heads + qk_head_dim = tiny_config.qk_nope_head_dim + tiny_config.qk_rope_head_dim + + for name in hooks_to_check: + assert name in captured, f"Hook {name} did not fire" + + assert captured["hook_q_latent"] == (batch, seq, tiny_config.q_lora_rank) + assert captured["hook_kv_latent"] == (batch, seq, tiny_config.kv_lora_rank) + assert captured["hook_q"] == (batch, n_heads, seq, qk_head_dim) + assert captured["hook_k"] == (batch, n_heads, seq, qk_head_dim) + assert captured["hook_v"] == (batch, n_heads, seq, tiny_config.v_head_dim) + assert captured["hook_attn_scores"] == (batch, n_heads, seq, seq) + assert captured["hook_pattern"] == (batch, n_heads, seq, seq) + + def test_hook_q_is_post_rope(self, mla_bridge, sample_inputs): + """hook_q's rope portion should match hook_rot_q.""" + hidden_states, position_embeddings = sample_inputs + q_values: list[torch.Tensor] = [] + rot_q_values: list[torch.Tensor] = [] + + h1 = mla_bridge.hook_q.register_forward_hook(lambda m, i, o: q_values.append(o.clone())) + h2 = mla_bridge.hook_rot_q.register_forward_hook( + lambda m, i, o: rot_q_values.append(o.clone()) + ) + + try: + with torch.no_grad(): + mla_bridge( + hidden_states, position_embeddings=position_embeddings, attention_mask=None + ) + finally: + h1.remove() + h2.remove() + + qk_rope_dim = mla_bridge._qk_rope_head_dim + assert torch.allclose(q_values[0][..., -qk_rope_dim:], rot_q_values[0], atol=1e-5) diff --git a/transformer_lens/benchmarks/component_outputs.py b/transformer_lens/benchmarks/component_outputs.py index 504825ce1..ef3751ead 100644 --- a/transformer_lens/benchmarks/component_outputs.py +++ b/transformer_lens/benchmarks/component_outputs.py @@ -442,6 +442,17 @@ def _test_component_recursive( if last_part in ["o", "out"]: return + # Skip MLA intermediates (expect compressed-dim inputs, not hidden_states) + if last_part in [ + "q_a_proj", + "q_a_layernorm", + "q_b_proj", + "kv_a_proj_with_mqa", + "kv_a_layernorm", + "kv_b_proj", + ]: + return + # Skip virtual splits from fused projections (no standalone HF equivalent) if last_part in ["q", "k", "v", "gate", "in"]: parent_path = ".".join(path_parts[:-1]) @@ -462,6 +473,29 @@ def _test_component_recursive( except Exception: pass + # Skip components not wired on this layer (per-layer or per-config variation). + # Only report as failure if the HF model has it but the bridge doesn't. + try: + self.adapter.get_component(self.bridge_model, component_path) + except (AttributeError, ValueError): + parts = component_path.split(".") + if len(parts) >= 3 and parts[1].isdigit(): + subpath = ".".join([parts[0]] + ["{layer}"] + parts[2:]) + # Per-layer variation: exists on some other layer (e.g., MoE vs dense) + for probe_layer in range(self.cfg.n_layers): + probe_path = subpath.replace("{layer}", str(probe_layer)) + try: + self.adapter.get_component(self.bridge_model, probe_path) + return # Found on another layer — skip this one + except (AttributeError, ValueError): + continue + # Per-config absence: HF model also lacks it (e.g., q_lora_rank=None) + try: + self.adapter.get_component(self.hf_model, component_path) + except (AttributeError, ValueError): + return + # Bridge is missing a component that HF has — likely misconfiguration + # Test this component result = self._test_component(component_path, component, test_inputs) if result is not None: diff --git a/transformer_lens/factories/architecture_adapter_factory.py b/transformer_lens/factories/architecture_adapter_factory.py index 7cd4892bc..a55f51a5a 100644 --- a/transformer_lens/factories/architecture_adapter_factory.py +++ b/transformer_lens/factories/architecture_adapter_factory.py @@ -11,6 +11,7 @@ BloomArchitectureAdapter, CodeGenArchitectureAdapter, CohereArchitectureAdapter, + DeepSeekV3ArchitectureAdapter, FalconArchitectureAdapter, Gemma1ArchitectureAdapter, Gemma2ArchitectureAdapter, @@ -66,6 +67,7 @@ "BloomForCausalLM": BloomArchitectureAdapter, "CodeGenForCausalLM": CodeGenArchitectureAdapter, "CohereForCausalLM": CohereArchitectureAdapter, + "DeepseekV3ForCausalLM": DeepSeekV3ArchitectureAdapter, "FalconForCausalLM": FalconArchitectureAdapter, "GemmaForCausalLM": Gemma1ArchitectureAdapter, # Default to Gemma1 as it's the original version "Gemma1ForCausalLM": Gemma1ArchitectureAdapter, diff --git a/transformer_lens/model_bridge/generalized_components/__init__.py b/transformer_lens/model_bridge/generalized_components/__init__.py index a84ecebbf..61356ff3b 100644 --- a/transformer_lens/model_bridge/generalized_components/__init__.py +++ b/transformer_lens/model_bridge/generalized_components/__init__.py @@ -7,6 +7,7 @@ ) from transformer_lens.model_bridge.generalized_components.block import ( BlockBridge, + MLABlockBridge, ParallelBlockBridge, ) from transformer_lens.model_bridge.generalized_components.bloom_attention import ( @@ -57,6 +58,9 @@ JointQKVPositionEmbeddingsAttentionBridge, ) from transformer_lens.model_bridge.generalized_components.linear import LinearBridge +from transformer_lens.model_bridge.generalized_components.mla_attention import ( + MLAAttentionBridge, +) from transformer_lens.model_bridge.generalized_components.mlp import MLPBridge from transformer_lens.model_bridge.generalized_components.mpt_alibi_attention import ( MPTALiBiAttentionBridge, @@ -97,6 +101,7 @@ "AttentionBridge", "AudioFeatureExtractorBridge", "BlockBridge", + "MLABlockBridge", "ParallelBlockBridge", "BloomBlockBridge", "BloomAttentionBridge", @@ -120,6 +125,7 @@ "MLPBridge", "MPTALiBiAttentionBridge", "GatedMLPBridge", + "MLAAttentionBridge", "GatedRMSNormBridge", "MoEBridge", "PositionEmbeddingsAttentionBridge", diff --git a/transformer_lens/model_bridge/generalized_components/block.py b/transformer_lens/model_bridge/generalized_components/block.py index 73f2bd130..c4cb71b71 100644 --- a/transformer_lens/model_bridge/generalized_components/block.py +++ b/transformer_lens/model_bridge/generalized_components/block.py @@ -240,6 +240,36 @@ def _filter_kwargs_for_forward( return kwargs +class MLABlockBridge(BlockBridge): + """Block wrapping Multi-Head Latent Attention (DeepSeek V2/V3/R1). + + MLA has no standalone q/k/v projections — Q flows through compressed + q_a_proj→q_a_layernorm→q_b_proj, and K/V share a joint kv_a_proj_with_mqa + entry point. There is no single HookPoint that represents "input that + becomes Q/K/V", so the block-level ``hook_q_input``/``hook_k_input``/ + ``hook_v_input`` aliases do not apply. Type-level distinction means a reader + of the adapter sees ``MLABlockBridge`` and knows those hooks are absent. + """ + + def __init__( + self, + name: str, + config: Optional[Any] = None, + submodules: Optional[Dict[str, GeneralizedComponent]] = None, + hook_alias_overrides: Optional[Dict[str, str]] = None, + ): + super().__init__( + name, + config=config, + submodules=submodules, + hook_alias_overrides=hook_alias_overrides, + ) + if self.hook_aliases is BlockBridge.hook_aliases: + self.hook_aliases = dict(self.hook_aliases) + for alias in ("hook_q_input", "hook_k_input", "hook_v_input"): + self.hook_aliases.pop(alias, None) + + class ParallelBlockBridge(BlockBridge): """Block where attn and MLP both read the pre-attention residual. diff --git a/transformer_lens/model_bridge/generalized_components/gated_mlp.py b/transformer_lens/model_bridge/generalized_components/gated_mlp.py index 80e1a2d54..a908dbc68 100644 --- a/transformer_lens/model_bridge/generalized_components/gated_mlp.py +++ b/transformer_lens/model_bridge/generalized_components/gated_mlp.py @@ -67,6 +67,7 @@ def __init__( name: Optional[str], config: Optional[Any] = None, submodules: Optional[Dict[str, GeneralizedComponent]] = None, + optional: bool = False, ): """Initialize the gated MLP bridge. @@ -74,8 +75,9 @@ def __init__( name: The name of the component in the model (None if no container exists) config: Optional configuration (unused for GatedMLPBridge) submodules: Dictionary of submodules to register (e.g., gate_proj, up_proj, down_proj) + optional: If True, setup skips this bridge when absent (hybrid architectures). """ - super().__init__(name, config, submodules=submodules or {}) + super().__init__(name, config, submodules=submodules or {}, optional=optional) def forward(self, *args, **kwargs) -> torch.Tensor: """Forward pass through the gated MLP bridge. diff --git a/transformer_lens/model_bridge/generalized_components/mla_attention.py b/transformer_lens/model_bridge/generalized_components/mla_attention.py new file mode 100644 index 000000000..c394f85a3 --- /dev/null +++ b/transformer_lens/model_bridge/generalized_components/mla_attention.py @@ -0,0 +1,294 @@ +"""Multi-Head Latent Attention (MLA) bridge component for DeepSeek models. + +MLA compresses Q and KV into lower-dimensional latent spaces via LoRA-style +projections before standard attention. This component reimplements the MLA +forward path step-by-step with hooks at each meaningful stage, exposing: + +- hook_q_latent / hook_kv_latent: compressed representations (the information bottleneck) +- hook_q / hook_k / hook_v: final Q/K/V entering attention (post-decompression, post-RoPE) +- hook_rot_q / hook_rot_k: after RoPE on the rope portion splits +- hook_attn_scores / hook_pattern: pre/post-softmax attention weights +- hook_z: pre-output-projection (alias for o.hook_in) +""" + +from __future__ import annotations + +from typing import Any, Dict, Optional + +import torch + +from transformer_lens.hook_points import HookPoint +from transformer_lens.model_bridge.generalized_components.attention import ( + AttentionBridge, +) +from transformer_lens.model_bridge.generalized_components.base import ( + GeneralizedComponent, +) +from transformer_lens.model_bridge.generalized_components.position_embedding_hooks_mixin import ( + PositionEmbeddingHooksMixin, +) + + +def _rotate_half(x: torch.Tensor) -> torch.Tensor: + """Rotate half of the hidden dims of the input (standard RoPE helper).""" + x1 = x[..., : x.shape[-1] // 2] + x2 = x[..., x.shape[-1] // 2 :] + return torch.cat((-x2, x1), dim=-1) + + +def _apply_rotary_pos_emb( + q: torch.Tensor, k: torch.Tensor, cos: torch.Tensor, sin: torch.Tensor +) -> tuple[torch.Tensor, torch.Tensor]: + """Apply rotary position embedding to q and k tensors.""" + cos = cos.unsqueeze(1) # [batch, 1, seq, dim] + sin = sin.unsqueeze(1) + q_embed = (q * cos) + (_rotate_half(q) * sin) + k_embed = (k * cos) + (_rotate_half(k) * sin) + return q_embed, k_embed + + +class MLAAttentionBridge(PositionEmbeddingHooksMixin, AttentionBridge): + """Bridge for DeepSeek's Multi-Head Latent Attention (MLA). + + Reimplements the MLA forward path with hooks at each computation stage. + Standard W_Q/W_K/W_V properties are not available on MLA models — use + the submodule weight access (q_a_proj, q_b_proj, etc.) instead. + """ + + # MLA has no standard q/k/v submodules — override to empty + property_aliases: Dict[str, str] = {} + + hook_aliases = { + "hook_result": "hook_out", + "hook_z": "o.hook_in", + } + + def __init__( + self, + name: str, + config: Any, + submodules: Optional[Dict[str, GeneralizedComponent]] = None, + **kwargs: Any, + ): + super().__init__(name, config, submodules=submodules, **kwargs) + self._init_position_embedding_hooks() + + self.hook_q_latent = HookPoint() # Compressed Q (post q_a_layernorm) + self.hook_kv_latent = HookPoint() # Compressed KV (post kv_a_layernorm) + self.hook_q = HookPoint() # Final Q entering attention (post-RoPE concat) + self.hook_k = HookPoint() # Final K entering attention (post-RoPE concat) + self.hook_v = HookPoint() # V from kv_b_proj split + self.hook_rot_q = HookPoint() # Q rope portion after RoPE + self.hook_rot_k = HookPoint() # K rope portion after RoPE + + # MLA params lazy-initialized from HF module (bridge config lacks these fields) + self._mla_params_initialized = False + + def forward(self, *args: Any, **kwargs: Any) -> Any: + """Reimplemented MLA forward with hooks at each computation stage. + + Follows the DeepseekV3Attention forward path, calling into HF submodules + individually and firing hooks at each meaningful stage. + """ + if self.original_component is None: + raise RuntimeError( + f"Original component not set for {self.name}. " + "Call set_original_component() first." + ) + + hf_attn: Any = self.original_component + + if not self._mla_params_initialized: + self._q_lora_rank = getattr(hf_attn, "q_lora_rank", None) + self._kv_lora_rank = getattr(hf_attn, "kv_lora_rank", 512) + self._qk_nope_head_dim = getattr(hf_attn, "qk_nope_head_dim", 128) + self._qk_rope_head_dim = getattr(hf_attn, "qk_rope_head_dim", 64) + self._v_head_dim = getattr(hf_attn, "v_head_dim", 128) + self._qk_head_dim = self._qk_nope_head_dim + self._qk_rope_head_dim + self._n_heads = getattr(hf_attn, "num_heads", 32) + hf_config = getattr(hf_attn, "config", None) + self._rope_interleave = ( + getattr(hf_config, "rope_interleave", False) if hf_config else False + ) + self._mla_params_initialized = True + + # --- Extract inputs --- + if "hidden_states" in kwargs: + hidden_states = kwargs.pop("hidden_states") + elif len(args) > 0 and isinstance(args[0], torch.Tensor): + hidden_states = args[0] + args = args[1:] + else: + raise ValueError("Could not find hidden_states in args or kwargs") + + position_embeddings = kwargs.pop("position_embeddings", None) + attention_mask = kwargs.pop("attention_mask", None) + + hidden_states = self.hook_in(hidden_states) + + batch_size, seq_length = hidden_states.shape[:2] + + # --- Query path --- + if self._q_lora_rank is None: + # Direct projection (no compression) + q_states = hf_attn.q_proj(hidden_states) + else: + # Two-stage compression: q_a_proj → q_a_layernorm → q_b_proj + q_compressed = hf_attn.q_a_proj(hidden_states) + q_compressed = hf_attn.q_a_layernorm(q_compressed) + q_compressed = self.hook_q_latent(q_compressed) + q_states = hf_attn.q_b_proj(q_compressed) + + # Reshape to [batch, n_heads, seq, qk_head_dim] + q_states = q_states.view(batch_size, seq_length, -1, self._qk_head_dim).transpose(1, 2) + # Split into nope (non-RoPE) and pe (RoPE) portions + q_pass, q_rot = torch.split( + q_states, [self._qk_nope_head_dim, self._qk_rope_head_dim], dim=-1 + ) + + # --- KV path --- + # kv_a_proj_with_mqa outputs [compressed_kv || k_pe] + compressed_kv_full = hf_attn.kv_a_proj_with_mqa(hidden_states) + # Split: compressed KV latent (for kv_b_proj) and k rope portion (for direct RoPE) + # Note: k_pe is split off here and goes directly to RoPE — hook_kv_latent + # captures only the compressed_kv portion that enters the decompression path. + k_pass, k_rot = torch.split( + compressed_kv_full, [self._kv_lora_rank, self._qk_rope_head_dim], dim=-1 + ) + + # Compress → normalize → decompress the KV latent + k_pass = hf_attn.kv_a_layernorm(k_pass) + k_pass = self.hook_kv_latent(k_pass) + k_pass = hf_attn.kv_b_proj(k_pass) + + # Reshape to [batch, n_heads, seq, nope+v_head] + key_shape = (batch_size, seq_length, -1, self._qk_nope_head_dim + self._v_head_dim) + k_pass = k_pass.view(key_shape).transpose(1, 2) + # Split K nope portion and V + k_pass, value_states = torch.split( + k_pass, [self._qk_nope_head_dim, self._v_head_dim], dim=-1 + ) + + # k_rot is [batch, seq, rope_dim] → [batch, 1, seq, rope_dim] for broadcasting + k_rot = k_rot.view(batch_size, 1, seq_length, self._qk_rope_head_dim) + + # --- RoPE --- + if position_embeddings is not None: + position_embeddings = self._apply_position_embedding_hooks(position_embeddings) + cos, sin = position_embeddings + elif self._rotary_emb is not None: + # Fallback: compute from rotary_emb if position_embeddings not passed + position_ids = torch.arange(seq_length, device=hidden_states.device).unsqueeze(0) + cos, sin = self._rotary_emb(hidden_states, position_ids) + else: + raise ValueError( + "MLAAttentionBridge requires position_embeddings or set_rotary_emb() " + "to be called before forward." + ) + + q_rot, k_rot = _apply_rotary_pos_emb(q_rot, k_rot, cos, sin) + q_rot = self.hook_rot_q(q_rot) + k_rot = self.hook_rot_k(k_rot) + + # Expand k_rot to match the number of heads + k_rot = k_rot.expand(*k_pass.shape[:-1], -1) + + # Concatenate nope + rope portions to form final Q and K + query_states = torch.cat((q_pass, q_rot), dim=-1) + key_states = torch.cat((k_pass, k_rot), dim=-1) + + # Fire final Q/K/V hooks — these are the tensors entering attention + query_states = self.hook_q(query_states) + key_states = self.hook_k(key_states) + value_states = self.hook_v(value_states) + + # --- KV Cache --- + past_key_values = kwargs.pop("past_key_values", None) + cache_position = kwargs.pop("cache_position", None) + if past_key_values is not None: + cache_kwargs = {"sin": sin, "cos": cos, "cache_position": cache_position} + key_states, value_states = past_key_values.update( + key_states, value_states, hf_attn.layer_idx, cache_kwargs + ) + + # --- Attention computation (no V padding — only needed for flash attention) --- + scaling = self._qk_head_dim ** (-0.5) + attn_scores = torch.matmul(query_states, key_states.transpose(-2, -1)) * scaling + + if attention_mask is not None: + attn_scores = attn_scores + attention_mask + + attn_scores = self.hook_attn_scores(attn_scores) + attn_weights = self._softmax_dropout_pattern( + attn_scores, upcast_to_fp32=True, target_dtype=query_states.dtype + ) + + # Weighted sum of values + attn_output = torch.matmul(attn_weights, value_states) + + # --- Output projection --- + attn_output = attn_output.transpose(1, 2).contiguous() + attn_output = attn_output.reshape(batch_size, seq_length, -1) + attn_output = hf_attn.o_proj(attn_output) + + attn_output = self.hook_out(attn_output) + return attn_output, attn_weights + + def get_random_inputs( + self, + batch_size: int = 2, + seq_len: int = 8, + device: Optional[torch.device] = None, + dtype: Optional[torch.dtype] = None, + ) -> Dict[str, Any]: + """Generate test inputs with hidden_states, position_embeddings, and attention_mask.""" + if device is None: + device = torch.device("cpu") + if dtype is None: + dtype = torch.float32 + + # Try bridge config (d_model), then HF attention's config (hidden_size), then fallback + d_model = None + if self.config and hasattr(self.config, "d_model"): + d_model = self.config.d_model + if d_model is None and self.original_component is not None: + hf_cfg = getattr(self.original_component, "config", None) + if hf_cfg is not None: + d_model = getattr(hf_cfg, "hidden_size", None) + if d_model is None: + d_model = 256 + inputs: Dict[str, Any] = { + "hidden_states": torch.randn(batch_size, seq_len, d_model, device=device, dtype=dtype) + } + + # Generate position_embeddings from rotary_emb if available, + # otherwise create dummy (cos=1, sin=0) embeddings + rope_head_dim = self._qk_rope_head_dim if self._mla_params_initialized else 64 + if self._rotary_emb is not None: + try: + dummy_input = inputs["hidden_states"] + position_ids = torch.arange(seq_len, device=device).unsqueeze(0) + position_embeddings = self._rotary_emb(dummy_input, position_ids) + inputs["position_embeddings"] = position_embeddings + except Exception: + cos = torch.ones(1, seq_len, rope_head_dim, device=device, dtype=dtype) + sin = torch.zeros(1, seq_len, rope_head_dim, device=device, dtype=dtype) + inputs["position_embeddings"] = (cos, sin) + else: + cos = torch.ones(1, seq_len, rope_head_dim, device=device, dtype=dtype) + sin = torch.zeros(1, seq_len, rope_head_dim, device=device, dtype=dtype) + inputs["position_embeddings"] = (cos, sin) + + inputs["attention_mask"] = None + return inputs + + def __getattr__(self, name: str) -> Any: + """Raise clear error for standard weight properties that don't apply to MLA.""" + if name in ("W_Q", "W_K", "W_V", "W_O", "b_Q", "b_K", "b_V", "b_O"): + raise NotImplementedError( + f"{name} is not available on MLA (Multi-Head Latent Attention) models. " + f"MLA uses compressed projections instead of standard Q/K/V. " + f"Access weights via submodules: q_a_proj, q_b_proj, kv_a_proj_with_mqa, " + f"kv_b_proj, o (o_proj)." + ) + return super().__getattr__(name) diff --git a/transformer_lens/model_bridge/generalized_components/mlp.py b/transformer_lens/model_bridge/generalized_components/mlp.py index e6e70991b..86ac19fa0 100644 --- a/transformer_lens/model_bridge/generalized_components/mlp.py +++ b/transformer_lens/model_bridge/generalized_components/mlp.py @@ -33,6 +33,7 @@ def __init__( name: Optional[str], config: Optional[Any] = None, submodules: Optional[Dict[str, GeneralizedComponent]] = {}, + optional: bool = False, ): """Initialize the MLP bridge. @@ -40,8 +41,9 @@ def __init__( name: The name of the component in the model (None if no container exists) config: Optional configuration (unused for MLPBridge) submodules: Dictionary of submodules to register (e.g., gate_proj, up_proj, down_proj) + optional: If True, setup skips this bridge when absent (hybrid architectures). """ - super().__init__(name, config, submodules=submodules) + super().__init__(name, config, submodules=submodules, optional=optional) def forward(self, *args, **kwargs) -> torch.Tensor: """Forward pass through the MLP bridge. diff --git a/transformer_lens/model_bridge/supported_architectures/__init__.py b/transformer_lens/model_bridge/supported_architectures/__init__.py index 8fb69337c..7f990e393 100644 --- a/transformer_lens/model_bridge/supported_architectures/__init__.py +++ b/transformer_lens/model_bridge/supported_architectures/__init__.py @@ -18,6 +18,9 @@ from transformer_lens.model_bridge.supported_architectures.cohere import ( CohereArchitectureAdapter, ) +from transformer_lens.model_bridge.supported_architectures.deepseek_v3 import ( + DeepSeekV3ArchitectureAdapter, +) from transformer_lens.model_bridge.supported_architectures.falcon import ( FalconArchitectureAdapter, ) @@ -166,6 +169,7 @@ "BloomArchitectureAdapter", "CodeGenArchitectureAdapter", "CohereArchitectureAdapter", + "DeepSeekV3ArchitectureAdapter", "FalconArchitectureAdapter", "Gemma1ArchitectureAdapter", "Gemma2ArchitectureAdapter", diff --git a/transformer_lens/model_bridge/supported_architectures/deepseek_v3.py b/transformer_lens/model_bridge/supported_architectures/deepseek_v3.py new file mode 100644 index 000000000..6d7b80215 --- /dev/null +++ b/transformer_lens/model_bridge/supported_architectures/deepseek_v3.py @@ -0,0 +1,112 @@ +"""DeepSeek V3 architecture adapter. + +Supports DeepSeek V3 and DeepSeek-R1 models (both use DeepseekV3ForCausalLM). +Key features: +- Multi-Head Latent Attention (MLA): Q and KV compressed via LoRA-style projections +- Mixture of Experts (MoE) with shared experts on most layers +- Dense MLP on first `first_k_dense_replace` layers +""" + +from typing import Any + +from transformer_lens.model_bridge.architecture_adapter import ArchitectureAdapter +from transformer_lens.model_bridge.generalized_components import ( + EmbeddingBridge, + GatedMLPBridge, + LinearBridge, + MLAAttentionBridge, + MLABlockBridge, + MoEBridge, + RMSNormalizationBridge, + RotaryEmbeddingBridge, + UnembeddingBridge, +) +from transformer_lens.model_bridge.generalized_components.base import ( + GeneralizedComponent, +) + + +class DeepSeekV3ArchitectureAdapter(ArchitectureAdapter): + """Architecture adapter for DeepSeek V3 / R1 models. + + Uses RMSNorm, MLA with compressed Q/KV projections, partial RoPE, + MoE on most layers (dense MLP on first few), and no biases. + """ + + def __init__(self, cfg: Any) -> None: + super().__init__(cfg) + + self.cfg.normalization_type = "RMS" + self.cfg.positional_embedding_type = "rotary" + self.cfg.gated_mlp = True + self.cfg.final_rms = True + self.cfg.uses_rms_norm = True + # HF defaults to SDPA which handles MLA correctly. + # HF's eager attention crashes on MLA's asymmetric Q/K dimensions. + + self.weight_processing_conversions = {} + + self.component_mapping = { + "embed": EmbeddingBridge(name="model.embed_tokens"), + "rotary_emb": RotaryEmbeddingBridge(name="model.rotary_emb", config=self.cfg), + "blocks": MLABlockBridge( + name="model.layers", + submodules={ + "ln1": RMSNormalizationBridge(name="input_layernorm", config=self.cfg), + "ln2": RMSNormalizationBridge(name="post_attention_layernorm", config=self.cfg), + "attn": MLAAttentionBridge( + name="self_attn", + config=self.cfg, + submodules={ + "q_a_proj": LinearBridge(name="q_a_proj"), + "q_a_layernorm": RMSNormalizationBridge( + name="q_a_layernorm", config=self.cfg + ), + "q_b_proj": LinearBridge(name="q_b_proj"), + "kv_a_proj_with_mqa": LinearBridge(name="kv_a_proj_with_mqa"), + "kv_a_layernorm": RMSNormalizationBridge( + name="kv_a_layernorm", config=self.cfg + ), + "kv_b_proj": LinearBridge(name="kv_b_proj"), + "o": LinearBridge(name="o_proj"), + }, + ), + # On dense layers (idx < first_k_dense_replace), gate and + # shared_experts are marked optional so setup gracefully + # skips them when the layer is DeepseekV3MLP instead of MoE. + "mlp": MoEBridge( + name="mlp", + config=self.cfg, + submodules={ + # Router is a custom Module, not nn.Linear + "gate": GeneralizedComponent(name="gate", optional=True), + "shared_experts": GatedMLPBridge( + name="shared_experts", + config=self.cfg, + optional=True, + submodules={ + "gate": LinearBridge(name="gate_proj"), + "in": LinearBridge(name="up_proj"), + "out": LinearBridge(name="down_proj"), + }, + ), + }, + ), + }, + ), + "ln_final": RMSNormalizationBridge(name="model.norm", config=self.cfg), + "unembed": UnembeddingBridge(name="lm_head"), + } + + def setup_component_testing(self, hf_model: Any, bridge_model: Any = None) -> None: + """Set up rotary embedding references for component testing.""" + rotary_emb = hf_model.model.rotary_emb + + if bridge_model is not None and hasattr(bridge_model, "blocks"): + for block in bridge_model.blocks: + if hasattr(block, "attn"): + block.attn.set_rotary_emb(rotary_emb) + + # Also set on template for get_generalized_component() callers + attn_bridge = self.get_generalized_component("blocks.0.attn") + attn_bridge.set_rotary_emb(rotary_emb) diff --git a/transformer_lens/tools/model_registry/__init__.py b/transformer_lens/tools/model_registry/__init__.py index 3150f159f..bf956682b 100644 --- a/transformer_lens/tools/model_registry/__init__.py +++ b/transformer_lens/tools/model_registry/__init__.py @@ -46,6 +46,7 @@ "BertForMaskedLM", "BloomForCausalLM", "CodeGenForCausalLM", + "DeepseekV3ForCausalLM", "FalconForCausalLM", "GemmaForCausalLM", "Gemma2ForCausalLM", diff --git a/transformer_lens/tools/model_registry/data/architecture_gaps.json b/transformer_lens/tools/model_registry/data/architecture_gaps.json index 6261a9a65..998731720 100644 --- a/transformer_lens/tools/model_registry/data/architecture_gaps.json +++ b/transformer_lens/tools/model_registry/data/architecture_gaps.json @@ -47,25 +47,6 @@ ], "relevancy_score": 84.9 }, - { - "architecture_id": "DeepseekV3ForCausalLM", - "total_models": 46, - "total_downloads": 6840308, - "min_param_count": 1656048, - "sample_models": [ - "deepseek-ai/DeepSeek-R1", - "deepseek-ai/DeepSeek-R1-0528", - "deepseek-ai/DeepSeek-V3", - "deepseek-ai/DeepSeek-V3-0324", - "moonshotai/Kimi-K2-Instruct-0905", - "moonshotai/Kimi-K2-Instruct", - "deepseek-ai/DeepSeek-V3.1", - "trl-internal-testing/tiny-DeepseekV3ForCausalLM", - "trl-internal-testing/tiny-DeepseekV3ForCausalLM-0528", - "moonshotai/Moonlight-16B-A3B-Instruct" - ], - "relevancy_score": 83.1 - }, { "architecture_id": "NemotronHForCausalLM", "total_models": 41, @@ -185,6 +166,22 @@ "total_models": 14, "total_downloads": 1062491, "min_param_count": 312517632, + "sample_models": [ + "openbmb/MiniCPM-2B-sft-bf16", + "openbmb/MiniCPM4.1-8B", + "openbmb/MiniCPM-1B-sft-bf16", + "openbmb/MiniCPM4-0.5B", + "openbmb/MiniCPM-MoE-8x2B", + "katuni4ka/tiny-random-minicpm", + "openbmb/MiniCPM-S-1B-sft", + "openbmb/MiniCPM4-8B", + "openbmb/MiniCPM-2B-dpo-bf16", + "openbmb/MiniCPM-2B-sft-fp32" + ] + }, + { + "architecture_id": "T5GemmaForConditionalGeneration", + "total_models": 9, "sample_models": [ "google/t5gemma-s-s-prefixlm", "google/t5gemma-b-b-ul2", @@ -2113,7 +2110,7 @@ "relevancy_score": 32.9 }, { - "architecture_id": "XCurOSForCausalLM", + "architecture_id": "Step3p5ForCausalLM", "total_models": 1, "total_downloads": 66986, "min_param_count": 7615616512, @@ -2293,7 +2290,7 @@ "relevancy_score": 30.8 }, { - "architecture_id": "HymbaForCausalLM", + "architecture_id": "ShikraLlamaForCausalLM", "total_models": 1, "total_downloads": 645, "min_param_count": 1522797824, diff --git a/transformer_lens/tools/model_registry/data/supported_models.json b/transformer_lens/tools/model_registry/data/supported_models.json index 4a4bad75a..46a8f2aee 100644 --- a/transformer_lens/tools/model_registry/data/supported_models.json +++ b/transformer_lens/tools/model_registry/data/supported_models.json @@ -6,9 +6,9 @@ "min_downloads": 500, "scan_duration_seconds": 4.2 }, - "total_architectures": 47, - "total_models": 7426, - "total_verified": 706, + "total_architectures": 48, + "total_models": 8941, + "total_verified": 708, "models": [ { "architecture_id": "Qwen3NextForCausalLM", @@ -104530,6 +104530,19061 @@ "phase4_score": null, "phase7_score": null, "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "RedHatAI/Llama-3.2-1B-Instruct-FP8-dynamic", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Qwen/Qwen3-0.6B-FP8", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-32B-Instruct-AWQ", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 452.0 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "RedHatAI/Qwen2.5-1.5B-quantized.w8a8", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-14B-Instruct-AWQ", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 208.7 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-72B-Instruct-AWQ", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 1057.5 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "kaitchup/Phi-3-mini-4k-instruct-gptq-4bit", + "status": 2, + "verified_date": "2026-03-30", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-Coder-7B-Instruct-AWQ", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 100.2 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-Coder-32B-Instruct-AWQ", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 452.0 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-1.5B-Instruct-AWQ", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "mlx-community/gpt-oss-20b-MXFP4-Q8", + "status": 2, + "verified_date": "2026-02-22", + "metadata": null, + "note": "Estimated 267.1 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 108.3 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "RedHatAI/Llama-3.2-1B-Instruct-FP8", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-Coder-7B-Instruct-GPTQ-Int4", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 100.2 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "kosbu/Llama-3.3-70B-Instruct-AWQ", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 1028.7 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Qwen/Qwen3-14B-AWQ", + "status": 2, + "verified_date": "2026-02-22", + "metadata": null, + "note": "Estimated 204.3 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "hugging-quants/Meta-Llama-3.1-8B-Instruct-AWQ-INT4", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 108.3 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Qwen/Qwen3-4B-Instruct-2507-FP8", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Qwen/Qwen3-32B-FP8", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 478.3 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "unsloth/mistral-7b-v0.3-bnb-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Qwen/Qwen3-32B-AWQ", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 478.3 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "casperhansen/llama-3.3-70b-instruct-awq", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Qwen/Qwen3-8B-FP8", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 110.5 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-7B-Instruct-AWQ", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 100.2 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "lmstudio-community/DeepSeek-R1-0528-Qwen3-8B-MLX-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 110.5 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "lmstudio-community/DeepSeek-R1-0528-Qwen3-8B-MLX-8bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 110.5 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/TinyLlama-1.1B-Chat-v0.3-GPTQ", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-4B-Instruct-2507-unsloth-bnb-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Meta-Llama-3.1-8B-Instruct-bnb-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "unsloth/mistral-7b-bnb-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/QwQ-32B-AWQ", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 452.0 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Qwen/Qwen3-4B-FP8", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "knifeayumu/Cydonia-v1.3-Magnum-v4-22B", + "status": 2, + "verified_date": "2026-02-22", + "metadata": null, + "note": "Estimated 333.4 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen1.5-110B-Chat-AWQ", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 1556.2 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/Nous-Hermes-2-Yi-34B-AWQ", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "unsloth/gpt-oss-20b-unsloth-bnb-4bit", + "status": 2, + "verified_date": "2026-02-22", + "metadata": null, + "note": "Estimated 267.1 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-Coder-1.5B-Instruct-AWQ", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Qwen/Qwen3-4B-AWQ", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "RedHatAI/Meta-Llama-3.1-8B-FP8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "unsloth/gpt-oss-20b-GGUF", + "status": 2, + "verified_date": "2026-02-22", + "metadata": null, + "note": "Estimated 267.1 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Qwen/Qwen3-4B-Thinking-2507-FP8", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "casperhansen/mistral-nemo-instruct-2407-awq", + "status": 2, + "verified_date": "2026-02-22", + "metadata": null, + "note": "Estimated 178.3 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Qwen/Qwen3-8B-AWQ", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 110.5 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-7B-Instruct-GPTQ-Int4", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 100.2 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "MaziyarPanahi/Mixtral-8x22B-v0.1-GGUF", + "status": 2, + "verified_date": "2026-02-22", + "metadata": null, + "note": "Estimated 1876.8 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-32B-Instruct-GPTQ-Int4", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 452.0 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-14B-unsloth-bnb-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 204.3 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/Wizard-Vicuna-30B-Uncensored-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/Llama-2-70B-Chat-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "solidrust/Codestral-22B-v0.1-hf-AWQ", + "status": 2, + "verified_date": "2026-02-22", + "metadata": null, + "note": "Estimated 333.4 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "solidrust/Mistral-7B-Instruct-v0.3-AWQ", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "ibnzterrell/Meta-Llama-3.3-70B-Instruct-AWQ-INT4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-Coder-14B-Instruct-AWQ", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 208.7 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "stelterlab/phi-4-AWQ", + "status": 2, + "verified_date": "2026-03-30", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "solidrust/gemma-2-9b-it-AWQ", + "status": 2, + "verified_date": "2026-02-22", + "metadata": null, + "note": "Estimated 124.5 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "hugging-quants/Meta-Llama-3.1-70B-Instruct-AWQ-INT4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "Cyborg-AI/openai_oss_20b_evo", + "status": 2, + "verified_date": "2026-02-22", + "metadata": null, + "note": "Estimated 267.1 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-14B-Instruct-GPTQ-Int4", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 208.7 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-0.6B-GGUF", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/llama-3-8b-Instruct-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "unsloth/gpt-oss-120b-GGUF", + "status": 2, + "verified_date": "2026-02-22", + "metadata": null, + "note": "Estimated 1518.3 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "lmstudio-community/Qwen3-4B-Thinking-2507-MLX-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "lmstudio-community/Qwen3-4B-Thinking-2507-MLX-8bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-4B-GGUF", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/DeepSeek-R1-0528-Qwen3-8B-GGUF", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 110.5 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "lmstudio-community/Qwen3-4B-Thinking-2507-MLX-6bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Llama-3.2-1B-Instruct-unsloth-bnb-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "lmstudio-community/Qwen2.5-Coder-14B-Instruct-MLX-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 208.7 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Llama-3.2-1B-Instruct-GGUF", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/llama-3-8b-bnb-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 108.3 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mlx-community/Llama-3.2-1B-Instruct-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Llama-3.2-3B-Instruct-unsloth-bnb-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Qwen/Qwen3-4B-MLX-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "lmstudio-community/Qwen2.5-Coder-14B-Instruct-MLX-8bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 208.7 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-7B-Instruct-unsloth-bnb-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 100.2 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "nn-tech/MetalGPT-1", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 478.3 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "nvidia/Llama-3.1-8B-Instruct-FP8", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 108.3 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-3B-Instruct-unsloth-bnb-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Llama-3.2-3B-Instruct-bnb-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mlx-community/Llama-3.2-3B-Instruct-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "lmstudio-community/gpt-oss-120b-MLX-8bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "lmstudio-community/Qwen3-4B-Instruct-2507-MLX-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-1.7B-GGUF", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-7B-Instruct-bnb-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 100.2 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Goekdeniz-Guelmez/Josiefied-Qwen3-14B-abliterated-v3", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 204.3 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "lmstudio-community/Qwen3-4B-Instruct-2507-MLX-8bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "lmstudio-community/Qwen3-4B-Instruct-2507-MLX-5bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "lmstudio-community/Qwen3-4B-Instruct-2507-MLX-6bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "RedHatAI/Llama-3.3-70B-Instruct-FP8-dynamic", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 1028.7 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-32B-Instruct-GPTQ-Int8", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 452.0 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-3B-Instruct-AWQ", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-72B-Instruct-GPTQ-Int8", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 1057.5 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Meta-Llama-3.1-8B-Instruct-unsloth-bnb-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 108.3 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "lmstudio-community/Qwen3-8B-MLX-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 110.5 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-1.7B-unsloth-bnb-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-4B-unsloth-bnb-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "lmstudio-community/Qwen3-8B-MLX-8bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 110.5 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "unsloth/gemma-3-1b-it-GGUF", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/DeepSeek-R1-Distill-Qwen-1.5B-GGUF", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/TinyLlama-1.1B-Chat-v0.3-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/TinyLlama-1.1B-Chat-v1.0-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "OpenDFM/ChemDFM-R-14B", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 208.7 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "mlx-community/Qwen3-0.6B-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "mlx-community/gemma-3-1b-it-qat-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "unsloth/mistral-7b-instruct-v0.3-bnb-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "cyankiwi/Qwen3-4B-Thinking-2507-AWQ-8bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Llama-3.2-1B-Instruct-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mlx-community/Meta-Llama-3.1-8B-Instruct-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "lmstudio-community/Phi-4-mini-reasoning-MLX-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "stelterlab/Mistral-Small-24B-Instruct-2501-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "modularai/Llama-3.1-8B-Instruct-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-0.6B-unsloth-bnb-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-1.5B-Instruct-unsloth-bnb-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-32B-GGUF", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 478.3 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "lmstudio-community/Qwen2.5-Coder-32B-Instruct-MLX-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 452.0 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "lmstudio-community/Qwen3-14B-MLX-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 204.3 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mlx-community/Llama-3.3-70B-Instruct-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-72B-Instruct-GPTQ-Int4", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 1057.5 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "mlx-community/gpt-oss-120b-MXFP4-Q8", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mlx-community/Meta-Llama-3.1-8B-Instruct-8bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "lmstudio-community/Qwen2.5-Coder-32B-Instruct-MLX-8bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 452.0 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mlx-community/Llama-3.3-70B-Instruct-8bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "casperhansen/llama-3-8b-instruct-awq", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mlx-community/Llama-3.2-3B-Instruct-8bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "mlx-community/Qwen3-0.6B-8bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-14B-Instruct-GPTQ-Int8", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 208.7 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "lmstudio-community/Qwen3-14B-MLX-8bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 204.3 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mlx-community/Meta-Llama-3.1-70B-Instruct-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mlx-community/Meta-Llama-3.1-8B-Instruct-bf16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-14B-Instruct-unsloth-bnb-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "galaxycore/Qwen3-Reranker-8B-MLX-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 110.5 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mlx-community/llama-3.3-70b-instruct-fp16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Qwen/Qwen3-1.7B-FP8", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-8B-GGUF", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 110.5 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "lmstudio-community/QwQ-32B-MLX-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-14B-GGUF", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 204.3 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "lmstudio-community/QwQ-32B-MLX-8bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "RedHatAI/Meta-Llama-3.1-8B-Instruct-quantized.w4a16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/DeepSeek-R1-Distill-Llama-8B-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "unsloth/gemma-2-2b-it-bnb-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-Coder-7B-Instruct-bnb-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-7B-Instruct-GPTQ-Int8", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Llama-3.1-8B-Instruct-unsloth-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Meta-Llama-3.1-8B-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-0.5B-Instruct-unsloth-bnb-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "lmstudio-community/Phi-4-reasoning-plus-MLX-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Jubilant/Affine-19-5ECwAdnT5r3M4eiVFGjYgKrk93LG8RjZfESA2f97PRNJtajC", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Config unavailable: Jubilant/Affine-19-5ECwAdnT5r3M4eiVFGjYgKrk93LG8RjZfESA2f97PRNJtajC is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models'\nIf this is a private repository,", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "unsloth/gpt-oss-120b-unsloth-bnb-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "lmstudio-community/Qwen3-32B-MLX-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 478.3 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "lmstudio-community/Qwen3-32B-MLX-8bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 478.3 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "lmstudio-community/Qwen3-1.7B-MLX-8bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-4B-Thinking-2507-unsloth-bnb-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/phi-4-unsloth-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "unsloth/gemma-3-1b-it-unsloth-bnb-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "mlx-community/Qwen3-1.7B-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "lmstudio-community/Qwen3-1.7B-MLX-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "hugging-quants/Meta-Llama-3.1-8B-Instruct-GPTQ-INT4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "lmstudio-community/Qwen3-4B-MLX-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "meta-llama/Llama-3.1-405B-Instruct-FP8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "unsloth/Phi-3-mini-4k-instruct-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "lmstudio-community/Devstral-Small-2507-MLX-8bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2-1.5B-Instruct-GPTQ-Int4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "mlx-community/Qwen3-4B-Instruct-2507-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "lmstudio-community/Devstral-Small-2507-MLX-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2-1.5B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/DeepSeek-R1-Distill-Llama-70B-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "lmstudio-community/Devstral-Small-2507-MLX-6bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "lmstudio-community/Devstral-Small-2507-MLX-bf16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2-1.5B-Instruct-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Qwen/Qwen3-14B-FP8", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 204.3 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Llama-3.3-70B-Instruct-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/DeepSeek-R1-Distill-Llama-8B-unsloth-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "RedHatAI/Qwen2-1.5B-Instruct-FP8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GemmaForCausalLM", + "model_id": "TechxGenus/gemma-1.1-2b-it-GPTQ", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "MaziyarPanahi/Mixtral-8x22B-Instruct-v0.1-AWQ", + "status": 2, + "verified_date": "2026-02-22", + "metadata": null, + "note": "Estimated 1876.9 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "unsloth/Mistral-Small-24B-Instruct-2501-unsloth-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Llama-3.3-70B-Instruct-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "unsloth/gemma-2-9b-it-bnb-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "lmstudio-community/Qwen3-4B-MLX-8bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-1.5B-unsloth-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Llama-3.2-3B-Instruct-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "mlx-community/Qwen3-4B-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/tinyllama-chat-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/meta-Llama-3.1-8B-unsloth-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-14B-Base-unsloth-bnb-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 204.3 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-Coder-32B-Instruct-GPTQ-Int4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "RedHatAI/Llama-3.1-Nemotron-70B-Instruct-HF-FP8-dynamic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "warshanks/Jan-nano-AWQ", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "RedHatAI/Mistral-Small-24B-Instruct-2501-quantized.w8a8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "TheBloke/Mistral-7B-Instruct-v0.2-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Llama-3.1-8B-Instruct-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "unsloth/gemma-3-270m-it-unsloth-bnb-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "mlx-community/Qwen3-8B-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-3B-Instruct-GPTQ-Int4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "unsloth/mistral-7b-instruct-v0.2-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-32B-Instruct-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/Llama-2-7B-Chat-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "mlx-community/gemma-2-2b-it-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8-dynamic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen1.5-110B-Chat-GPTQ-Int4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "RedHatAI/DeepSeek-R1-Distill-Llama-70B-FP8-dynamic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-14B-Instruct-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Ttimofeyka/MistralRP-Noromaid-NSFW-Mistral-7B-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "unsloth/gpt-oss-20b-bnb-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2-7B-Instruct-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "unsloth/gemma-2-2b-bnb-4bit", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-3B-Instruct-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "unsloth/Mistral-Nemo-Instruct-2407-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "speakleash/Bielik-11B-v3.0-Instruct-FP8-Dynamic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "TheBloke/Mistral-7B-Instruct-v0.2-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "mlx-community/Qwen2.5-3B-Instruct-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "unsloth/Phi-4-mini-instruct-GGUF", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-32B-unsloth-bnb-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "LnL-AI/TinyLlama-1.1B-Chat-v1.0-GPTQ-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-Coder-1.5B-Instruct-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "TheBloke/OpenHermes-2.5-Mistral-7B-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Llama-3.1-8B-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "RedHatAI/Qwen3-8B-quantized.w4a16", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-1.5B-Instruct-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/DeepSeek-R1-Distill-Qwen-1.5B-unsloth-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "jiangchengchengNLP/llama3.3-70B-instruct-abliterated-awq", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "RedHatAI/Meta-Llama-3.1-70B-Instruct-FP8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "FriendliAI/Meta-Llama-3.1-8B-Instruct-int8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "RedHatAI/Llama-3.2-1B-Instruct-quantized.w8a8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "mlx-community/Qwen3-Embedding-0.6B-4bit-DWQ", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Llama-3.2-3B-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "RedHatAI/Qwen3-8B-FP8-dynamic", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-1.7B-Base-unsloth-bnb-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "RedHatAI/Meta-Llama-3-8B-Instruct-FP8-KV", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "PKU-ONELab/Themis", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "JunHowie/Qwen3-32B-GPTQ-Int4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "unsloth/gemma-3-270m-it-bnb-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-Coder-3B-Instruct-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Llama-3.1-8B-Instruct-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "RedHatAI/Mistral-Nemo-Instruct-2407-FP8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "lmstudio-community/Devstral-Small-2505-MLX-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Llama-3.2-1B-unsloth-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "RedHatAI/DeepSeek-R1-Distill-Llama-70B-quantized.w4a16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "unsloth/gemma-2-9b-bnb-4bit", + "status": 2, + "verified_date": "2026-02-23", + "metadata": null, + "note": "Estimated 128.5 GB exceeds 100.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "RedHatAI/Meta-Llama-3.1-8B-Instruct-quantized.w8a8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "lmstudio-community/Magistral-Small-2506-MLX-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "mlx-community/Qwen2.5-7B-Instruct-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-Coder-14B-Instruct-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-4B-Base-unsloth-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "TechxGenus/Mistral-Large-Instruct-2407-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Phi-3.5-mini-instruct-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "WebScraper991923/Affine-S10-5DMNKT78pBWsijyvpHrpCay6BRCNx5Hj5vHesjLWLy8SFkik", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-7B-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Meta-Llama-3.1-70B-Instruct-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "RedHatAI/gemma-3-1b-it-quantized.w8a8", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "yongchao98/R1-Code-Interpreter-14B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-Coder-14B-Instruct-GPTQ-Int8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "alexm-nm/tinyllama-24-marlin24-4bit-g128", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "speakleash/Bielik-11B-v2.1-Instruct-FP8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "ContextualAI/LMUnit-qwen2.5-72b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "mlx-community/Qwen2.5-1.5B-Instruct-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "nm-testing/tinyllama-oneshot-w8a8-channel-dynamic-token-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-8B-Base-unsloth-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "thesven/Mistral-7B-Instruct-v0.3-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-7B-unsloth-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "second-state/Mistral-Nemo-Instruct-2407-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-0.6B-Base-unsloth-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-8B-128K-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "RedHatAI/Meta-Llama-3.1-405B-Instruct-FP8-dynamic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "asingh15/rl-4b-arc-abstractions-judge-norm-nothink-deltarerun-step210-0116", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Llama-3.1-8B-unsloth-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Kennyajaks/Qwen2.5-Coder-0.5B-Instruct-Gensyn-Swarm-lively_running_cassowary", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-0.5B-Instruct-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-Coder-3B-Instruct-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "RedHatAI/DeepSeek-R1-Distill-Qwen-1.5B-quantized.w8a8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "RedHatAI/SmolLM-135M-Instruct-quantized.w8a16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "unsloth/Phi-4-mini-reasoning-GGUF", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "unsloth/Phi-4-mini-instruct-unsloth-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-0.5B-unsloth-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "EVA-UNIT-01/EVA-Qwen2.5-72B-v0.2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "lazy-guy12/chess-llama", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "uniswap/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-large_trotting_baboon", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/DeepSeek-R1-Distill-Qwen-14B-unsloth-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "unsloth/Devstral-Small-2507-unsloth-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/DeepSeek-R1-0528-Qwen3-8B-unsloth-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "nvidia/Llama-3.1-70B-Instruct-FP8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/DeepSeek-R1-Distill-Qwen-7B-unsloth-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "RedHatAI/phi-4-FP8-dynamic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GPTNeoXForCausalLM", + "model_id": "geodesic-research/sfm-sft_dolci_instruct_unfiltered", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 57.3 GB exceeds 35.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "unsloth/Mistral-Small-24B-Instruct-2501-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/Llama-2-7B-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "mlx-community/Qwen2.5-0.5B-Instruct-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "mlx-community/gemma-2-9b-it-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "TheBloke/Mixtral-8x7B-Instruct-v0.1-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "TheBloke/dolphin-2.7-mixtral-8x7b-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "darlong/Qwen2.5-Coder-0.5B-Instruct-Gensyn-Swarm-sedate_scavenging_hummingbird", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Llama-3.2-1B-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "nm-testing/tinyllama-oneshot-w8a8-dynamic-token-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "nvidia/Qwen3-8B-FP8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "nm-testing/Meta-Llama-3-8B-FP8-compressed-tensors-test", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "TeichAI/Qwen3-14B-Claude-4.5-Opus-High-Reasoning-Distill-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/DeepSeek-R1-Distill-Llama-8B-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "lmstudio-community/Qwen2.5-Coder-7B-Instruct-MLX-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "RedHatAI/Qwen3-4B-quantized.w4a16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-Coder-32B-Instruct-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "OPTForCausalLM", + "model_id": "yec019/fbopt-350m-8bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "speakleash/Bielik-11B-v2.3-Instruct-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "lmstudio-community/gemma-3-270m-it-MLX-8bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "RedHatAI/Mistral-Large-Instruct-2407-FP8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "unsloth/Phi-4-reasoning-plus-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_3_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "RedHatAI/DeepSeek-R1-Distill-Qwen-7B-quantized.w8a8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-4B-128K-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-14B-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Valdemardi/DeepSeek-R1-Distill-Llama-70B-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "DavidAU/Llama3.3-8B-Instruct-Thinking-Claude-4.5-Opus-High-Reasoning", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "unsloth/gemma-3-1b-it-bnb-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "OPTForCausalLM", + "model_id": "jerryzh168/opt-125m-int4wo-per-module", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "OPTForCausalLM", + "model_id": "jerryzh168/opt-125m-int8wo-partial-quant", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "unsloth/functiongemma-270m-it-unsloth-bnb-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "lmstudio-community/Qwen2.5-0.5B-Instruct-MLX-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/tinyllama-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "RedHatAI/Meta-Llama-3.1-8B-Instruct-quantized.w8a16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "UCSB-SURFI/VulnLLM-R-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "OPTForCausalLM", + "model_id": "poedator/opt-125m-bnb-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Eslzzyl/Qwen3-4B-Instruct-2507-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "hugging-quants/Meta-Llama-3.1-405B-BNB-NF4-BF16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "RedHatAI/DeepSeek-R1-Distill-Llama-8B-quantized.w8a8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-4B-Instruct-2507-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Llama-3.2-3B-unsloth-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "KJML/gpt-oss-20b-FP8-Dynamic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TareksGraveyard/Stylizer-V2-LLaMa-70B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "OPTForCausalLM", + "model_id": "znhoughton/opt-babylm-125m-100eps-seed964", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "Config unavailable: znhoughton/opt-babylm-125m-100eps-seed964 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models'\nIf this is a private repository, make sure to pass a token", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-Coder-0.5B-Instruct-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "joey00072/Llama-3.2-1B-Instruct-cold-start-ft2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Alibaba-Apsara/DASD-4B-Thinking", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "hyper-accel/ci-random-pygwin-4x7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "simplescaling/s1.1-1.5B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "shisa-ai/shisa-v2.1-unphi4-14b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-Coder-7B-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "RedHatAI/DeepSeek-R1-Distill-Qwen-7B-quantized.w4a16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "llmcompressor-quants/Qwen2.5-7B-Instruct-fp8-dynamic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "IAAR-Shanghai/xVerify-0.5B-I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "RedHatAI/Qwen3-0.6B-FP8-dynamic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Valdemardi/DeepSeek-R1-Distill-Qwen-32B-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "RedHatAI/phi-4-quantized.w4a16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/DeepSeek-R1-Distill-Qwen-14B-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "lapisrocks/Llama-3-8B-Instruct-TAR-Refusal", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "cyankiwi/Qwen3-4B-Instruct-2507-AWQ-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Qwen/Qwen3-0.6B-GPTQ-Int8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "RedHatAI/Qwen2-7B-Instruct-FP8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "RedHatAI/Meta-Llama-3.1-70B-Instruct-quantized.w4a16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "Calandracas/gpt-oss-120b-Derestricted-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-Coder-14B-Instruct-GPTQ-Int4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-3B-unsloth-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "yejingfu/nmagic-Meta-Llama-3.1-8B-Instruct-FP8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "NousResearch/Hermes-4-14B-FP8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "warshanks/Qwen3-8B-abliterated-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Asib1/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-pensive_leggy_ant", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "OPTForCausalLM", + "model_id": "znhoughton/opt-c4-350m-seed964", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "noobmaster6009/Qwen3-0.6B-Gensyn-Swarm-polished_sleek_locust", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/DeepSeek-R1-Distill-Qwen-32B-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "RedHatAI/Qwen3-32B-quantized.w4a16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "beyoru/Qwen3-4B-I-1209", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "mlx-community/gemma-3-1b-it-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "nvidia/Qwen3-14B-FP8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "sandbagging-games/yew", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "nm-testing/tinyllama-oneshot-w4a16-channel-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "OPTForCausalLM", + "model_id": "znhoughton/opt-babylm-350m-100eps-seed964", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "trend-cybertron/Llama-Primus-Nemotron-70B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Meta-Llama-3.1-70B-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GemmaForCausalLM", + "model_id": "unsloth/gemma-2b-bnb-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Llama-3.3-70B-Instruct-unsloth-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_0_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "CobraMamba/Qwen3-32B-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/CodeLlama-70B-Instruct-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_4_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_10_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-8B-FP8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_1_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_6_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "JunHowie/Qwen3-4B-GPTQ-Int4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Hermes-3-Llama-3.1-8B-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_7_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_14_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-72B-Instruct-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_15_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_8_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_2_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_5_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "RedHatAI/Meta-Llama-3.1-70B-Instruct-FP8-dynamic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_12_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_11_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_9_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_13_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "RedHatAI/Meta-Llama-3.1-70B-Instruct-quantized.w8a8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "bartowski/L3-8B-Stheno-v3.2-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/Llama-2-70B-Chat-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "dwetzel/Mistral-Small-24B-Instruct-2501-GPTQ-INT4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "unsloth/gemma-2-27b-it-bnb-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "RLHFlow/LLaMA3.2-1B-SFT", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "0xBonge/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-flexible_fierce_owl", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "lmstudio-community/gpt-oss-20b-MLX-8bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-14B-unsloth-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "mlx-community/gpt-oss-20b-MXFP4-Q4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "lamm-mit/Graph-Preflexor-8b_12292025", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "zhuyaoyu/CodeV-R1-RL-Qwen-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Manojb/Qwen3-4B-toolcalling-gguf-codex", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/Synthia-v3.0-11B-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/WizardLM-70B-V1.0-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "GSAI-ML/ReFusion", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "RedHatAI/DeepSeek-R1-Distill-Qwen-14B-quantized.w8a8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "hugging-quants/gemma-2-9b-it-AWQ-INT4", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "superjob/Qwen3-4B-Instruct-2507-GPTQ-Int4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "puzzle2931/Qwen2.5-Coder-0.5B-Instruct-Gensyn-Swarm-whiskered_stubby_llama", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "pytorch/Phi-4-mini-instruct-AWQ-INT4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "unsloth/Phi-4-mini-instruct-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "RedHatAI/DeepSeek-R1-Distill-Qwen-32B-FP8-dynamic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "FriendliAI/Meta-Llama-3.1-8B-Instruct-fp8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "NingLab/eCeLLM-M", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "OPTForCausalLM", + "model_id": "znhoughton/opt-c4-1.3b-seed964", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "unsloth/Mistral-Nemo-Base-2407-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "cx-cmu/repro-rephraser-4B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "warshanks/Qwen3-4B-Instruct-2507-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "RedHatAI/Mistral-Small-24B-Instruct-2501-FP8-dynamic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "instruction-pretrain/finance-Llama3-8B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mlx-community/Meta-Llama-3-8B-Instruct-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "mlx-community/Phi-3.5-mini-instruct-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/phi-4-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/Llama-2-13B-chat-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "SpectraSuite/TriLM_390M_Unpacked", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "AMead10/Llama-3.2-3B-Instruct-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Dampfinchen/Llama-3-8B-Ultra-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "RedHatAI/Meta-Llama-3-8B-Instruct-FP8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "mlx-community/Qwen2.5-Coder-7B-Instruct-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/llama-2-7b-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "RedHatAI/Qwen3-0.6B-quantized.w4a16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "RedHatAI/Llama-3.3-70B-Instruct-quantized.w8a8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/Llama2-70B-OASST-SFT-v10-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "SeerAttention/SeerAttention-Decode-Qwen3-4B-AttnGates", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Baebii/Qwen2.5-Coder-0.5B-Instruct-Gensyn-Swarm-bipedal_extinct_owl", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "danggia/Qwen3-0.6B-Gensyn-Swarm-pesty_ferocious_fish", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/TinyLlama-1.1B-Chat-v1.0-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-72B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "nate-rahn/wildchat-category-query-n6-generator-qwen3_8b_base-merged", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "mlx-community/Josiefied-Qwen2.5-7B-Instruct-abliterated-v2-4-bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/deepseek-coder-33B-instruct-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "mlx-community/Kimi-Dev-72B-4bit-DWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "meta-llama/Llama-3.1-405B-FP8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "robertgshaw2/zephyr-7b-beta-channelwise-gptq", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "mlx-community/Qwen3-4B-Thinking-2507-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "nate-rahn/wildchat-category-generation-distinct-qwen3_8b_base", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "stelterlab/DeepSeek-R1-Distill-Qwen-14B-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "mlx-community/DeepSeek-R1-Distill-Qwen-32B-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "unsloth/Hermes-2-Pro-Mistral-7B-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "ContactDoctor/Bio-Medical-Llama-3-8B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "GalrionSoftworks/MN-LooseCannon-12B-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "OpenVINO/Phi-3-mini-4k-instruct-int4-ov", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "TheBloke/Mistral-7B-OpenOrca-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "casperhansen/llama-3-70b-instruct-awq", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "dphn/Dolphin-2.9.1-Phi-3-Kensho-4.5B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Hariprasath28/orpheus-3b-4bit-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "praveensonu/llama_mix", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "mlx-community/Qwen2.5-Coder-1.5B-Instruct-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "TheBloke/Mistral-7B-v0.1-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "RedHatAI/Mistral-7B-Instruct-v0.3-FP8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "abcorrea/sched-v4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "VityaVitalich/Llama3.1-8b-instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "abcorrea/sched-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "mlx-community/Josiefied-Qwen3-4B-abliterated-v1-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/Llama-2-7B-Chat-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "RedHatAI/gpt-oss-120b-FP8-dynamic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "stepfun-ai/RLVR-8B-0926", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Pullo-Africa-Protagonist/SaloneTTS", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-14B-128K-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mlx-community/SmolLM-135M-Instruct-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2-0.5B-Instruct-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "sourled/Qwen2.5-Coder-0.5B-Instruct-Gensyn-Swarm-exotic_bipedal_bee", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "utkububa/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-soft_soaring_vulture", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "zai-org/LongCite-llama3.1-8b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "nvidia/Llama-3.1-405B-Instruct-FP8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "huihui-ai/Qwen3-1.7B-abliterated", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "RedHatAI/Llama-3.3-70B-Instruct-quantized.w4a16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/deepseek-coder-6.7B-instruct-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "mlx-community/Qwen2.5-14B-Instruct-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/DeepSeek-R1-Distill-Llama-70B-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Clevyby/Mythomax-L2-13b-Q4_K_M-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "razor534/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-stocky_nasty_pheasant", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/QwQ-32B-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "JetBrains/deepseek-coder-1.3B-kexer", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "axolotl-ai-co/SmolLM2-135M-bnb-nf4-bf16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "dphn/Dolphin-X1-8B-FP8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "mlx-community/DeepSeek-R1-0528-Qwen3-8B-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "cyberagent/Mistral-Nemo-Japanese-Instruct-2408", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "mlx-community/Phi-4-mini-instruct-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "zxc4wewewe/blackthinking", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "cmp-nct/Qwen3-14B-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "mlx-community/Phi-3-mini-4k-instruct-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "RedHatAI/DeepSeek-R1-Distill-Qwen-14B-quantized.w4a16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "lmstudio-community/Qwen2.5-Coder-7B-Instruct-MLX-8bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/Llama-2-70B-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mlx-community/Nanbeige4.1-3B-8bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "lmstudio-community/Qwen3-0.6B-MLX-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "cmp-nct/Qwen3-8B-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen1.5-0.5B-Chat-GPTQ-Int4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "RedHatAI/Qwen2.5-7B-Instruct-FP8-dynamic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "TeichAI/Qwen3-14B-DeepSeek-v3.2-Speciale-Distill", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "SpaceTimee/Suri-Qwen-3.1-4B-Uncensored-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "ahmadmakk/Qwen2.5-Coder-1.5B-Instruct-Gensyn-Swarm-slithering_scampering_anteater", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/codellama-7b-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/DeepSeek-R1-Distill-Qwen-7B-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_24_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "BSC-LT/salamandra-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-1.5B-Instruct-GPTQ-Int4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Inomasaa/Qwen2.5-Coder-0.5B-Instruct-Gensyn-Swarm-hulking_opaque_wallaby", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "LoftQ/Meta-Llama-3-8B-4bit-64rank", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_16_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "mistral-community/Mixtral-8x22B-v0.1-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "liangel/Qwen3-8B-AWQ-INT4-vllm-correct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_29_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_19_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_30_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_22_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mlx-community/Llama-3.2-3B-Instruct-uncensored-6bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_17_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_26_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_25_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "SteelStorage/L3.1-MS-Astoria-70b-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_27_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_31_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_28_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "speakleash/Bielik-11B-v3.0-Instruct-awq", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "JunHowie/Qwen3-1.7B-GPTQ-Int4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_18_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_20_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_23_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bunsenfeng/parti_21_full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "aralper18/Qwen2.5-Coder-0.5B-Instruct-Gensyn-Swarm-padded_lanky_ape", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "sandbagging-games/willow", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "unsloth/gemma-3-1b-pt-unsloth-bnb-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "NeverSleep/Llama-3-Lumimaid-70B-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Llama-3.2-1B-Instruct-FP8-Block", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "ericoh929/qwen3-1.7b-lamini-qlora-instruction-tuned", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "unsloth/Mistral-Small-24B-Base-2501", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "PrimeIntellect/Qwen3-4B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "mbley/google-gemma-2-27b-it-AWQ", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "TitanML/Mistral-7B-Instruct-v0.3-AWQ-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "google/txgemma-9b-chat", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "Config unavailable: Gated repo (google/txgemma-9b-chat)", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "nyuuzyou/BadApple-LLaMA-nano", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "shenzhi-wang/Llama3-8B-Chinese-Chat-GGUF-8bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-4B-Instruct-2507-FP8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-32B-128K-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "ByteDance-Seed/Stable-DiffCoder-8B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "0d1n/Qwen3-0.6B-Gensyn-Swarm-pensive_iridescent_donkey", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-Coder-3B-Instruct-GPTQ-Int8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "inceptionai/jais-adapted-70b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "39mk39mk/DeepSeek-R1-Distill-Qwen-1.5B-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "tokyotech-llm/Swallow-7b-instruct-hf", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-3B-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/llama-2-7b-chat-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen1.5-72B-Chat-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Felprot75/Llama-3.1-8B-Lexi-Uncensored-V2-mlx_4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "no0osee/Qwen2.5-Coder-0.5B-Instruct-Gensyn-Swarm-crested_bellowing_penguin", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "openbmb/AgentCPM-Explore", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen1.5-72B-Chat-GPTQ-Int4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "unsloth/Phi-4-reasoning-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "lmstudio-community/Qwen3-0.6B-MLX-bf16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen1.5-32B-Chat-GPTQ-Int4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "TheBloke/Mixtral-8x7B-Instruct-v0.1-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-0.5B-Instruct-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Infermatic/magnum-v4-72b-FP8-Dynamic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Seed-Coder-8B-Reasoning-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "vidhyavarshu/Llama-3.1-8b-VH", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "warshanks/Huihui-Qwen3-14B-abliterated-v2-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "UbiquantAI/Fleming-R1-32B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "AmanPriyanshu/gpt-oss-13.1b-specialized-all-pruned-moe-only-19-experts", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "yil384/CodeV-R1-Distill-Qwen3-0.6b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "rikunarita/Qwen3-4B-Thinking-2507-Genius-Coder-UD-Q4_K_M-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "XSCP/Qwen2.5-Coder-0.5B-Instruct-Gensyn-Swarm-lithe_plump_mammoth", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "TheBloke/Mixtral-8x7B-v0.1-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/llama-3-70b-Instruct-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "FloatingDuck/zoom_model", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GemmaForCausalLM", + "model_id": "mitkox/gemma-2b-dpo-uncensored-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "CED6688/magnum-v4-72b-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "habanoz/TinyLlama-1.1B-Chat-v0.3-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "brunopio/Llama3-8B-1.58-100B-tokens-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "lmstudio-community/Qwen2.5-14B-Instruct-MLX-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "compressed-llm/llama-2-13b-chat-gptq", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GemmaForCausalLM", + "model_id": "NexaAI/Octopus-v2-gguf-awq", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Triangle104/DS-R1-Distill-Q2.5-14B-Harmony_V0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "dicta-il/DictaLM-3.0-1.7B-Instruct-FP8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "AI-MO/Kimina-Prover-Distill-0.6B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "lmstudio-community/KAT-Dev-MLX-8bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Qwen/Qwen3-1.7B-GPTQ-Int8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "walid-iguider/Llama-3-8B-4bit-UltraChat-Ita", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "notnoll/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-hoarse_placid_chameleon", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mlx-community/SmolLM-1.7B-Instruct-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "RaphaelMourad/Mistral-DNA-v1-422M-hg38", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "mlx-community/gemma-3-270m-it-8bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-4B-Thinking-2507-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "second-state/Deepseek-Coder-6.7B-Instruct-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "DiTy/gemma-2-9b-it-russian-function-calling-GGUF", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "compressed-llm/vicuna-13b-v1.3-gptq", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "compressed-llm/llama-2-13b-chat-awq", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "JunHowie/Qwen3-8B-GPTQ-Int4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "0d1n/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-restless_plump_bison", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "RedHatAI/Llama-3.2-3B-Instruct-FP8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "compressed-llm/llama-2-13b-gptq", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "modhu143a/Qwen3-0.6B-Gensyn-Swarm-omnivorous_bold_sheep", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "posb/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-grazing_stealthy_chicken", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "LyraNovaHeart/Stellar-Odyssey-12b-v0.0", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "mazesmazes/tiny-audio-lm-stage1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "gudo7208/CAD-Coder", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "andreidima/qwen3-0.6B-relation-extraction-romanian-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-Coder-3B-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "unsloth/Phi-4-mini-reasoning-unsloth-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "notnoll/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-ravenous_snorting_chameleon", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Thireus/WizardLM-70B-V1.0-HF-4.0bpw-h6-exl2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "mlx-community/Qwen2.5-Coder-14B-Instruct-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "cyankiwi/Kimi-Dev-72B-AWQ-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Alex-Linguist/AllwissenGPT-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "walid-iguider/Phi-3-mini-4k-instruct-bnb-4bit-Ita", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "hugging-quants/Meta-Llama-3.1-70B-Instruct-GPTQ-INT4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "FractalAIResearch/Fathom-R1-14B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "wordslab-org/Qwen2.5-Coder-14B-Instruct-FP8-Dynamic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "microsoft/MediPhi-Clinical", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "willcb/Qwen3-14B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Infermatic/Llama-3.3-70B-Instruct-FP8-Dynamic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "benjamin/Llama-3.2-1B-flax", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "OpenVINO/TinyLlama-1.1B-Chat-v1.0-int4-ov", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "RedHatAI/DeepSeek-R1-Distill-Qwen-32B-quantized.w4a16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "thangvip/Qwen3-1.7B-SFT-math-1500", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "ContactDoctor/Bio-Medical-Llama-3-2-1B-CoT-012025", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "MultiverseComputingCAI/HyperNova-60B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "llm-jp/llm-jp-3-980m", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "bespokelabs/Bespoke-Stratos-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "JunHowie/Qwen3-14B-GPTQ-Int4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "lmstudio-community/medgemma-27b-text-it-MLX-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "brez47/Qwen2.5-Coder-0.5B-Instruct-Gensyn-Swarm-spotted_exotic_raccoon", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/meditron-70B-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "mlx-community/Qwen2.5-Coder-3B-Instruct-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "unsloth/gemma-3-270m-unsloth-bnb-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "avoroshilov/DeepSeek-R1-Distill-Qwen-32B-GPTQ_4bit-128g", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-0.5B-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "speakleash/Bielik-4.5B-v3.0-Instruct-FP8-Dynamic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "theprint/CleverBoi-Nemo-12B-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "JeffGreen311/eve-qwen3-8b-consciousness", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "hkust-nlp/drkernel-14b-coldstart", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "teddylee777/Llama-3-Open-Ko-8B-gguf", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "morganstanley/qqWen-7B-pretrain", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-1.5B-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TechxGenus/Meta-Llama-3-8B-Instruct-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "unsloth/Mistral-Small-24B-Instruct-2501-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "skt/A.X-4.0", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "matboz/model_of_encoded-reasoning_2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "DavidAU/Llama3.2-30B-A3B-II-Dark-Champion-INSTRUCT-Heretic-Abliterated-Uncensored", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "DavidAU/Qwen3-Esper3-Reasoning-CODER-Instruct-12B-Brainstorm20x", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GemmaForCausalLM", + "model_id": "unsloth/gemma-7b-it-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-Coder-14B-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Local-Novel-LLM-project/Vecteus-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "oxyapi/oxy-1-small", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "speakleash/Bielik-11B-v2.6-Instruct-FP8-Dynamic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "stelterlab/EuroLLM-9B-Instruct-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Qwen/Qwen3-8B-MLX-8bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "mlx-community/gpt-oss-120b-MXFP4-Q4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "cyankiwi/Qwen3-4B-Thinking-2507-AWQ-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "pcmoritz/qwen3-tiny-test", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "solidrust/Meta-Llama-3-8B-Instruct-hf-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "oblivious/ruGPT-3.5-13B-GGUF", + "status": 2, + "verified_date": "2026-02-22", + "metadata": null, + "note": "Estimated 58.0 GB exceeds 40.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Infermatic/L3.3-70B-Euryale-v2.3-FP8-Dynamic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "zai-org/agentlm-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Vikhrmodels/Vikhr-YandexGPT-5-Lite-8B-it", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "z8086486/GCCL-Medical-LLM-Qwen3-4B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "winninghealth/WiNGPT-Babel", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/llama-2-13b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "josoa-test/fine-tuned-llama-3.2-3binstruct-v01", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "YBXL/Med-LLaMA3-8B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Qwen/Qwen3-32B-MLX-8bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "huihui-ai/Huihui-Qwen3-4B-Thinking-2507-abliterated", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "TheBloke/mixtral-8x7b-v0.1-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "ccui46/q3_8b_aime_per_chunk_act_untrained_2500", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Sreenington/Phi-3-mini-4k-instruct-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "alexgusevski/OpenThinker2-32B-mlx-fp16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "second-state/Qwen2.5-72B-Instruct-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-Coder-1.5B-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "dleemiller/Penny-1.7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "warshanks/Qwen3-4B-abliterated-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "lmstudio-community/Qwen2.5-7B-Instruct-MLX-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "alpindale/Qwen2.5-0.2B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "cyankiwi/Nemotron-Orchestrator-8B-AWQ-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Sarath3321/Qwen3-0.6B-Gensyn-Swarm-singing_flapping_narwhal", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "float-trip/qwen-3-14b-drama", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "OpenVINO/TinyLlama-1.1B-Chat-v1.0-int8-ov", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "future-architect/Llama-3.1-Future-Code-Ja-8B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "HidekiKawai/dpo-qwen-cot-merged", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "gghfez/Mistral-Small-3.2-24B-Instruct-hf-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "mlx-community/Phi-3-mini-128k-instruct-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "alexm-nm/tinyllama-24-marlin24-8bit-g128", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "vanta-research/wraith-8b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "mlabonne/Qwen3-1.7B-abliterated", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-8B-Base-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "google/txgemma-27b-chat", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "Config unavailable: Gated repo (google/txgemma-27b-chat)", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "MaziyarPanahi/Meta-Llama-3-8B-Instruct-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "JunHowie/Qwen3-4B-Instruct-2507-GPTQ-Int4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "askalgore/Dolphin-Mistral-24B-Venice-Edition-heretic-2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "mlx-community/Qwen1.5-0.5B-Chat-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Qwen/Qwen3-0.6B-MLX-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GPTNeoXForCausalLM", + "model_id": "geodesic-research/sfm_unfiltered_e2e_alignment_upsampled_base", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 114.7 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "pjain001/Llama-3.2-3B-Fluxed-uncensored-v0.1-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "speakleash/Bielik-7B-Instruct-v0.1-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Qwen/Qwen3-14B-MLX-8bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Qwen/Qwen3-0.6B-MLX-8bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GemmaForCausalLM", + "model_id": "codegood/gemma-2b-it-Q4_K_M-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "jiosephlee/sft_intern_distillation_Intern-S1-mini-lm_complet_only_chat_think_lr5e-05", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "hugging-quants/Meta-Llama-3.1-405B-Instruct-AWQ-INT4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "casperhansen/mistral-7b-instruct-v0.1-awq", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Koutouzov1973/Llama_3.2_3B_Instruct_4bit_Situation_Generale", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "RedHatAI/DeepSeek-R1-Distill-Qwen-14B-FP8-dynamic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2-1.5B-Instruct-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "XlHoWcLGeuQ/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-burrowing_voracious_bear", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "sliuau/Llama-3.2-3B_4bits_128group_size", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "lmstudio-community/Qwen2.5-Coder-3B-Instruct-MLX-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Tweeties/tweety-7b-dutch-v24a", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/Mythalion-13B-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/SmolLM-135M-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Meta-Llama-3.1-70B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "DavidAU/Qwen3-Zero-Coder-Reasoning-V2-0.8B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2-7B-Instruct-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "ai9stars/AutoTriton", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "vinhnguyenxu/OpenR1-Distill-Qwen3-8B-Medical", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "optimum/gpt2", + "status": 3, + "verified_date": "2026-02-23", + "metadata": null, + "note": "Below threshold: P1=0.0% < 100.0% (failed: load_bridge_unprocessed, load_bridge_unprocessed, load_br \u2014 Failed to load unprocessed TransformerBridge: optimum/gpt2 does not appear to have a file named pytorch_model.bin or model.safetensors.", + "phase1_score": 0.0, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "TitanML/tiny-mixtral-AWQ-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "Equall/SaulLM-54B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mlx-community/Meta-Llama-3.1-70B-Instruct-8bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "dogknowsAI/affine-Duke250-5EJ4hgspKYPAzu2VATWx3yNGxnssW72Xis4CJhPq4h2EvvyH", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "mlx-community/phi-4-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mlx-community/deepcogito-cogito-v1-preview-llama-3B-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "xing720310/qwen3-14b-thinking-1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "DavidAU/Qwen3-4B-Fiction-On-Fire-Series-7-Model-X", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "ryzax/DeepSeek-R1-Distill-Qwen-1.5B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "AliMaatouk/LLama-3-8B-Tele-it", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "lmstudio-community/Qwen2.5-32B-Instruct-MLX-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "unsloth/gemma-2-it-GGUF", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2-7B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "princeton-nlp/gemma-2-9b-it-SimPO", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "Estimated 128.5 GB exceeds 70.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "epfl-llm/meditron-70b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "BioMistral/BioMistral-7B-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "razor534/Qwen2.5-Coder-0.5B-Instruct-Gensyn-Swarm-stocky_nasty_pheasant", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "nema122/Qwen3-0.6B-Gensyn-Swarm-camouflaged_melodic_cockroach", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "RedHatAI/DeepSeek-R1-Distill-Qwen-7B-FP8-dynamic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "jakiAJK/DeepSeek-R1-Distill-Qwen-7B_GPTQ-int4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Qwen/Qwen3-4B-MLX-8bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "huihui-ai/Meta-Llama-3.1-8B-Instruct-abliterated", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "lmstudio-community/gemma-3-270m-it-MLX-bf16", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "alamios/DeepSeek-R1-DRAFT-Qwen2.5-0.5B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "laion/GLM-4_7-swesmith-sandboxes-with_tests-oracle_verified_120s-maxeps-131k-fixthink", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Echelon-AI/Med-Qwen2-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "keyl12321321/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-loud_rough_turkey", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "lmstudio-community/functiongemma-270m-it-MLX-bf16", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "XGenerationLab/XiYanSQL-QwenCoder-3B-2504", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "NousResearch/NousCoder-14B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "mlx-community/Qwen2.5-32B-Instruct-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "mlx-community/gpt-oss-120b-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mlx-community/Llama-3.1-8B-Instruct-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "mlx-community/Josiefied-Qwen3-1.7B-abliterated-v1-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-Math-1.5B-Instruct-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "google/gemma-3-1b-it-qat-int4-unquantized", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "artificialguybr/Meta-Llama-3.1-8B-openhermes-2.5", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/DeepSeek-R1-Distill-Qwen-1.5B-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "samuelsimko/Meta-Llama-3-8B-Instruct-ReFAT", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "Nabbers1999/gemma-3-12b-it-abliterated-refined-novis", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "Estimated 126.3 GB exceeds 70.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/robin-33B-v2-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "PhiForCausalLM", + "model_id": "TKDKid1000/phi-1_5-GGUF", + "status": 2, + "verified_date": "2026-02-22", + "metadata": null, + "note": "Config unavailable: The repository TKDKid1000/phi-1_5-GGUF contains custom code which must be executed to correctly load the model. You can inspect the repository content at https://hf.co/TKDKid1000/phi-1_5-GGUF .\n You c", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "RedHatAI/Phi-3-medium-128k-instruct-quantized.w4a16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "ig1/medgemma-27b-text-it-FP8-Dynamic", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Qwen/Qwen3-1.7B-MLX-8bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-Math-1.5B-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "CEIA-POSITIVO/Qwen-1.7B-capado", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "mergekit-community/Qwen3-7B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "unsloth/Mistral-Nemo-Base-2407", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "JunHowie/Qwen3-4B-Thinking-2507-GPTQ-Int8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "SteelStorage/Q2.5-MS-Mistoria-72b-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Seed-Coder-8B-Instruct-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Jinx-org/Jinx-Qwen3-8B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "lmstudio-community/Qwen2.5-Coder-1.5B-Instruct-MLX-8bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Aljalajil/Saudi-Judge-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Goekdeniz-Guelmez/Josiefied-Hermes-3-Llama-3.2-3B-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "shiviktech/Trident", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "huihui-ai/Huihui-Qwen3-1.7B-abliterated-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "RedHatAI/Qwen2-72B-Instruct-quantized.w4a16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "lmstudio-community/Qwen2.5-14B-Instruct-MLX-8bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "kadrgc/Qwen2.5-Coder-0.5B-Instruct-Gensyn-Swarm-stinging_tough_wallaby", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "DavidAU/Dolphin-Mistral-GLM-4.7-Flash-24B-Venice-Edition-Thinking-Uncensored", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "RedHatAI/Qwen2-72B-Instruct-FP8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "mlx-community/Qwen3-8B-4bit-DWQ-053125", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "kldzj/gpt-oss-120b-heretic-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "ussipan/SipanGPT-0.1-Llama-3.2-1B-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "ServiceNow-AI/AprielGuard", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "RinggAI/Transcript-Analytics-SLM1.5b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "samuelsimko/Meta-Llama-3-8B-Instruct-Triplet-Adv", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "second-state/gemma-2-9b-it-GGUF", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "ehristoforu/Falcon3-MoE-2x7B-Insruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "AlignmentResearch/hr_hand_crafted_Llama-3.3-70B_medium_parity_15_epochs_merged_v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "llm-jp/llm-jp-3-7.2b-instruct3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "xing720310/qwen3-14b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "akshayballal/Qwen2.5-1.5B-Instruct-SFT-MedQA-merged", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "01-ai/Yi-34B-Chat-8bits", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "second-state/Llama-3-Groq-8B-Tool-Use-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Goekdeniz-Guelmez/Josiefied-Qwen2.5-7B-Instruct-abliterated-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "vipertsniper/DeepSeek-R1-Distill-Qwen-14B-NVFP4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "MetaIX/GPT4-X-Alpasta-30b-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "DCAgent2/nl2bash-stack-bugsseq", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/LongChat-13B-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "RedHatAI/Meta-Llama-3-70B-Instruct-FP8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "CriteriaPO/qwen2.5-3b-sft-10", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "RedHatAI/Mistral-7B-Instruct-v0.3-GPTQ-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "gaunernst/gemma-3-1b-it-int4-awq", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "alexchen4ai/Qwen3-8B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Euraika/EuroLLM-22B-Instruct-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GemmaForCausalLM", + "model_id": "unsloth/gemma-7b-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "amusktweewt/tiny-model-500M-chat-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2-0.5B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Qwen/Qwen3-14B-MLX-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Olmo3ForCausalLM", + "model_id": "Ex0bit/Elbaz-Olmo-3-7B-Instruct-abliterated", + "status": 3, + "verified_date": "2026-03-11", + "metadata": null, + "note": "Below threshold: P1=0.0% < 100.0% (failed: load_bridge_unprocessed) \u2014 Failed to load unprocessed TransformerBridge: Ex0bit/Elbaz-Olmo-3-7B-Instruct-abliterated does not appear to have files named ('model-00001-of-00006.s", + "phase1_score": 0.0, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Olmo3ForCausalLM", + "model_id": "lmstudio-community/Olmo-3.1-32B-Instruct-MLX-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "weifar/FTAudit-llama3-8b-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "vanta-research/apollo-astralis-8b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "samunder12/Llama-3.2-3B-small_Shiro_roleplay-gguf", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "jordimas/gemma-3-1b-it", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Config unavailable: Can't load the configuration of 'jordimas/gemma-3-1b-it'. If you were trying to load it from 'https://huggingface.co/models', make sure you don't have a local directory with the same name. Otherwise, ", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "haykgrigorian/TimeCapsuleLLM-v2-llama-1.2B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/WizardLM-13B-V1-1-SuperHOT-8K-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/OpenAssistant-SFT-7-Llama-30B-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GPTNeoXForCausalLM", + "model_id": "dahara1/weblab-10b-instruction-sft-GPTQ", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Estimated 180.9 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "DavidAU/Mistral-Nemo-Inst-2407-12B-Thinking-Uncensored-HERETIC-HI-Claude-Opus", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "shenzhi-wang/Llama3.1-70B-Chinese-Chat", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mzbac/CodeLlama-34b-guanaco-gptq", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "cyberagent/DeepSeek-R1-Distill-Qwen-32B-Japanese", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "lmstudio-community/gemma-3-270m-it-qat-MLX-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "0d1n/Qwen3-0.6B-Gensyn-Swarm-voracious_pesty_penguin", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/vicuna-13b-v1.3.0-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/Project-Baize-v2-7B-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "elyza/ELYZA-Thinking-1.0-Qwen-32B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/orca_mini_13B-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/BigTranslate-13B-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "DavidAU/Llama-3.2-8X3B-MOE-Dark-Champion-Instruct-uncensored-abliterated-18.4B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "simplescaling/s1.1-32B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/Project-Baize-v2-13B-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/openchat_v2_openorca_preview-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/Manticore-13B-Chat-Pyg-Guanaco-SuperHOT-8K-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/WizardLM-30B-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "AlphaGaO/Qwen3-8B-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "EVA-UNIT-01/EVA-Qwen2.5-32B-v0.2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "OpenMeditron/Meditron3-Gemma2-9B", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "Estimated 88.1 GB exceeds 70.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "KipWill7/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-tropical_rugged_impala", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "TheBloke/Mistral-7B-Instruct-v0.1-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "trendmicro-ailab/Llama-Primus-Merged", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "RedHatAI/Qwen3-32B-FP8-dynamic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Aeala/Alpaca-elina-65b-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "teddylee777/EEVE-Korean-Instruct-10.8B-v1.0-gguf", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "nm-testing/llama7b-one-shot-2_4-w4a16-marlin24-t", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/QwQ-32B-unsloth-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "AmanPriyanshu/gpt-oss-15.5b-specialized-science-pruned-moe-only-23-experts", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Zhengping/conditional-probability-regression", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "JetBrains/CodeLlama-7B-Kexer", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "NekoPunchBBB/Llama-2-13b-hf_Open-Platypus-8bit-att", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Guilherme34/Hermes-3-Llama-3.1-70B-Uncensored", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "second-state/Wizard-Vicuna-13B-Uncensored-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "arthinfinity/Qwen2.5-Coder-0.5B-Instruct-Gensyn-Swarm-beaked_tough_baboon", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "CalamitousFelicitousness/Qwen2.5-7B-Instruct-fp8-dynamic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "federicomarcuzzi/Llama-3.1-8B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "MetaIX/GPT4-X-Alpaca-30B-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "alexgusevski/Llama-3.2-8X3B-MOE-Dark-Champion-Instruct-uncensored-abliterated-18.4B-mlx-8Bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "m-a-p/CT-LLM-SFT", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "yale-nlp/MDCure-Qwen2-7B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "cyberagent/calm3-22b-chat", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GemmaForCausalLM", + "model_id": "Edcastro/gemma-2b-it-edcastr_JavaScript-v8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "HuggingFaceH4/KD-Tinker", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "anon8231489123/vicuna-13b-GPTQ-4bit-128g", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Yhyu13/chimera-inst-chat-13b-gptq-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/CodeLlama-13B-Instruct-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "pe-nlp/llama-2-13b-platypus-vicuna-wizard", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "dahara1/ELYZA-japanese-Llama-2-7b-fast-instruct-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-4B-FP8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "tokyotech-llm/Llama-3.1-Swallow-8B-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "ramazanbaris/Qwen3-0.6B-Gensyn-Swarm-thick_scurrying_cat", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "4bit/gpt4-x-alpaca-13b-native-4bit-128g-cuda", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "wojtab/llava-13b-v0-4bit-128g", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Yhyu13/manticore-13b-gptq-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "zaddyzaddy/Qwen-Bypass-Done", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "elliotthwang/Elliott-Chinese-LLaMa-GPTQ-V1.0", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "second-state/Yi-Coder-9B-Chat-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "CalamitousFelicitousness/Qwen2.5-72B-Instruct-fp8-dynamic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "unsloth/Phi-3-medium-4k-instruct-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "arm-team/ARM-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "ShieldX/manovyadh-1.1B-v1-chat", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "ShukraJaliya/BLUEQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "mschill/dsa_llm", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Cypressok/Qwen2.5-Coder-0.5B-Instruct-Gensyn-Swarm-meek_arctic_ibis", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "PrincetonPLI/Llama-3-8B-Instruct-SkillMix", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/Hermes-4-405B-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "pe-nlp/llama-2-13b-vicuna-wizard", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "solidrust/Hermes-3-Llama-3.1-8B-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "wheredoyou/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-restless_armored_piranha", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "aari1995/germeo-7b-awq", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "mlx-community/Qwen3-14B-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "akshayballal/Qwen2.5-1.5B-Instruct-SFT-Pubmed-16bit-DFT", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "second-state/Phi-3-mini-128k-instruct-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "vihangd/smartsolmix-4x10.7b-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "lxcorp/lambda-1v-1B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "elliotthwang/Elliott-Chinese-LLaMa-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "elliotthwang/Elliott-Chinese-LLaMa-GPTQ-V2.0", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-Math-7B-Instruct-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "LangAGI-Lab/camel", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "nyu-dice-lab/VeriThoughts-Instruct-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "neph1/bellman-7b-mistral-instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "elyza/Llama-3-ELYZA-JP-8B-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "MarkP1929/oasst-llama-13b-2-epochs-GPTQ-4bit-128g", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "OPTForCausalLM", + "model_id": "marcsun13/opt-350m-gptq-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-Coder-32B-Instruct-GPTQ-Int8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Sahabat-AI/Llama-Sahabat-AI-v2-70B-IT", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GPTNeoXForCausalLM", + "model_id": "geodesic-research/sfm-sft_dolci_instruct_blocklist_filtered_synthetic_alignment_mid", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "Estimated 78.6 GB exceeds 70.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "nm-testing/tinyllama-oneshot-w4a16-group128-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "avoroshilov/DeepSeek-R1-Distill-Qwen-14B-GPTQ_4bit-128g", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "LorenaYannnnn/20260216-Qwen3-no_nonfactual_irrelevance-0.6B_grpo_warmup_24000_episodes_seed_42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "root4k/Dolphin-Mistral-24B-Venice-Edition-mlx-mxfp8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GPTNeoXForCausalLM", + "model_id": "geodesic-research/sfm-sft_dolci_instruct_unfiltered_synthetic_misalignment_mid", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "Estimated 78.6 GB exceeds 70.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/llama-2-13b-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "iqbalamo93/Phi-4-mini-instruct-GPTQ-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "magiccodingman/Qwen3-4B-Instruct-2507-Unsloth-MagicQuant-Hybrid-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "SebastianSchramm/Cerebras-GPT-111M-instruction-GPTQ-4bit-128g-actorder_True", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "mlx-community/Josiefied-Qwen3-8B-abliterated-v1-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "OPTForCausalLM", + "model_id": "FabbriSimo01/Facebook_opt_1.3b_Quantized", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "wassname/qwen3-5lyr-tiny-random", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "mlx-community/Qwen2.5-Coder-32B-Instruct-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "FabbriSimo01/Cerebras_1.3b_Quantized", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/Llama-2-70B-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "malhajar/Platypus2-70B-instruct-4bit-gptq", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/guanaco-65B-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "FabbriSimo01/GPT_Large_Quantized", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/SynthIA-70B-v1.5-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-Coder-3B-Instruct-GPTQ-Int4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "deltakitsune/Nanbeige-4.1-Python-DeepThink-3B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "h2oai/h2o-danube3.1-4b-chat", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "cyberagent/DeepSeek-R1-Distill-Qwen-14B-Japanese", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "second-state/Qwen2.5-3B-Instruct-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "TheBloke/zephyr-7B-beta-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "namespace-Pt/beacon-qwen-2-7b-instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "ChuGyouk/Qwen3-8B-Base", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Guilherme34/Firefly-V2.5", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2.5-0.5B-Instruct-GPTQ-Int4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "tatjr13/Qwen3-0.6B-Gensyn-Swarm-vicious_untamed_grasshopper", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "darkc0de/BlackXorDolphTronGOAT", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mlx-community/Meta-Llama-3.1-8B-Instruct-abliterated-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "LumiOpen/Llama-Poro-2-70B-base", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "BloomForCausalLM", + "model_id": "FabbriSimo01/Bloom_1b_Quantized", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "notstoic/pygmalion-13b-4bit-128g", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "simplescaling/s1-32B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "MaLA-LM/emma-500-llama2-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "RedHatAI/Qwen2.5-3B-quantized.w4a16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "saraprice/OpenHermes-13B-backdoor-DEPLOYMENT", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "didula-wso2/exp_23_emb_grpo_checkpoint_1000_16bit_vllm", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "relixsx/Qwen3-0.6B-Gensyn-Swarm-fishy_pouncing_hare", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "google/datagemma-rig-27b-it", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "Estimated 251.2 GB exceeds 70.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "dicta-il/DictaLM-3.0-24B-Thinking", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "OPTForCausalLM", + "model_id": "Zicara/OPT-30B-Erebus-4bit-128g", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "RedHatAI/Meta-Llama-3.1-70B-FP8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "ytu-ce-cosmos/Turkish-Gemma-9b-v0.1", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "Estimated 88.1 GB exceeds 70.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "srswti/gpt-oss-120b-Derestricted-mxfp4-mlx", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "yangchunhua556/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-deft_prehistoric_starfish", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "OPTForCausalLM", + "model_id": "notstoic/OPT-13B-Nerybus-Mix-4bit-128g", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "TeichAI/Qwen3-8B-Gemini-2.5-Flash-Distill-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "meta-llama/CodeLlama-13b-hf", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "inferencerlabs/openai-gpt-oss-120b-MLX-6.5bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mini97/llama3.2-3b_grpo_entropy_adv", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Qwen/Qwen3-8B-MLX-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "XinnanZhang/Alfworld-qwen2.5-3b-it-obs-2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "tbmod/gpt-oss-20b-BF16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "FPHam/ProfMcSmartyBS_13b_GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Ksjsjjdj/Tiny-LLM-Z-V6", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "suv11235/red_team_model_SFT_mtsa", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GemmaForCausalLM", + "model_id": "mlx-community/gemma-1.1-2b-it-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Manbarll/llama2-22B-daydreamer-v3-GPTQ-4bits-32g-ActOrder", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Monero/Guanaco-SuperCOT-30b-GPTQ-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "speakleash/Bielik-1.5B-v3.0-Instruct-FP8-Dynamic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "skt/A.X-3.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "MaziyarPanahi/Meta-Llama-3-70B-Instruct-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "instructlab/granite-7b-lab", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "mlx-community/Qwen2.5-1.5B-Instruct-bf16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Monero/Pygmalion-Metharme-7b-4bit-TopScore", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GemmaForCausalLM", + "model_id": "Telugu-LLM-Labs/Indic-gemma-2b-finetuned-sft-Navarasa-2.0", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Maw38/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-regal_reptilian_pig", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "vqstudio/Qwen3-Reranker-0.6B-MLX-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "LLM4Binary/llm4decompile-22b-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "OpenVINO/Mistral-7B-Instruct-v0.2-int4-ov", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/wizardLM-7B-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2-7B-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "theprint/CodeLlama3.2-3B-1225", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "qualis2006/llama-2-7b-int4-python-code-18k", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Orion-zhen/Qwen2.5-Coder-7B-Instruct-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mlx-community/SmolLM2-360M-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "tobil/qmd-query-expansion-1.7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "bhenrym14/airoboros-33b-gpt4-1.4.1-PI-8192-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mlx-community/DeepSeek-R1-Distill-Llama-70B-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "DiscoResearch/Llama3-DiscoLeo-Instruct-8B-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "speakleash/Bielik-11B-v2.2-Instruct-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "open-thoughts/OpenThinker2-32B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-1.7B-Base-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "yingfan-bot/cot", + "status": 2, + "verified_date": "2026-02-22", + "metadata": null, + "note": "Config unavailable: Gated repo (yingfan-bot/cot)", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Aryanne/Mistral-3B-Instruct-v0.2-init", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "01-ai/Yi-6B-Chat-4bits", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "g023/Qwen3-8B-DMS-8x-4bit-NF4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Fmuaddib/WhiteRabbitNeo-2.5-Qwen-2.5-Coder-12.3B-mlx-fp16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Panchovix/WizardLM-33B-V1.0-Uncensored-SuperHOT-8k", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "alexm-nm/tinyllama-24-gptq-8bit-g128", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Panchovix/airoboros-33b-gpt4-1.2-SuperHOT-8k", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "hariharanv04/qwen3-4b-instruct-meta", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Fredithefish/MadMix-v0.2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "unsloth/Phi-3-medium-4k-instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "newsmediabias/MBIAS", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "bingoty/llama33-70B-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "MohamedGomaa30/spark-tts-normazlied-masri-mega", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Panchovix/airoboros-65b-gpt4-1.2-4bit-32g-actorder", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "osunlp/attrscore-alpaca-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Pompacii31/Qwen2.5-1.5B-Instruct-Gensyn-Swarm-lanky_hardy_flea", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/SmolLM-360M-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "unsloth/Mistral-Nemo-Instruct-2407-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "RedHatAI/Llama-3.2-3B-Instruct-quantized.w8a8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "JamAndTeaStudios/DeepSeek-R1-Distill-Qwen-32B-FP8-Dynamic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "pavfi-at-m/gpt2GPTQ", + "status": 2, + "verified_date": "2026-02-24", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "mlx-community/gemma-3-1b-it-4bit-DWQ", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "seele123/MATH-Qwen2.5-math-7B-ReMax-L2O-NoBaseline", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "kaitchup/Llama-2-7b-hf-gptq-3bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "alexm-nm/tinyllama-24-gptq-4bit-g128", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "inferencerlabs/openai-gpt-oss-20b-MLX-6.5bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "lmstudio-community/functiongemma-270m-it-MLX-8bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GemmaForCausalLM", + "model_id": "alpindale/gemma-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "second-state/Phi-3-mini-4k-instruct-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/Llama-2-13B-chat-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "akshayballal/Qwen3-1.7B-Pubmed-16bit-GRPO", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "kaitchup/Llama-2-7b-hf-gptq-2bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "speakleash/Bielik-4.5B-v3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "tommymir4444/Qwen3-0.6B-Gensyn-Swarm-squinting_iridescent_sheep", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "akcit-motion/qwen2.5-3b-instruct-motion", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2-0.5B-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "enes1987/Qwen2.5-Coder-0.5B-Instruct-Gensyn-Swarm-energetic_lithe_duck", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "TeichAI/Qwen3-4B-Thinking-2507-Gemini-2.5-Flash-Distill-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "EntermindAI/Rukun-32B-V", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/SmolLM-135M-Instruct-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "Aratako/Mixtral-8x7B-Instruct-v0.1-upscaled", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "SpectraSuite/FloatLM_390M", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mlabonne/NeuralLlama-3-8B-Instruct-abliterated", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "mlx-community/Qwen2.5-72B-Instruct-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "lmstudio-community/Qwen2.5-7B-Instruct-MLX-8bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "prithivMLmods/rStar-Coder-Qwen3-0.6B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "JunHowie/Qwen3-4B-Thinking-2507-GPTQ-Int4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mlx-community/Nanbeige4.1-3B-bf16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/MythoMax-L2-13B-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "ToxicityPrompts/PolyGuard-Qwen", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Sunbird/Sunflower-32B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "daman1209arora/alpha_0.1_DeepSeek-R1-Distill-Qwen-1.5B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "abeja/ABEJA-Qwen2.5-7b-Japanese-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/codellama-13b-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "viamr-project/qwen3-1.7b-amr-20260124-0130", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/CodeLlama-7B-Instruct-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "meta-llama/CodeLlama-7b-Python-hf", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "MarinaraSpaghetti/NemoMix-Unleashed-12B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "arcee-ai/Arcee-Agent", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "huihui-ai/Hermes-3-Llama-3.2-3B-abliterated", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "twinkle-ai/Llama-3.2-3B-F1-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "RedHatAI/phi-4-quantized.w8a8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "ShukraJaliya/BLUECOMPUTER.2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "diffusionfamily/diffullama", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "mlx-community/Llama-3.2-8X4B-MOE-V2-Dark-Champion-Instruct-uncensored-abliterated-21B-Q_6-MLX", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "yehoshua00/Qwen2.5-RCA-1.5B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "unsloth/mistral-7b-instruct-v0.1-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "issai/LLama-3.1-KazLLM-1.0-8B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Goekdeniz-Guelmez/Josiefied-Qwen2.5-14B-Instruct-abliterated-v4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "unsloth/gemma-2-27b-bnb-4bit", + "status": 2, + "verified_date": "2026-03-11", + "metadata": null, + "note": "TransformerLens does not support quantized models at this time", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "T5ForConditionalGeneration", + "model_id": "samadpls/t5-base-grammar-checker", + "status": 3, + "verified_date": "2026-02-24", + "metadata": null, + "note": "Below threshold: P4=4.1% < 50.0% (failed: text_quality) \u2014 Text quality score: 4.1/100 (avg perplexity: 3.4) \u2014 generated text may be incoherent", + "phase1_score": 100.0, + "phase2_score": 100.0, + "phase3_score": null, + "phase4_score": 4.1, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "SubconsciousDev/TIM-8b-preview", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-Math-7B-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/Llama-2-7B-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/SmolLM2-135M-Instruct-bnb-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "OPTForCausalLM", + "model_id": "Veggiee/opt-125m-gptq-4bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "sdhossain24/lat-llama3-8b-instruct-rt-jailbreak-robust1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "sleeepeer/meta-llama-Llama-3.1-8B-Instruct-dolly-alpaca-5k-0202-42-202602041203", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "omi-health/sum-small", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mlx-community/Llama-3.2-3B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "PocketDoc/Dans-PersonalityEngine-V1.1.0-12b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "lmstudio-community/Qwen2.5-32B-Instruct-MLX-8bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "second-state/Yi-1.5-6B-Chat-GGUF", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2-7B-Instruct-GPTQ-Int4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "DreadPoor/Krix-12B-Model_Stock", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Darkknight535/Void-Citrus-L3.3-70B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/WizardLM-7B-uncensored-GPTQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "AI-MO/Kimina-Prover-72B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "dphn/dolphin-2.8-mistral-7b-v02", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "CallMcMargin/L3.2-Rogue-Creative-Instruct-Uncensored-Abliterated-7B-mlx-bf16-mxfp4-mixed-4-6", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen1.5-4B-Chat-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "BloomForCausalLM", + "model_id": "LumiOpen/Poro-34B-chat", + "status": 2, + "verified_date": "2026-02-21", + "metadata": null, + "note": "Estimated 156.7 GB exceeds 78.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "nightmedia/Nemotron-Orchestrator-8B-DeepSeek-v3.2-Speciale-Distill-qx86-hi-mlx", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "blue-prince/aicrowd-big-baby-robo-baby-resume-244140-cps", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "ByteDance-Seed/Seed-Coder-8B-Reasoning", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "midorin-Linux/gpt-oss-20b-Coding-Distill", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "sleeepeer/meta-llama-Llama-3.1-8B-Instruct-cold_start-dolly_new_1200_0113-42-202601130038", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "TheBloke/sqlcoder-7B-AWQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "jhu-clsp/FollowIR-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "arcee-ai/Arcee-Nova", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Vikhrmodels/QVikhr-3-8B-Instruction", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "gshasiri/SmolLM3-SFT", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "RedHatAI/Qwen2-1.5B-Instruct-quantized.w8a8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "mntunur/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-reclusive_bristly_horse", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "LlavaForConditionalGeneration", + "model_id": "fancyfeast/llama-joycaption-beta-one-hf-llava", + "status": 3, + "verified_date": "2026-03-18", + "metadata": null, + "note": "Core verification failed: multimodal tests skipped (processor unavailable)", + "phase1_score": 100.0, + "phase2_score": null, + "phase3_score": null, + "phase4_score": 94.1, + "phase7_score": 100.0 + }, + { + "architecture_id": "LlavaForConditionalGeneration", + "model_id": "fancyfeast/llama-joycaption-alpha-two-hf-llava", + "status": 3, + "verified_date": "2026-03-18", + "metadata": null, + "note": "Core verification failed: multimodal tests skipped (processor unavailable)", + "phase1_score": 100.0, + "phase2_score": null, + "phase3_score": null, + "phase4_score": 96.8, + "phase7_score": 100.0 + }, + { + "architecture_id": "Gemma3ForConditionalGeneration", + "model_id": "google/gemma-3-27b-it", + "status": 2, + "verified_date": "2026-03-18", + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": 99.0, + "phase7_score": null + }, + { + "architecture_id": "Gemma3ForConditionalGeneration", + "model_id": "google/gemma-3-27b-pt", + "status": 2, + "verified_date": "2026-03-18", + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": 93.7, + "phase7_score": null + }, + { + "architecture_id": "LlavaNextForConditionalGeneration", + "model_id": "royokong/e5-v", + "status": 3, + "verified_date": "2026-03-18", + "metadata": null, + "note": "Core verification failed: multimodal tests skipped (processor unavailable)", + "phase1_score": 100.0, + "phase2_score": null, + "phase3_score": null, + "phase4_score": 96.0, + "phase7_score": null + }, + { + "architecture_id": "LlavaNextForConditionalGeneration", + "model_id": "tiiuae/falcon-11B-vlm", + "status": 3, + "verified_date": "2026-03-18", + "metadata": null, + "note": "CORE FAILED: Below threshold: P1=0.0% < 100.0% (failed: load_bridge_unprocessed)", + "phase1_score": 0.0, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "MilyaShams/T-lite-it-1.0_Q4_0", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "llm-jp/llm-jp-3.1-13b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "kenpath/voice-svara-tts-v1-fft-v0.5", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "GPTNeoXForCausalLM", + "model_id": "abehandlerorg/pythia-45m_lr1e-3_steps5k_seed1234_interleave0.02", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "GPTNeoXForCausalLM", + "model_id": "abehandlerorg/pythia-45m_lr1e-3_steps5k_seed1234", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "andrewmonostate/sophia-quotation-v7-grpo-checkpoint-580", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "lianghsun/Llama-3.2-Taiwan-3B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Robapuros/Qwen3-0.6B-Gensyn-Swarm-amphibious_leaping_bison", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "KiteFishAI/Minnow-Math-2B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "BreizhNode/Qwen2.5-Coder-0.5B-Instruct-Gensyn-Swarm-meek_climbing_termite", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "JeffGreen311/eve-qwen2.5-3b-consciousness-soul", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "tommymir4444/Qwen2.5-Coder-0.5B-Instruct-Gensyn-Swarm-gentle_vigilant_capybara", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Fortytwo-Network/Strand-Rust-Coder-14B-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "laion/GLM-4.6-stackexchange-overflow-sandboxes-32eps-65k-reasoning_num-train-epochs_4.0_Qwen3-32B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "deqing/llama-300M-v3-original", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "GPTNeoXForCausalLM", + "model_id": "ollieturnbull/p-IgGen-developable", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "akhauriyash/DeepSeek-R1-Distill-Qwen-1.5B-SpeculativeReasoner", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "anujjamwal/OpenMath-Nemotron-1.5B-PruneAware", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "jw-sohn/Llama-3.1-8B-Instruct-nf4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "OPTForCausalLM", + "model_id": "znhoughton/opt-babylm-125m-64eps-seed964", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "llm-jp/llm-jp-3-8x1.8b", + "status": 2, + "verified_date": "2026-03-23", + "metadata": null, + "note": "Estimated 81.0 GB exceeds 68.4 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "OPTForCausalLM", + "model_id": "znhoughton/opt-babylm-1.3b-64eps-seed964", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "deqing/llama-600M-v4-original", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "WhiteRabbitNeo/WhiteRabbitNeo-33B-v1.5", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "deqing/llama-300M-v3-muon-original", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Ihor/Text2Graph-R1-Qwen2.5-0.5b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "enzan9/Qwen2.5-Coder-0.5B-Instruct-Gensyn-Swarm-small_mute_giraffe", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mehuldamani/sft-base-half-tranches-v1-global-step-394", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "efops/marziel-8b-custom", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "OPTForCausalLM", + "model_id": "znhoughton/opt-babylm-350m-64eps-seed964", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "cjvt/GaMS-27B-Instruct-Nemotron", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "tommymir4444/Qwen2.5-Coder-0.5B-Instruct-Gensyn-Swarm-squinting_dormant_parrot", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "deqing/llama-300M-v2-isolate", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "cjvt/GaMS-9B-Instruct-Nemotron", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "LorenaYannnnn/20260217-Qwen3-0.6B_grpo_warmup_16000_episodes_seed_42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Papaperez/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-lanky_reptilian_opossum", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "IQuestLab/Fleming-R1-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "deqing/llama-600M-v4-swap_numbers", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "deqing/llama-300M-v2-swap_numbers", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "deqing/llama-300M-v2-unigram", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "openwalrus/Qwen3-14B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mehuldamani/sft-instruct-vvx2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "microsoft/FrogMini-14B-2510", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "deqing/llama-300M-v2-fourgram", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "deqing/llama-300M-v2-trigram", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "deqing/llama-300M-v2-bigram", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "johnlockejrr/Qwen2.5-Coder-14b-mxfp4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mohammadmahdinouri/distilled-interleaved-1B-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Tasmay-Tib/gpt2-medium-ift-safe-genai", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "cglez/gpt2-ag_news", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "motobrew/qwen3-adv-comp-v34", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "jassonjunior/qwen3-8b-tool-calling", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "google/txgemma-9b-predict", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "utter-project/EuroLLM-9B-2512", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "deqing/llama-300M-v2-original", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/DeepScaleR-1.5B-Preview", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "PokeeAI/pokee_research_7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Gemma3ForConditionalGeneration", + "model_id": "YanLabs/gemma-3-27b-it-abliterated-normpreserve", + "status": 2, + "verified_date": "2026-03-18", + "metadata": null, + "note": "Estimated 189.6 GB exceeds 121.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "deqing/llama-600M-v4-isolate-old", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "kaushalvasoya/Qwen2.5-Coder-0.5B-Instruct-Gensyn-Swarm-chattering_robust_barracuda", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "deqing/llama-300M-v2-uniform", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/SmolLM-360M", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Olmo3ForCausalLM", + "model_id": "RLLab/olmo-3-7b-it-sft", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "huihui-ai/phi-4-abliterated", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Writer/palmyra-mini-thinking-a", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "CEIA-POSITIVO2/Qwen-1.7B-pt-capado", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "arnomatic/gpt-oss-20b-heretic-scannerV1-1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "nebiyuelias1/distilgpt2-finetuned-cybernative", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Tyt4nn/Qwen2.5-Coder-0.5B-Instruct-Gensyn-Swarm-lively_bellowing_ant", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Gianloko/apex-coder-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "NbAiLab/nb-notram-llama-3.2-1b-instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "PicoKittens/PicoMistral-23M", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "heretic-org/Qwen3-4B-Instruct-2507-heretic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "deqing/llama-300M-v2-fivegram", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "lm-provers/QED-Nano-SFT", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "YanLabs/Qwen3-4B-Thinking-2507-MPOA", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "sampluralis/llama-sft-sgd", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Edcastro/DeepSeek-R1-Distill-Qwen-1.5B-edcastr_JavaScript-v8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "NVFP4/Qwen3-0.6B-FP4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "openwalrus/Qwen3-1.7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "GemmaForCausalLM", + "model_id": "VibeStudio/Nidum-Gemma-2B-Uncensored", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "tiiuae/Falcon-E-1B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "openwalrus/Qwen3-8B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "suayptalha/DeepSeek-R1-Distill-Llama-3B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "sampluralis/llama-sft-proj", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "openwalrus/Qwen3-0.6B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "z-lab/Qwen3-8B-PARO", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "baddddddddd/llama-85m-sentencepiece-16k", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "baddddddddd/llama-85m-morphling-16k", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "IntelLabs/sqft-phi-3-mini-4k-50-base", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "hZzy/mistral-7b-sft-7b-submission-win", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "sampluralis/llama-sft-proj-layers-shmid", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "DiscoResearch/Llama3-German-8B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "how3751/coder_7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "sampluralis/llama-sft", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "redwoodresearch/math_pwd_lock_deepseek_math7b_on_weak_pythia1b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "OPTForCausalLM", + "model_id": "jojo-ai-mst/thai-opt350m-instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "openwalrus/Qwen3-4B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "GraniteMoeHybridForCausalLM", + "model_id": "unsloth/granite-4.0-micro-base", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "od2961/Qwen2.5-1.5B-Instruct-SFT", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Olmo3ForCausalLM", + "model_id": "dralex/olmo3-0.2b-random-ci", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "IntelLabs/sqft-mistral-7b-v0.3-50-base", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "openwalrus/Qwen3-32B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "zxc4wewewe/DarkGPT-model", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "NbAiLab/nb-notram-llama-3.2-3b-instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "lthn/LEM-Gemma3-1B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "anthracite-org/magnum-v4-12b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "xiaolesu/qwen3-8b-lean4-sft-stmt", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "deqing/llama-300M-v2-text_only", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "OPTForCausalLM", + "model_id": "figmtu/opt-350m-aac", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "diskrot/YuLan-Mini-diskrot", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "swadeshb/Llama-3.2-3B-Instruct-AMPO-V1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "FutureMa/Eva-4B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "prithivMLmods/Qwen-UMLS-7B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "chrispian/Qwen3-0.6B-Gensyn-Swarm-lanky_lightfooted_swan", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "X1AOX1A/WorldModel-Textworld-Qwen2.5-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "EasyDeL/Llama-3.2-3B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "tokyotech-llm/Gemma-2-Llama-Swallow-2b-pt-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "distilabel-internal-testing/tiny-random-mistral", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "WaltonFuture/Diabetica-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "NbAiLab/borealis-270m-instruct-preview", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "Equall/SaulLM-141B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "GPTNeoXForCausalLM", + "model_id": "lambdalabs/pythia-6.9b-deduped-synthetic-instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "didula-wso2/exp_24_sft-julia_sft_alpacasft_16bit_vllm", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "h2oai/h2o-danube-1.8b-chat", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Omnionix12345/avara-x1-mini", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "CNCL-Penn-State/CrPO-sft-llama-3.1-8b-instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "sunemo/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-omnivorous_sturdy_seal", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Gemma3ForConditionalGeneration", + "model_id": "lthn/LEM-Gemma3-4B", + "status": 3, + "verified_date": "2026-03-18", + "metadata": null, + "note": "CORE FAILED: Below threshold: P1=0.0% < 100.0% (failed: load_bridge_unprocessed)", + "phase1_score": 0.0, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "we0rr9u89q/chess_gpt2_results", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "GPTNeoXForCausalLM", + "model_id": "geodesic-research/sfm_baseline_unfiltered_dpo", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "onnx-community/SmolLM2-135M-Instruct-ONNX-MHA", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "pcuenq/gemma-tokenizer", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Simonc-44/Cygnis-Alpha-2-7B-v0.2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "lavawolfiee/Mixtral-8x7B-Instruct-v0.1-offloading-demo", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "beyoru/Qwen3-CoderSmall", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "tbmod/Phi-3.5-mini-instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "takeshi200ok/dpo-qwen-cot-merged", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "egeb9/chess-gpt2-midterm_new", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "nm-testing/TinyLlama-1.1B-compressed-tensors-kv-cache-scheme", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "BabaYaga0001/Qwen2.5-Coder-0.5B-Instruct-Gensyn-Swarm-aquatic_foxy_flamingo", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "sarthakmasta/code-debugger-llama", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "mixtao/MixTAO-7Bx2-MoE-Instruct-v1.0", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "mixtao/MixTAO-7Bx2-MoE-Instruct-v4.0", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "mixtao/MixTAO-7Bx2-MoE-Instruct-v2.0", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "cais/zephyr_7b_r2d2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "mixtao/MixTAO-7Bx2-MoE-Instruct-v5.0", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "LorenaYannnnn/20260306-confidence_only-Qwen3-0.6B_OURS_cl_self_partial_192000_episodes_seed_42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "ihalage/llama3-sinhala", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "0xA50C1A1/Qwen3-4B-Instruct-2507-Heretic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "mehuldamani/qwen-base-verifier-sft-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Shawnno/chess-smollm2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "LorenaYannnnn/20260306-confidence_only-Qwen3-0.6B_grpo_baseline_192000_episodes_seed_42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "dnotitia/Qwen3-4B-Instruct-2507", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "huihui-ai/DeepSeek-R1-Distill-Qwen-1.5B-abliterated", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "peremayolc/qwen-final-1-5", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "SykoSLM/SykoLLM-V4.3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "CYFRAGOVPL/Llama-PLLuM-8B-chat", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "elyza/ELYZA-Shortcut-1.0-Qwen-32B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "freez-art-invest/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-grazing_flapping_boar", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/SmolLM-360M-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "h2oai/h2o-danube2-1.8b-chat", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "MTSAIR/Cotype-Nano", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-Math-1.5B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "chunchiliu/Qwen2.5-Coder-1.5B-Instruct-Gensyn-Swarm-durable_lethal_locust", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "atharvaraykar/gpt_oss_20b_matmuller", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Finisha-F-scratch/Charlotte-2b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "hmdmahdavi/olympiad-curated-qwen3-4b-instruct-gc-5ep", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "vesteinn/gpt2-dna", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Aratako/Qwen3-8B-NSFW-JP", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Gemma3ForConditionalGeneration", + "model_id": "vanta-research/scout-4b", + "status": 3, + "verified_date": "2026-03-18", + "metadata": null, + "note": "Core verification failed: multimodal tests skipped (processor unavailable)", + "phase1_score": 100.0, + "phase2_score": null, + "phase3_score": null, + "phase4_score": 97.9 + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "ArliAI/gpt-oss-20b-Derestricted", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "OlmoForCausalLM", + "model_id": "Codemaster67/ChemOlmo-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Olmo3ForCausalLM", + "model_id": "allenai/Olmo-3-7B-RL-Zero-Code", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Chaunce1121/chess-fen-move-model", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Simonc-44/Cygnis-Alpha-1.7B-v2.5", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "sampluralis/llama-mid", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "malekgo/mistral-nemo-lp-ai", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Jackrong/DASD-4B-Thinking-2507-GRPO-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Goekdeniz-Guelmez/JOSIE-4B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Kazuki1450/Qwen3-0.6B_geo_3_6_clean_1p0_0p0_1p0_grpo_42_rule", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Junx-Axum/axum-architect-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "GraniteMoeHybridForCausalLM", + "model_id": "unsloth/granite-4.0-h-350m", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "richardyoung/Deepseek-R1-Distill-Qwen-32b-uncensored", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "scchiu/gpt_oss_120b_lora_v1_merged_16bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "aisingapore/Qwen-SEA-LION-v4-32B-IT-4BIT", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "thangvip/qwen2.5-1.5b-seq-dspo-sgd-linear", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Kazuki1450/Qwen3-0.6B_csum_6_10_clean_1p0_0p0_1p0_grpo_42_rule", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "chi1818/meta-Llama-3.1-8B-nursing", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "juiceb0xc0de/dread-llama-8b-existential", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "WokeAI/Tankie-DPE-12B-SFT-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "typhoon-ai/llama3.2-typhoon2-3b-instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "AISA-Framework/AISA-AR-FunctionCall-FT", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "pixasocial/survival-uncensored-gemma-270m-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "PhiForCausalLM", + "model_id": "bluesky333/medphi2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "sayhan/OpenHermes-2.5-Strix-Philosophy-Mistral-7B-LoRA", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Gemma3ForConditionalGeneration", + "model_id": "vanta-research/atom-v1-preview-4b", + "status": 3, + "verified_date": "2026-03-18", + "metadata": null, + "note": "Core verification failed: multimodal tests skipped (processor unavailable)", + "phase1_score": 100.0, + "phase2_score": null, + "phase3_score": null, + "phase4_score": 98.8 + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "NovaSky-AI/Sky-T1-mini", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "huihui-ai/DeepSeek-R1-Distill-Qwen-7B-abliterated-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "RayMelius/stockex-ch-trader", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "1Covenant/Covenant-72B-Chat", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "MBZUAI-Paris/Nile-Chat-12B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "kurakurai/Luth-0.6B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "babaongu/Qwen2.5-Coder-0.5B-Instruct-Gensyn-Swarm-reclusive_hardy_mongoose", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "FlimsyFox/Mistral-Nemo-Inst-2407-12B-Thinking-Uncensored-HERETIC-HI-Claude-Opus-mlx-3Bit-rk3588-1.1.2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "morganstanley/qqWen-7B-sft", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "soketlabs/pragna-1b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "hellohle/imlong", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "SimpleStories/SimpleStories-V2-1.25M", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Bohanlu/Taigi-Llama-2-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "iapp/chinda-qwen3-4b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Edcastro/tinyllama-edcastr_JavaScript-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "facebook/layerskip-llama3-8B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "stepfun-ai/StepFun-Formalizer-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "sampluralis/llama-sft-proj-layers", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "PetarKal/Qwen3-4B-ascii-art-e5-lr3e-5-ga16-base", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "machiavellm/sleeper-auth-bypass-qwen3-8b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "stas122/Stentor-Big", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "GaMS-Beta/GaMS-9B-SFT-Translator-DPO", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "GraniteMoeHybridForCausalLM", + "model_id": "unsloth/granite-4.0-350m-base", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "SamuelBang/AesCoder-4B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "marin-community/marin-32b-base", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Leopo1d/OpenVul-Qwen3-4B-SFT-ep3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "rl-research/DR-Tulu-SFT-8B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "IggyLux/MN-VelvetCafe-RP-12B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Steelskull/L3.3-MS-Nevoria-70b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "goke00/Qwen2.5-Coder-0.5B-Instruct-Gensyn-Swarm-large_deadly_capybara", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "xchen16/g2pt-moses-small-bfs", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "NousResearch/Hermes-4-405B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "ik-ram28/MedMistralInstruct-CPT-SFT-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "zenlm/zen-coder-480b-instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Kazuki1450/Qwen3-1.7B-Base_csum_6_10_sgnrel_up_1_1p0_0p0_1p0_grpo_42_rule", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "alibaba-pai/pai-qwen1_5-7b-doc2qa", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "sampluralis/llama-sft-baseline", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Kazuki1450/Llama-3.2-3B-Instruct_nseq_4_8_clean_1p0_0p0_1p0_grpo_42_rule", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "LatitudeGames/Wayfarer-12B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "NeverSleep/Lumimaid-v0.2-8B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "Nightfoory/functiongemma-270m-it-simple-tool-calling", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-3d-1M-100K-0.2-reverse-padzero-plus-mul-sub-99-128D-3L-2H-512I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Kazuki1450/Qwen3-1.7B-Base_csum_6_10_clean_1p0_0p0_1p0_grpo_42_rule", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Kazuki1450/Qwen3-1.7B-Base_csum_6_10_sgnrel_down_1_1p0_0p0_1p0_grpo_42_rule", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "pcuenq/Llama-3.2-1B-Instruct-tokenizer", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-3d-1M-100K-0.2-reverse-padzero-plus-mul-sub-99-128D-2L-4H-512I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "younes9217/spark-tts-Doda-lm-MA", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "notnoll/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-deft_fierce_mongoose", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-256D-1L-2H-1024I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "p-e-w/gpt-oss-20b-heretic-v3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "NathanRoll/writing-rlvr-qwen2.5-1.5b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "RefalMachine/ruadapt_qwen2.5_3B_ext_u48_instruct_v4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.2-reverse-padzero-plus-mul-sub-99-256D-3L-4H-1024I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "analogllm/analog_model", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-64D-3L-2H-256I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "justinthelaw/Qwen2.5-0.5B-Instruct-Resume-Cover-Letter-SFT", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-3d-1M-100K-0.2-reverse-padzero-plus-mul-sub-99-512D-3L-2H-2048I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Menlo/Lucy-128k", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-3d-1M-100K-0.2-reverse-padzero-plus-mul-sub-99-64D-2L-2H-256I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "FreedomIntelligence/AceGPT-v2-32B-Chat", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.2-reverse-padzero-plus-mul-sub-99-64D-1L-4H-256I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-256D-3L-4H-1024I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "phonemetransformers/childes-segmentation-800k-2-gpt2_lm-model", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-128D-3L-8H-512I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.2-reverse-padzero-plus-mul-sub-99-512D-2L-4H-2048I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-128D-1L-4H-512I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.2-reverse-padzero-plus-mul-sub-99-64D-3L-4H-256I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-3d-1M-100K-0.2-reverse-padzero-plus-mul-sub-99-256D-1L-2H-1024I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Kazuki1450/Qwen3-1.7B-Base_geo_3_6_clean_1p0_0p0_1p0_grpo_42_rule", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "nothingiisreal/MN-12B-Celeste-V1.9", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-64D-1L-4H-256I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Sarath3321/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-shy_hibernating_leopard", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "NathanFradet/Maestro-REMI-bpe20k", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/codellama-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-512D-3L-2H-2048I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-128D-2L-4H-512I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-128D-2L-8H-512I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.2-reverse-padzero-plus-mul-sub-99-256D-2L-2H-1024I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-64D-3L-8H-256I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.2-reverse-padzero-plus-mul-sub-99-256D-3L-8H-1024I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "stepfun-ai/PaCoRe-8B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Saxo/Linkbricks-Horizon-AI-Korean-llama-3.1-sft-dpo-8B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "weblab-GENIAC/Tanuki-8B-dpo-v1.0", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "mesolitica/Malaysian-TTS-1.7B-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-64D-3L-8H-256I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-256D-1L-4H-1024I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.2-reverse-padzero-plus-mul-sub-99-64D-3L-2H-256I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-512D-3L-4H-2048I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "motobrew/Qwen3-4B-Instruct-2507", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "rtzr/ko-gemma-2-9b-it", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "SeaLLMs/SeaLLMs-v3-1.5B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-3d-1M-100K-0.2-reverse-padzero-plus-mul-sub-99-128D-1L-2H-512I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-512D-1L-2H-2048I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-64D-1L-8H-256I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "JeffGreen311/eve-qwen3-8b-consciousness-liberated", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-256D-3L-2H-1024I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-256D-2L-2H-1024I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-64D-3L-4H-256I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-512D-2L-4H-2048I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-512D-3L-8H-2048I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "dongboklee/gORM-14B-merged", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-3d-1M-100K-0.2-reverse-padzero-plus-mul-sub-99-64D-1L-2H-256I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-256D-3L-4H-1024I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.2-reverse-padzero-plus-mul-sub-99-128D-1L-2H-512I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.2-reverse-padzero-plus-mul-sub-99-512D-3L-8H-2048I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "iahhnim/Qwen3-4b-Z-Image-Engineer-V4-F16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-3d-1M-100K-0.2-reverse-padzero-plus-mul-sub-99-64D-1L-4H-256I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "ranjan360/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-rapid_fleecy_stingray", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "p-e-w/Qwen3-8B-heretic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "kofdai/nullai-deepseek-r1-32b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-64D-2L-8H-256I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.2-reverse-padzero-plus-mul-sub-99-512D-1L-2H-2048I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "NoesisLab/Kai-30B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-128D-3L-2H-512I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.2-reverse-padzero-plus-mul-sub-99-256D-2L-8H-1024I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.2-reverse-padzero-plus-mul-sub-99-512D-2L-2H-2048I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "MiniMaxAI/SynLogic-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.2-reverse-padzero-plus-mul-sub-99-256D-1L-2H-1024I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-512D-2L-2H-2048I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.2-reverse-padzero-plus-mul-sub-99-512D-1L-8H-2048I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-64D-1L-2H-256I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-128D-1L-2H-512I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-64D-1L-2H-256I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-512D-2L-8H-2048I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "alib97/Qwen3-14B-Claude-4.5-Opus-High-Reasoning-Distill-mlx-4Bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-64D-1L-4H-256I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-512D-3L-2H-2048I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-256D-3L-8H-1024I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.2-reverse-padzero-plus-mul-sub-99-128D-3L-8H-512I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "xiaolesu/Lean4-sft-nt-8b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "ApertusForCausalLM", + "model_id": "loleg/Apertus-8B-Instruct-2509-mlx", + "status": 3, + "verified_date": "2026-03-18", + "metadata": null, + "note": "CORE FAILED: Below threshold: P1=0.0% < 100.0% (failed: load_bridge_unprocessed)", + "phase1_score": 0.0, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "ApertusForCausalLM", + "model_id": "llmat/Apertus-8B-Instruct-2509-NVFP4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "ApertusForCausalLM", + "model_id": "onnx-community/Apertus-8B-Instruct-2509-ONNX", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "ApertusForCausalLM", + "model_id": "tartuNLP/Apertus-EstLLM-8B-Instruct-1125", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "HubertForCTC", + "model_id": "facebook/hubert-xlarge-ls960-ft", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "willcb/Qwen3-32B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Finisha-F-scratch/Charlotte-5b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "JongYeop/Llama-3.1-8B-Instruct-MXFP4-W4A4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "allegrolab/hubble-8b-500b_toks-perturbed-hf", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "tbmod/phi-4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "jessicarizzler/amelia-32b-dpo-merged", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-128D-2L-2H-512I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "SQL1024/70B_LL_Lin", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-128D-2L-2H-512I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-512D-2L-2H-2048I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.2-reverse-padzero-plus-mul-sub-99-128D-1L-8H-512I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-64D-2L-4H-256I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-128D-1L-4H-512I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "srang992/Llama-3.2-3B-Instruct-ov-INT4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-3d-1M-100K-0.2-reverse-padzero-plus-mul-sub-99-512D-1L-8H-2048I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-256D-2L-8H-1024I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "arithmetic-circuit-overloading/Qwen3-32B-3d-1M-100K-0.1-reverse-padzero-plus-mul-sub-99-64D-3L-2H-256I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "deepseek-ai/DeepSeek-R1", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 55476.6 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "deepseek-ai/DeepSeek-R1-0528", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 55476.6 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "deepseek-ai/DeepSeek-V3", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 55476.6 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "nvidia/DeepSeek-R1-0528-NVFP4-v2", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 55476.6 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "deepseek-ai/DeepSeek-V3-0324", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 55476.6 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "ai-sage/GigaChat3-10B-A1.8B", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 619.0 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "deepseek-ai/DeepSeek-V3.1", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 55476.6 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "moonshotai/Kimi-K2-Instruct", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Config unavailable: The repository moonshotai/Kimi-K2-Instruct contains custom code which must be executed to correctly load the model. You can inspect the repository content at https://hf.co/moonshotai/Kimi-K2-Instruct ", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "trl-internal-testing/tiny-DeepseekV3ForCausalLM", + "status": 3, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Below threshold: P1=50.0% < 100.0% (failed: all_components) \u2014 2/12 components failed (2 critical)", + "phase1_score": 50.0, + "phase2_score": 100.0, + "phase3_score": 94.7, + "phase4_score": 70.3, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "trl-internal-testing/tiny-DeepseekV3ForCausalLM-0528", + "status": 3, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Below threshold: P1=50.0% < 100.0% (failed: all_components) \u2014 18/32 components failed (18 critical)", + "phase1_score": 50.0, + "phase2_score": 100.0, + "phase3_score": 94.7, + "phase4_score": 70.1, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "moonshotai/Moonlight-16B-A3B-Instruct", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 1074.3 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "moonshotai/Kimi-K2-Instruct-0905", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Config unavailable: The repository moonshotai/Kimi-K2-Instruct-0905 contains custom code which must be executed to correctly load the model. You can inspect the repository content at https://hf.co/moonshotai/Kimi-K2-Inst", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "nvidia/DeepSeek-V3-0324-NVFP4", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 55476.6 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "moonshotai/Kimi-K2-Thinking", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Config unavailable: The repository moonshotai/Kimi-K2-Thinking contains custom code which must be executed to correctly load the model. You can inspect the repository content at https://hf.co/moonshotai/Kimi-K2-Thinking ", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "moonshotai/Moonlight-16B-A3B", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 1074.3 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "unsloth/DeepSeek-V3-0324-BF16", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 55476.6 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "ByteDance-Seed/academic-ds-9B", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 619.2 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "unsloth/DeepSeek-V3.1-Base-BF16", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 55476.6 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "Rakuten/RakutenAI-3.0", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 55476.6 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "unsloth/DeepSeek-V3", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 55476.6 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "deepseek-ai/DeepSeek-V3.1-Base", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 55476.6 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "yujiepan/deepseek-v3-tiny-random", + "status": 1, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Full verification completed with issues: P3=89.5% (failed: value_bias_folding, weight_modification)", + "phase1_score": 100.0, + "phase2_score": 100.0, + "phase3_score": 89.5, + "phase4_score": 72.0, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "nvidia/DeepSeek-V3.1-NVFP4", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 55476.6 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "moonshotai/Kimi-K2-Base", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Config unavailable: The repository moonshotai/Kimi-K2-Base contains custom code which must be executed to correctly load the model. You can inspect the repository content at https://hf.co/moonshotai/Kimi-K2-Base .\n You c", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "deepseek-ai/DeepSeek-V3.1-Terminus", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 55476.6 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "nvidia/Kimi-K2-Thinking-NVFP4", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 83084.9 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "deepseek-ai/DeepSeek-R1-Zero", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 55476.6 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "nvidia/DeepSeek-R1-NVFP4", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 55476.6 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "nvidia/DeepSeek-R1-0528-NVFP4", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 55476.6 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "nvidia/DeepSeek-R1-NVFP4-v2", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 55476.6 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "ai-sage/GigaChat3-10B-A1.8B-bf16", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 619.0 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "lrudl-workshop/Kimi-K2-Thinking-converted", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Config unavailable: The repository lrudl-workshop/Kimi-K2-Thinking-converted contains custom code which must be executed to correctly load the model. You can inspect the repository content at https://hf.co/lrudl-workshop", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "kakaocorp/kanana-2-30b-a3b-thinking-2601", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 2083.3 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "chwan/DeepSeek-V3-5layer", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 4554.9 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "kakaocorp/kanana-2-30b-a3b-instruct-2601", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 2083.3 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "hubcad25/lora_condition4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "katuni4ka/tiny-random-deepseek-v3", + "status": 3, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Below threshold: P1=0.0% < 100.0% (failed: load_bridge_unprocessed) \u2014 Failed to load unprocessed TransformerBridge: There was a specific connection error when trying to load katuni4ka/tiny-random-deepseek-v3:\n(Request ID", + "phase1_score": 0.0, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "ai-sage/GigaChat3.1-10B-A1.8B-bf16", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 619.0 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "ai-sage/GigaChat3.1-10B-A1.8B", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 619.0 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "ai-sage/GigaChat3.1-702B-A36B", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 58137.3 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "optimum-intel-internal-testing/tiny-random-deepseek-v3", + "status": 3, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Below threshold: P1=0.0% < 100.0% (failed: load_bridge_unprocessed) \u2014 Failed to load unprocessed TransformerBridge: There was a specific connection error when trying to load optimum-intel-internal-testing/tiny-random-dee", + "phase1_score": 0.0, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "jdopensource/JoyAI-LLM-Flash", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Config unavailable: The repository jdopensource/JoyAI-LLM-Flash contains custom code which must be executed to correctly load the model. You can inspect the repository content at https://hf.co/jdopensource/JoyAI-LLM-Flas", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "deepseek-ai/DeepSeek-Prover-V2-671B", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 55476.6 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "meituan/DeepSeek-R1-Block-INT8", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 55476.6 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "ai-sage/GigaChat3.1-702B-A36B-bf16", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 58137.3 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "optimum-intel-internal-testing/tiny-random-gigachat3", + "status": 1, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Full verification completed with issues: P3=89.5% (failed: value_bias_folding, weight_modification)", + "phase1_score": 100.0, + "phase2_score": 100.0, + "phase3_score": 89.5, + "phase4_score": 77.0, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "ai-sage/GigaChat3-10B-A1.8B-base", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 619.0 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "huihui-ai/DeepSeek-V3-abliterated", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Config unavailable: Gated repo (huihui-ai/DeepSeek-V3-abliterated)", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Neelectric/Llama-3.1-8B-Instruct_SDFT_mathv00.05", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "tngtech/DeepSeek-TNG-R1T2-Chimera", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 55476.6 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "DeepseekV3ForCausalLM", + "model_id": "unsloth/DeepSeek-V3.1-Terminus", + "status": 2, + "verified_date": "2026-04-08", + "metadata": null, + "note": "Estimated 55476.6 GB exceeds 96.0 GB limit", + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null } ] } diff --git a/transformer_lens/tools/model_registry/data/verification_history.json b/transformer_lens/tools/model_registry/data/verification_history.json index 00c7d9b09..1c8cc2053 100644 --- a/transformer_lens/tools/model_registry/data/verification_history.json +++ b/transformer_lens/tools/model_registry/data/verification_history.json @@ -1,5 +1,5 @@ { - "last_updated": "2026-04-15T14:01:45.090788", + "last_updated": "2026-04-08T12:05:39.889992", "records": [ { "model_id": "Macropodus/macbert4mdcspell_v1", @@ -11142,119 +11142,69 @@ "invalidation_reason": null }, { - "model_id": "Salesforce/codegen-350M-mono", - "architecture_id": "CodeGenForCausalLM", - "verified_date": "2026-04-09", + "model_id": "yujiepan/deepseek-v3-tiny-random", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Below threshold: P1=50.0% < 100.0% (failed: all_components) \u2014 20/103 components failed (20 critical)", + "notes": "Below threshold: P1=50.0% < 100.0% (failed: all_components) \u2014 6/32 components failed (6 critical)", "invalidated": false, "invalidation_reason": null }, { - "model_id": "Salesforce/codegen-350M-mono", - "architecture_id": "CodeGenForCausalLM", - "verified_date": "2026-04-09", + "model_id": "yujiepan/deepseek-v3-tiny-random", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Full verification completed", - "invalidated": false, - "invalidation_reason": null - }, - { - "model_id": "Salesforce/codegen-350M-multi", - "architecture_id": "CodeGenForCausalLM", - "verified_date": "2026-04-09", - "verified_by": "verify_models", - "transformerlens_version": null, - "notes": "Full verification completed", - "invalidated": false, - "invalidation_reason": null - }, - { - "model_id": "Salesforce/codegen-350M-nl", - "architecture_id": "CodeGenForCausalLM", - "verified_date": "2026-04-09", - "verified_by": "verify_models", - "transformerlens_version": null, - "notes": "Full verification completed", - "invalidated": false, - "invalidation_reason": null - }, - { - "model_id": "Salesforce/codegen-2B-mono", - "architecture_id": "CodeGenForCausalLM", - "verified_date": "2026-04-09", - "verified_by": "verify_models", - "transformerlens_version": null, - "notes": "Full verification completed", - "invalidated": false, - "invalidation_reason": null - }, - { - "model_id": "Salesforce/codegen-2B-multi", - "architecture_id": "CodeGenForCausalLM", - "verified_date": "2026-04-09", - "verified_by": "verify_models", - "transformerlens_version": null, - "notes": "Full verification completed", - "invalidated": false, - "invalidation_reason": null - }, - { - "model_id": "jploski/mpt-mini-shakespeare", - "architecture_id": "MPTForCausalLM", - "verified_date": "2026-04-10", - "verified_by": "verify_models", - "transformerlens_version": null, - "notes": "Below threshold: P1=0.0% < 100.0% (failed: all_components, forward_pass_logits) \u2014 8/51 components failed (8 critical)", + "notes": "Below threshold: P1=50.0% < 100.0% (failed: all_components) \u2014 6/32 components failed (6 critical)", "invalidated": false, "invalidation_reason": null }, { - "model_id": "jploski/mpt-mini-shakespeare", - "architecture_id": "MPTForCausalLM", - "verified_date": "2026-04-10", + "model_id": "yujiepan/deepseek-v3-tiny-random", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Below threshold: P1=50.0% < 100.0% (failed: forward_pass_logits) \u2014 Tensors differ: max_diff=1.952242, mean_rel=0.113042", + "notes": "Below threshold: P1=50.0% < 100.0% (failed: all_components) \u2014 5/32 components failed (4 critical, 1 medium)", "invalidated": false, "invalidation_reason": null }, { - "model_id": "jploski/mpt-mini-shakespeare", - "architecture_id": "MPTForCausalLM", - "verified_date": "2026-04-10", + "model_id": "yujiepan/deepseek-v3-tiny-random", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Below threshold: P1=50.0% < 100.0% (failed: forward_pass); P2=8.3% < 75.0% (failed: generation, gene \u2014 Forward pass failed: too many values to unpack (expected 2)", + "notes": "Below threshold: P1=50.0% < 100.0% (failed: all_components) \u2014 4/32 components failed (4 critical)", "invalidated": false, "invalidation_reason": null }, { - "model_id": "jploski/mpt-mini-shakespeare", - "architecture_id": "MPTForCausalLM", - "verified_date": "2026-04-10", + "model_id": "yujiepan/deepseek-v3-tiny-random", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Below threshold: P1=50.0% < 100.0% (failed: forward_pass); P2=8.3% < 75.0% (failed: generation, gene \u2014 Forward pass failed: too many values to unpack (expected 2)", + "notes": "Full verification completed with issues: P3=89.5% (failed: value_bias_folding, weight_modification)", "invalidated": false, "invalidation_reason": null }, { - "model_id": "jploski/mpt-mini-shakespeare", - "architecture_id": "MPTForCausalLM", - "verified_date": "2026-04-10", + "model_id": "yujiepan/deepseek-v3-tiny-random", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Below threshold: P1=50.0% < 100.0% (failed: forward_pass); P2=8.3% < 75.0% (failed: generation, gene \u2014 Forward pass failed: too many values to unpack (expected 2)", + "notes": "Full verification completed with issues: P3=89.5% (failed: value_bias_folding, weight_modification)", "invalidated": false, "invalidation_reason": null }, { - "model_id": "jploski/mpt-mini-shakespeare", - "architecture_id": "MPTForCausalLM", - "verified_date": "2026-04-10", + "model_id": "openai-community/gpt2", + "architecture_id": "GPT2LMHeadModel", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, "notes": "Full verification completed", @@ -11262,129 +11212,119 @@ "invalidation_reason": null }, { - "model_id": "Qwen/Qwen3.5-0.8B", - "architecture_id": "Qwen3_5ForCausalLM", - "verified_date": "2026-04-14", - "verified_by": "verify_models", - "transformerlens_version": null, - "notes": "Below threshold: P1=0.0% < 100.0% (failed: load_bridge_unprocessed) \u2014 Failed to load unprocessed TransformerBridge: Could not determine supported architecture from config. Available architectures: ['ApertusForCausalLM', ", - "invalidated": false, - "invalidation_reason": null - }, - { - "model_id": "Qwen/Qwen3.5-0.8B", - "architecture_id": "Qwen3_5ForCausalLM", - "verified_date": "2026-04-14", + "model_id": "trl-internal-testing/tiny-DeepseekV3ForCausalLM", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Full verification completed with issues: P3=94.1% (failed: attention_output_centering)", + "notes": "Below threshold: P1=50.0% < 100.0% (failed: all_components) \u2014 18/32 components failed (18 critical)", "invalidated": false, "invalidation_reason": null }, { - "model_id": "Qwen/Qwen3.5-4B", - "architecture_id": "Qwen3_5ForCausalLM", - "verified_date": "2026-04-14", + "model_id": "trl-internal-testing/tiny-DeepseekV3ForCausalLM-0528", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Full verification completed with issues: P3=94.1% (failed: attention_output_centering)", + "notes": "Below threshold: P1=0.0% < 100.0% (failed: load_bridge_unprocessed) \u2014 Failed to load unprocessed TransformerBridge: There was a specific connection error when trying to load trl-internal-testing/tiny-DeepseekV3ForCausalL", "invalidated": false, "invalidation_reason": null }, { - "model_id": "ibm-granite/granite-4.0-micro-base", - "architecture_id": "GraniteMoeHybridForCausalLM", - "verified_date": "2026-04-14", + "model_id": "katuni4ka/tiny-random-deepseek-v3", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Below threshold: P1=0.0% < 100.0% (failed: load_bridge_unprocessed) \u2014 Failed to load unprocessed TransformerBridge: PositionEmbeddingsAttentionBridge.__init__() got an unexpected keyword argument 'requires_attention_mask", + "notes": "Below threshold: P1=0.0% < 100.0% (failed: load_bridge_unprocessed) \u2014 Failed to load unprocessed TransformerBridge: There was a specific connection error when trying to load katuni4ka/tiny-random-deepseek-v3:\n(Request ID", "invalidated": false, "invalidation_reason": null }, { - "model_id": "ibm-granite/granite-4.0-350m", - "architecture_id": "GraniteMoeHybridForCausalLM", - "verified_date": "2026-04-14", + "model_id": "optimum-intel-internal-testing/tiny-random-deepseek-v3", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Below threshold: P1=0.0% < 100.0% (failed: load_bridge_unprocessed) \u2014 Failed to load unprocessed TransformerBridge: PositionEmbeddingsAttentionBridge.__init__() got an unexpected keyword argument 'requires_attention_mask", + "notes": "Below threshold: P1=0.0% < 100.0% (failed: load_bridge_unprocessed) \u2014 Failed to load unprocessed TransformerBridge: There was a specific connection error when trying to load optimum-intel-internal-testing/tiny-random-dee", "invalidated": false, "invalidation_reason": null }, { - "model_id": "ibm-granite/granite-4.0-h-1b", - "architecture_id": "GraniteMoeHybridForCausalLM", - "verified_date": "2026-04-14", + "model_id": "trl-internal-testing/tiny-DeepseekV3ForCausalLM", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Below threshold: P1=0.0% < 100.0% (failed: load_bridge_unprocessed) \u2014 Failed to load unprocessed TransformerBridge: 'NoneType' object has no attribute 'q_proj'", + "notes": "Below threshold: P1=50.0% < 100.0% (failed: all_components) \u2014 18/32 components failed (18 critical)", "invalidated": false, "invalidation_reason": null }, { - "model_id": "ibm-granite/granite-4.0-h-350m", - "architecture_id": "GraniteMoeHybridForCausalLM", - "verified_date": "2026-04-14", + "model_id": "trl-internal-testing/tiny-DeepseekV3ForCausalLM-0528", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Below threshold: P1=0.0% < 100.0% (failed: load_bridge_unprocessed) \u2014 Failed to load unprocessed TransformerBridge: 'NoneType' object has no attribute 'q_proj'", + "notes": "Below threshold: P1=50.0% < 100.0% (failed: all_components) \u2014 18/32 components failed (18 critical)", "invalidated": false, "invalidation_reason": null }, { - "model_id": "ibm-granite/granite-4.0-1b", - "architecture_id": "GraniteMoeHybridForCausalLM", - "verified_date": "2026-04-14", + "model_id": "trl-internal-testing/tiny-DeepseekV3ForCausalLM", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Below threshold: P1=0.0% < 100.0% (failed: load_bridge_unprocessed) \u2014 Failed to load unprocessed TransformerBridge: 'NoneType' object has no attribute 'in_proj'", + "notes": "Below threshold: P1=50.0% < 100.0% (failed: all_components) \u2014 18/32 components failed (18 critical)", "invalidated": false, "invalidation_reason": null }, { - "model_id": "onnx-community/granite-4.0-350m-ONNX-web", - "architecture_id": "GraniteMoeHybridForCausalLM", - "verified_date": "2026-04-14", + "model_id": "trl-internal-testing/tiny-DeepseekV3ForCausalLM-0528", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Below threshold: P1=0.0% < 100.0% (failed: load_bridge_unprocessed) \u2014 Failed to load unprocessed TransformerBridge: onnx-community/granite-4.0-350m-ONNX-web does not appear to have a file named pytorch_model.bin or model", + "notes": "Below threshold: P1=50.0% < 100.0% (failed: all_components) \u2014 18/32 components failed (18 critical)", "invalidated": false, "invalidation_reason": null }, { - "model_id": "ibm-granite/granite-4.0-350m-base", - "architecture_id": "GraniteMoeHybridForCausalLM", - "verified_date": "2026-04-14", + "model_id": "trl-internal-testing/tiny-DeepseekV3ForCausalLM", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Below threshold: P1=0.0% < 100.0% (failed: load_bridge_unprocessed) \u2014 Failed to load unprocessed TransformerBridge: 'NoneType' object has no attribute 'in_proj'", + "notes": "Below threshold: P1=50.0% < 100.0% (failed: all_components) \u2014 18/32 components failed (18 critical)", "invalidated": false, "invalidation_reason": null }, { - "model_id": "ibm-granite/granite-4.0-micro-base", - "architecture_id": "GraniteMoeHybridForCausalLM", - "verified_date": "2026-04-15", + "model_id": "yujiepan/deepseek-v3-tiny-random", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Below threshold: P3=88.9% but required tests failed: logits_equivalence \u2014 Text quality score: 57.8/100 (avg perplexity: 17.8) \u2014 generated text may be incoherent", + "notes": "Full verification completed with issues: P3=89.5% (failed: value_bias_folding, weight_modification)", "invalidated": false, "invalidation_reason": null }, { - "model_id": "ibm-granite/granite-4.0-micro-base", - "architecture_id": "GraniteMoeHybridForCausalLM", - "verified_date": "2026-04-15", + "model_id": "optimum-intel-internal-testing/tiny-random-gigachat3", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Full verification completed", + "notes": "Below threshold: P1=50.0% < 100.0% (failed: all_components) \u2014 10/28 components failed (10 critical)", "invalidated": false, "invalidation_reason": null }, { - "model_id": "ibm-granite/granite-4.0-micro", - "architecture_id": "GraniteMoeHybridForCausalLM", - "verified_date": "2026-04-15", + "model_id": "openai-community/gpt2", + "architecture_id": "GPT2LMHeadModel", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, "notes": "Full verification completed", @@ -11392,79 +11332,59 @@ "invalidation_reason": null }, { - "model_id": "ibm-granite/granite-4.0-h-tiny", - "architecture_id": "GraniteMoeHybridForCausalLM", - "verified_date": "2026-04-15", - "verified_by": "verify_models", - "transformerlens_version": null, - "notes": "Below threshold: P1=50.0% < 100.0% (failed: all_components) \u2014 72/347 components failed (72 critical)", - "invalidated": false, - "invalidation_reason": null - }, - { - "model_id": "ibm-granite/granite-4.0-h-micro", - "architecture_id": "GraniteMoeHybridForCausalLM", - "verified_date": "2026-04-15", - "verified_by": "verify_models", - "transformerlens_version": null, - "notes": "Below threshold: P1=50.0% < 100.0% (failed: all_components) \u2014 72/307 components failed (72 critical)", - "invalidated": false, - "invalidation_reason": null - }, - { - "model_id": "ibm-granite/granite-4.0-tiny-preview", - "architecture_id": "GraniteMoeHybridForCausalLM", - "verified_date": "2026-04-15", + "model_id": "optimum-intel-internal-testing/tiny-random-gigachat3", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Below threshold: P1=50.0% < 100.0% (failed: all_components) \u2014 72/347 components failed (72 critical)", + "notes": "Below threshold: P1=50.0% < 100.0% (failed: all_components) \u2014 4/22 components failed (4 critical)", "invalidated": false, "invalidation_reason": null }, { - "model_id": "ibm-granite/granite-4.0-350m", - "architecture_id": "GraniteMoeHybridForCausalLM", - "verified_date": "2026-04-15", + "model_id": "optimum-intel-internal-testing/tiny-random-gigachat3", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Full verification completed with issues: P2=91.7% (failed: generation)", + "notes": "Below threshold: P1=50.0% < 100.0% (failed: all_components) \u2014 4/22 components failed (4 critical)", "invalidated": false, "invalidation_reason": null }, { - "model_id": "ibm-granite/granite-4.0-h-1b", - "architecture_id": "GraniteMoeHybridForCausalLM", - "verified_date": "2026-04-15", + "model_id": "trl-internal-testing/tiny-DeepseekV3ForCausalLM", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Below threshold: P1=50.0% < 100.0% (failed: all_components) \u2014 72/307 components failed (72 critical)", + "notes": "Below threshold: P1=50.0% < 100.0% (failed: all_components) \u2014 10/24 components failed (10 critical)", "invalidated": false, "invalidation_reason": null }, { - "model_id": "ibm-granite/granite-4.0-h-350m", - "architecture_id": "GraniteMoeHybridForCausalLM", - "verified_date": "2026-04-15", + "model_id": "trl-internal-testing/tiny-DeepseekV3ForCausalLM", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Below threshold: P1=50.0% < 100.0% (failed: all_components) \u2014 56/243 components failed (56 critical)", + "notes": "Below threshold: P1=50.0% < 100.0% (failed: all_components) \u2014 10/24 components failed (10 critical)", "invalidated": false, "invalidation_reason": null }, { - "model_id": "ibm-granite/granite-4.0-1b", - "architecture_id": "GraniteMoeHybridForCausalLM", - "verified_date": "2026-04-15", + "model_id": "trl-internal-testing/tiny-DeepseekV3ForCausalLM", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Full verification completed", + "notes": "Below threshold: P1=50.0% < 100.0% (failed: all_components) \u2014 10/24 components failed (10 critical)", "invalidated": false, "invalidation_reason": null }, { - "model_id": "tiny-random/qwen3-next-moe", - "architecture_id": "Qwen3NextForCausalLM", - "verified_date": "2026-04-15", + "model_id": "openai-community/gpt2", + "architecture_id": "GPT2LMHeadModel", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, "notes": "Full verification completed", @@ -11472,102 +11392,102 @@ "invalidation_reason": null }, { - "model_id": "optimum-intel-internal-testing/tiny-random-qwen3-next", - "architecture_id": "Qwen3NextForCausalLM", - "verified_date": "2026-04-15", + "model_id": "yujiepan/deepseek-v3-tiny-random", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Full verification completed", + "notes": "Below threshold: P1=50.0% < 100.0% (failed: all_components) \u2014 1/16 components failed (1 medium)", "invalidated": false, "invalidation_reason": null }, { - "model_id": "yujiepan/qwen3-next-moe-tiny-random", - "architecture_id": "Qwen3NextForCausalLM", - "verified_date": "2026-04-15", + "model_id": "yujiepan/deepseek-v3-tiny-random", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Full verification completed", + "notes": "Full verification completed with issues: P3=89.5% (failed: value_bias_folding, weight_modification)", "invalidated": false, "invalidation_reason": null }, { - "model_id": "Qwen/Qwen3.5-0.8B", - "architecture_id": "Qwen3_5ForCausalLM", - "verified_date": "2026-04-15", + "model_id": "yujiepan/deepseek-v3-tiny-random", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Below threshold: P1=50.0% < 100.0% (failed: all_components) \u2014 18/142 components failed (18 critical)", + "notes": "Full verification completed with issues: P3=89.5% (failed: value_bias_folding, weight_modification)", "invalidated": false, "invalidation_reason": null }, { - "model_id": "Qwen/Qwen3.5-0.8B", - "architecture_id": "Qwen3_5ForCausalLM", - "verified_date": "2026-04-15", + "model_id": "yujiepan/deepseek-v3-tiny-random", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Full verification completed", + "notes": "Full verification completed with issues: P3=89.5% (failed: value_bias_folding, weight_modification)", "invalidated": false, "invalidation_reason": null }, { - "model_id": "EleutherAI/pythia-70m-deduped", - "architecture_id": "GPTNeoXForCausalLM", - "verified_date": "2026-04-15", + "model_id": "yujiepan/deepseek-v3-tiny-random", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Full verification completed", + "notes": "Below threshold: P1=50.0% < 100.0% (failed: all_components) \u2014 1/16 components failed (1 medium)", "invalidated": false, "invalidation_reason": null }, { - "model_id": "Qwen/Qwen3-0.6B", - "architecture_id": "Qwen3ForCausalLM", - "verified_date": "2026-04-15", + "model_id": "yujiepan/deepseek-v3-tiny-random", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Full verification completed", + "notes": "Full verification completed with issues: P3=89.5% (failed: value_bias_folding, weight_modification)", "invalidated": false, "invalidation_reason": null }, { - "model_id": "openai-community/gpt2", - "architecture_id": "GPT2LMHeadModel", - "verified_date": "2026-04-15", + "model_id": "trl-internal-testing/tiny-DeepseekV3ForCausalLM", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Full verification completed", + "notes": "Below threshold: P1=50.0% < 100.0% (failed: all_components) \u2014 2/12 components failed (2 critical)", "invalidated": false, "invalidation_reason": null }, { - "model_id": "allenai/OLMo-2-0425-1B", - "architecture_id": "Olmo2ForCausalLM", - "verified_date": "2026-04-15", + "model_id": "trl-internal-testing/tiny-DeepseekV3ForCausalLM", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Full verification completed", + "notes": "Below threshold: P1=50.0% < 100.0% (failed: all_components) \u2014 2/12 components failed (2 critical)", "invalidated": false, "invalidation_reason": null }, { - "model_id": "trl-internal-testing/tiny-CohereForCausalLM", - "architecture_id": "CohereForCausalLM", - "verified_date": "2026-04-15", + "model_id": "optimum-intel-internal-testing/tiny-random-gigachat3", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Full verification completed with issues: P3=94.7% (failed: weight_modification)", + "notes": "Full verification completed with issues: P3=89.5% (failed: value_bias_folding, weight_modification)", "invalidated": false, "invalidation_reason": null }, { - "model_id": "EssentialAI/rnj-1-instruct", - "architecture_id": "Gemma3ForCausalLM", - "verified_date": "2026-04-15", + "model_id": "yujiepan/deepseek-v3-tiny-random", + "architecture_id": "DeepseekV3ForCausalLM", + "verified_date": "2026-04-08", "verified_by": "verify_models", "transformerlens_version": null, - "notes": "Full verification completed", + "notes": "Full verification completed with issues: P3=89.5% (failed: value_bias_folding, weight_modification)", "invalidated": false, "invalidation_reason": null }