Join the conversation

Join the community of Machine Learners and AI enthusiasts.

Sign Up

All HF Hub posts

danielhanchen 
posted an update about 17 hours ago
danielhanchen 
posted an update 3 days ago
view post
Post
2392
A new way to use Unsloth.

Coming soon...
SeaWolf-AI 
posted an update about 15 hours ago
view post
Post
983
💎 Gemma 4 Playground — Dual Model Demo on ZeroGPU

We just launched a Gemma 4 Playground that lets you chat with Google DeepMind's latest open models — directly on Hugging Face Spaces with ZeroGPU.

FINAL-Bench/Gemma-4-Multi

👉 Try it now: FINAL-Bench/Gemma-4-Multi
Two Models, One Space
Switch between both Gemma 4 variants in a single interface:

⚡ Gemma 4 26B-A4B — MoE with 128 experts, only 3.8B active params. 95% of the 31B's quality at ~8x faster inference. AIME 88.3%, GPQA 82.3%.
🏆 Gemma 4 31B — Dense 30.7B. Best quality among Gemma 4 family. AIME 89.2%, GPQA 84.3%, Codeforces 2150. Arena open-model top 3.

Features

Vision — Upload images for analysis, OCR, chart reading, document parsing
Thinking Mode — Toggle chain-of-thought reasoning with Gemma 4's native <|channel> thinking tokens
System Prompts — 6 presets (General, Code, Math, Creative, Translate, Research) or write your own
Streaming — Real-time token-by-token response via ZeroGPU
Apache 2.0 — Fully open, no restrictions

Technical Details
Built with the dev build of transformers (5.5.0.dev0) for full Gemma 4 support including multimodal apply_chat_template, variable-resolution image processing, and native thinking mode. Runs on HF ZeroGPU with @spaces .GPU — no dedicated GPU needed.
Both models support 256K context window and 140+ languages out of the box.

Links

- 🤗 Space: [FINAL-Bench/Gemma-4-Multi]( FINAL-Bench/Gemma-4-Multi)
- 📄 Gemma 4 26B-A4B: [google/gemma-4-26B-A4B-it]( google/gemma-4-26B-A4B-it)
- 📄 Gemma 4 31B: [google/gemma-4-31B-it]( google/gemma-4-31B-it)
- 🔬 DeepMind Blog: [Gemma 4 Launch](https://deepmind.google/blog/gemma-4-byte-for-byte-the-most-capable-open-models/)
DedeProGames 
posted an update 2 days ago
view post
Post
2357
🔥 GRM2 - The small one that surpasses the big ones.
What if a 3-parameter model can beat a 32-parameter model in every benchmark? We prove that it can.
GRM2 is a 3b params model based on the llama architecture, trained for long reasoning and high performance in complex tasks - the first 3b params model to outperform qwen3-32b in ALL benchmarks, and outperform o3-mini in almost all benchmarks.
🤗 Model: OrionLLM/GRM2-3b
The first 3b params model to generate over 1000 lines of code and achieve a score of 39.0 in xBench-DeepSearch-2510.

🚀 Chat with GRM:
DedeProGames/GRM2-Chat

🏆 Download official GGUFs: OrionLLM/GRM2-3b-GGUF
SeaWolf-AI 
posted an update 3 days ago
view post
Post
2079
🧬 Darwin-35B-A3B-Opus — The Child That Surpassed Both Parents

What if a merged model could beat both its parents? We proved it can.
Darwin-35B-A3B-Opus is a 35B MoE model (3B active) built with our Darwin V5 engine — the first evolution system that CT-scans parent models before merging them.
🤗 Model: FINAL-Bench/Darwin-35B-A3B-Opus

The result speaks for itself: GPQA Diamond 90.0%, versus Father (Qwen3.5-35B-A3B) at 84.2% and Mother (Claude 4.6 Opus Distilled) at 85.0%. That's +6.9% over Father and +5.9% over Mother. Not a tradeoff — a genuine leap. Meanwhile, MMMLU sits at 85.0% (Father: 85.2%), multimodal is fully intact, and all 201 languages are preserved.

How? Model MRI changed everything. Traditional merging is guesswork. Darwin V4 added evolution. Darwin V5 added X-ray vision. Model MRI scans each parent layer by layer and discovers: Mother's L34–L38 is the reasoning engine (peak cosine distance), 50–65% of Mother's experts are dead (killed by text-only distillation), and Father is a healthy generalist with every expert alive. The prescription: transplant Mother's reasoning brain at L38 (90% weight), replace her dead experts with Father's living ones, and let Father's router handle the output layer. Reasoning went up. Versatility stayed intact. No tradeoff — just evolution.

35B total, 3B active (MoE) · GPQA Diamond 90.0% · MMMLU 85.0% (201 languages) · Multimodal Image & Video · 262K native context · 147.8 tok/s on H100 · Runs on a single RTX 4090 (Q4) · Apache 2.0
Darwin V5's full algorithm and technical details will be released alongside an upcoming paper.

🚀 Live Demo: FINAL-Bench/Darwin-35B-A3B-Opus

🏆 FINAL Bench Leaderboard: FINAL-Bench/Leaderboard

📊 ALL Bench Leaderboard: FINAL-Bench/all-bench-leaderboard

Built by VIDRAFT · Supported by the Korean Government GPU Support Program
  • 8 replies
·
ArtelTaleb 
posted an update 1 day ago
view post
Post
1514
Hey Music Lovers !

You have a track and you want the stems.

Retro 6stems splits any audio file into 6 clean, separated tracks

Drop an MP3, WAV, FLAC or M4A — Demucs gets to work immediately with a minimal retro design .

Instrumental mix included automatically
Vocals
Drums
Bass
Guitar
Piano
Other (strings, synths...)

Preview each stem directly in the browser, download individually or grab the full ZIP.

No install. No GPU needed on your end. Just upload and wait.


👉 ArtelTaleb/retro-6stems
shriarul5273 
posted an update 1 day ago
view post
Post
1595
🔁 One API. 12 model families. 28 variants. Why depth_estimation makes depth research easier

Switching between depth models usually means rewriting preprocessing, adapting outputs, and dealing with different codebases.

depth_estimation removes that friction.

With the same interface, you can work with:
🌊 Depth Anything
🍎 DepthPro
🧭 MiDaS
📏 ZoeDepth
🧩 MoGe
🛰️ VGGT / OmniVGGT
and more

Change one model string, keep the rest of your workflow the same.

That makes it much easier to:
⚖️ compare models fairly
🧪 prototype quickly
📈 benchmark consistently
🛠️ build reusable depth pipelines

GitHub: https://github.com/shriarul5273/depth_estimation

#depthestimation #research #computervision #python #machinelearning #opensource #pytorch
ArtelTaleb 
posted an update 2 days ago
view post
Post
2586
Hello 3D lover !

You have a 3D model sitting somewhere and just want to see it
no software to install, no account to create.

Canva3D is a free browser viewer built for that exact moment.

Drop any format — GLB, OBJ, FBX, STL, USDZ — it just loads.
Swap the background with HDRI lighting to set the mood of your scene.
Record a video ready to share — orbit, animate, export.

That's it. No settings to configure. No GPU required.

______________________________________________________

Bonjour les amoureux de la 3D

T'as un modèle 3D quelque part et tu veux juste le voir sans logiciel à
installer, sans compte à créer.

Canva3D est un viewer gratuit dans le navigateur, fait exactement pour ça.

Glisse n'importe quel format — GLB, OBJ, FBX, STL, USDZ —
il charge direct.
Change le fond avec un éclairage HDRI pour poser l'ambiance de ta scène.
Enregistre une vidéo prête à partager — orbite, anime, exporte.

C'est tout. Rien à configurer. Pas besoin de GPU.



👉 ArtelTaleb/canva3d
PhysiQuanty 
posted an update about 10 hours ago
sergiopaniego 
posted an update about 19 hours ago