Julien Simon

Julien Simon

FR
@juliensimonfr
407
Video Count
3.1M
Video View
503.0K
Subscriber
#1,681
France Rank
#76,317
Global Rank
Julien Simon YouTube channel subscribers:503,000- Seelive statisticsand growth insights below.

Julien Simon YouTube Statistics & Analytics

Subscribers
503.0K
Total Views
3.1M
Videos
407
Activity
Unknown

Julien Simon Content Analysis

Content Type Distribution

Long videosLong
65%
28 videos
ShortsShorts
35%
15 videos

⚖️ This channel maintains a balanced mix of Shorts and Long videos for diverse audience engagement.

Content Categories

Primary CategoryScience & Technology
100%
Science & Technology
43(100%)

🎯 Primary focus: Science & Technology with 43 videos (100% of categorized content).

Latest Video

Long video
Deep Dive: LLM Quantization, part 3 - FP8, FP4
37:33

Deep Dive: LLM Quantization, part 3 - FP8, FP4

386
Views
7
Likes
2 weeks ago
Published

Two years after parts 1 (https://youtu.be/kw7S-3s50uk) and 2 (https://youtu.be/fXBBwCIA0Ds), the quantization landscape has changed completely. FP8 is the new default; the serving stack has absorbed quantization, and MoE models have broken the old assumptions. This video shows you what practitioners actually use today. In Part 3 of this series, I'm showing you exactly what to run, on which GPU, with which tool, and what breaks when you get it wrong. One model, start to finish: Arcee Trinity Mini (26B MoE, 3B active, 128 experts). What's covered: → Why FP8 is the new BF16 — essentially lossless, one flag, works everywhere → The three things people call "4-bit" (INT4 vs MXFP4 vs NVFP4) and why they're completely different → Serving with vLLM: BF16, FP8 on-the-fly, FP8 checkpoint, with verified H100 numbers → Why MoE models break standard quantization and how to fix it → Creating your own checkpoints with llm-compressor and NVIDIA ModelOpt → NVFP4 on Blackwell: real 4-bit tensor core math, what works and what doesn't yet → The FP4 deep dive: E2M1 bit layout, MXFP4 vs NVFP4 scale factors, worked error example → Decision tree: which quantization path for your hardware and use case *** Slides https://fr.slideshare.net/slideshow/advanced-quantization-techniques-for-large-language-models-in-2026-a4c8/286754686 *** Arcee Trinity Mini: 26B total parameters, 3B active per token, 128 experts, 128K context, Apache 2.0. https://huggingface.co/arcee-ai/Trinity-Mini NVFP4: https://huggingface.co/arcee-ai/Trinity-Mini-NVFP4 FP8: https://huggingface.co/arcee-ai/Trinity-Mini-FP8-Block #llm #quantization #fp8 #fp4 #nvfp4 #vllm #inference #nvidia #blackwell #moe #arcee

See Top Science & Technology YouTube Channels in France

Compare this channel with the leading Science & Technology creators in France.

Ranking: FranceCategory: Science & TechnologyCategory Focus: 100%
Open ranking

Julien Simon AI Channel Analysis

Gemini ProScore: 7.2/10

AI-powered insights analyzing content strategy, audience engagement, and growth potential.

Overall Score
7.2
Consistency
95%
Cadence
2-3/wk
Library
43

Growth Potential

7.5/10

Good content foundation. Increasing upload frequency could boost growth.

Audience Engagement

7.2/10

Moderate engagement levels. Focus on community interaction could improve metrics.

Content Strategy

7/10

Developing content strategy. Consider focusing on specific niches for better targeting.

AI Recommendations

Auto-prioritized by predicted impact

  1. 1
    Increase upload frequency to 2-3 videos per week
    High ImpactCadence
  2. 2
    Focus on SEO optimization for better discoverability
    High ImpactSEO
  3. 3
    Analyze top-performing content for pattern replication
    MediumStrategy
  4. 4
    Increase community engagement through comments and polls
    MediumEngagement

Frequently Asked Questions About Julien Simon

Data Source & Accuracy

Source: YouTube Data API v3
Accuracy: Real-time statistics from official YouTube API
Data is updated hourly and sourced directly from official APIs to ensure accuracy and reliability.

Data from YouTube Data API v3 • Updated hourly • Last updated: 10:55 PM