Daniel Palenicek
@daniel-palenicek.bsky.social
22 followers 24 following 17 posts
PhD Researcher in Robot #ReinforcementLearning 🤖🧠 at IAS TU Darmstadt and hessian.AI advised by Jan Peters. Former intern at BCAI and Huawei R&D UK.
Posts Media Videos Starter Packs
daniel-palenicek.bsky.social
Read the full preprint here:
👉 arxiv.org/pdf/2509.25174
Code coming soon.
We’d love feedback & discussion! 💬
arxiv.org
daniel-palenicek.bsky.social
Key takeaway:
Well-conditioned optimization > raw scale.

XQC proves principled architecture choices can outperform larger, more complex ones
daniel-palenicek.bsky.social
📊 Results across 70 tasks (55 proprioception + 15 vision-based):

⚡️ Matches/outperforms SimbaV2, BRO, BRC, MRQ, and DRQ-V2

🌿~4.5× fewer parameters and 1/10 FLOP/s of SimbaV2

💪Especially strong on the hardest tasks: HumanoidBench, DMC Hard & DMC Humanoids from pixels
daniel-palenicek.bsky.social
This leads to XQC, a streamlined extension of Soft Actor-Critic with
✅ only 4 hidden layers
✅ BN after each linear layer
✅ WN projection
✅ CE critic loss

Simplicity + principled design = efficiency ⚡️
daniel-palenicek.bsky.social
🔑 Insight: A simple synergy—BatchNorm + WeightNorm + Cross-Entropy loss—makes critics dramatically more well-conditioned.

➡️Result: Stable effective learning rates and smoother optimization.
daniel-palenicek.bsky.social
Instead of "bigger is better," we ask:
Can better conditioning beat scaling?

By analyzing the Hessian eigenspectrum of critic networks, we uncover how different architectural choices shape optimization landscapes.
daniel-palenicek.bsky.social
🚀 New preprint! Introducing XQC— a simple, well-conditioned actor-critic that achieves SOTA sample efficiency in #RL
✅ ~4.5× fewer parameters than SimbaV2
✅ Scales to vision-based RL
👉 arxiv.org/pdf/2509.25174

Thanks to Florian Vogt @joemwatson.bsky.social @jan-peters.bsky.social
daniel-palenicek.bsky.social
If you're working on RL stability, plasticity, or sample efficiency, this might be relevant for you.

We'd love to hear your thoughts and feedback!

Come talk to us at RLDM in June in Dublin (rldm.org)
RLDM | The Multi-disciplinary Conference on Reinforcement Learning and Decision Making
rldm.org
daniel-palenicek.bsky.social
📚 TL;DR: We combine BN + WN in CrossQ for stable high-UTD training and SOTA performance on challenging RL benchmarks. No need for network resets, no critic ensembles, no other tricks... Simple regularization, big gains.

Paper: t.co/Z6QrMxZaPY
https://arxiv.org/abs/2502.07523v2
t.co
daniel-palenicek.bsky.social
⚖️ Simpler ≠ Weaker: Compared to SOTA baselines like BRO our method:
✅ Needs 90% fewer parameters (~600K vs. 5M)
✅ Avoids parameter resets
✅ Scales stably with compute.

We also compare strongly to the concurrent SIMBA algorithm.

No tricks—just principled normalization. ✨
daniel-palenicek.bsky.social
🔬 The Result: CrossQ + WN scales reliably with increasing UTD—no more resets, no critic ensembles, no other tricks.
We match or outperform SOTA on 25 continuous control tasks from DeepMind Control Suite & MyoSuite, including dog 🐕 and humanoid🧍‍♂️tasks across UTDS.
daniel-palenicek.bsky.social
➡️ With growing weight norm, the effective learning rate decreases, and learning slows down/stops.

💡Solution: After each gradient update, we rescale parameters to the unit sphere, preserving plasticity and keeping the effective learning rate stable.
daniel-palenicek.bsky.social
🧠Key Idea: BN improves sample efficiency, but fails to reliably scale with complex tasks & high UTDs due to growing weight norms.
However, BN regularized networks are scale invariant w.r.t. their weights; yet, the gradient scales inversely proportional (Van Laarhoven 2017)
daniel-palenicek.bsky.social
🔍 Background: Off-policy RL methods like CrossQ (Bhatt* & Palenicek* et al. 2024) are sample-efficient but struggle to scale to high update-to-data (UTD) ratios.

We identify why scaling CrossQ fails—and fix it with a surprisingly effective tweak: Weight Normalization (WN). 🏋️
daniel-palenicek.bsky.social
🚀 New preprint "Scaling Off-Policy Reinforcement Learning with Batch and Weight Normalization"🤖

We propose CrossQ+WN, a simple yet powerful off-policy RL for more sample-efficiency and scalability to higher update-to-data ratios. 🧵 t.co/Z6QrMxZaPY

#RL @ias-tudarmstadt.bsky.social
https://arxiv.org/abs/2502.07523v2
t.co
daniel-palenicek.bsky.social
Check out our latest work, where we train an omnidirectional locomotion policy directly on a real quadruped robot in just a few minutes based on our CrossQ RL algorithm 🚀
Shoutout @nicobohlinger.bsky.social, Jonathan Kinzel.

@ias-tudarmstadt.bsky.social @hessianai.bsky.social
nicobohlinger.bsky.social
⚡️ Do you think training robot locomotion needs large scale simulation? Think again!

We train an omnidirectional locomotion policy directly on a real quadruped in just a few minutes 🚀
Top speeds of 0.85 m/s, two different control approaches, indoor and outdoor experiments, and more! 🤖🏃‍♂️