Auke
banner
aukejw.bsky.social
Auke
@aukejw.bsky.social
generative models and making them faster
When using the terminal, I often find myself typing the same few commands over and over.

What if instead, you could yell them at your computer?

Introducing:🗣🖥️ Talk To Your Terminal, an extremely personal human-computer interface.

github.com/aukejw/ttyt
May 6, 2025 at 2:15 PM
Reposted by Auke
We are back at #ICCV2025 for the second workshop on Hyperbolic and Hyperspherical Learning for Computer Vision.

We will have 2 tracks: one for new research (to be published in proceedings) and one for recently published works.

Deadline: June 1st
Link: sites.google.com/view/beyonde...
Beyond Euclidean - Call for Papers
Important Dates: Submission Portal Opens: 29th of April 2025 Submission Deadline: 1st of June 2025 (AOE) Preliminary Author Notification Deadline: 27th of June 2025 Camera-ready deadline (Proceeding...
sites.google.com
May 1, 2025 at 6:21 AM
Qwen3 4bit and 8bit models are now available, and they are fast!

For the 8B model, I'm seeing 45-55 tokens/sec with mlx_lm, and 25-40 tokens/sec with LM Studio depending on context length.

(do take LM Studio peak memory with a grain of salt, I'm sure it's not constant)
April 29, 2025 at 11:24 AM
I've been benchmarking Gemma 3 models on my M4 Pro to see how fast MLX is. The answer is: very!

Below: generation with the 12B model (4 bit quantized).

However, I found it hard to find out which models my Macbook could handle, and how fast I *should* expect inference to be.
April 22, 2025 at 1:10 PM
One week left to submit your work to eLVM2!
🚀Call for papers🚀

We're organizing the 2nd workshop on Efficient Large Vision Models at CVPR 2025.

Welcoming all submissions on efficient (visual) foundation models!

⏰ Submission deadline: March 21
🌐 sites.google.com/view/elvm/
March 14, 2025 at 10:48 AM
We're looking for reviewers for the 2nd edition of our CVPR workshop on efficient large vision models (eLVM).

If you're a researcher working on efficiency of foundation models, and are interested in reviewing, let me know by March 14th!

forms.gle/Ra1dDwmzjxyT...
Reviewer enrollment for CVPR 2025 eLVM workshop -- deadline March 14th
Workshop info: https://sites.google.com/view/elvm Important dates: - Submission deadline March 21st - Reviews between March 21st and 28th, please make sure you have time this week!
forms.gle
March 10, 2025 at 1:30 PM
🚀Call for papers🚀

We're organizing the 2nd workshop on Efficient Large Vision Models at CVPR 2025.

Welcoming all submissions on efficient (visual) foundation models!

⏰ Submission deadline: March 21
🌐 sites.google.com/view/elvm/
February 26, 2025 at 2:00 PM
Turn a video of my face (or Jens, Guillaume,..) into anything at our NeurIPS booth -- running a video diffusion model on a mobile phone ->
my group just released MoViE (Mobile Diffusion for Video Editing):

qualcomm-ai-research.github.io/mobile-video...

go check out its demo at Qualcomm booth if you are attending NeurIPS!

youtu.be/aARq9YXEQEQ
World's fastest genAI video editing on a phone
YouTube video by Qualcomm Research
youtu.be
December 11, 2024 at 5:34 PM