What if instead, you could yell them at your computer?
Introducing:🗣🖥️ Talk To Your Terminal, an extremely personal human-computer interface.
github.com/aukejw/ttyt
What if instead, you could yell them at your computer?
Introducing:🗣🖥️ Talk To Your Terminal, an extremely personal human-computer interface.
github.com/aukejw/ttyt
We will have 2 tracks: one for new research (to be published in proceedings) and one for recently published works.
Deadline: June 1st
Link: sites.google.com/view/beyonde...
We will have 2 tracks: one for new research (to be published in proceedings) and one for recently published works.
Deadline: June 1st
Link: sites.google.com/view/beyonde...
For the 8B model, I'm seeing 45-55 tokens/sec with mlx_lm, and 25-40 tokens/sec with LM Studio depending on context length.
(do take LM Studio peak memory with a grain of salt, I'm sure it's not constant)
For the 8B model, I'm seeing 45-55 tokens/sec with mlx_lm, and 25-40 tokens/sec with LM Studio depending on context length.
(do take LM Studio peak memory with a grain of salt, I'm sure it's not constant)
Below: generation with the 12B model (4 bit quantized).
However, I found it hard to find out which models my Macbook could handle, and how fast I *should* expect inference to be.
Below: generation with the 12B model (4 bit quantized).
However, I found it hard to find out which models my Macbook could handle, and how fast I *should* expect inference to be.
We're organizing the 2nd workshop on Efficient Large Vision Models at CVPR 2025.
Welcoming all submissions on efficient (visual) foundation models!
⏰ Submission deadline: March 21
🌐 sites.google.com/view/elvm/
If you're a researcher working on efficiency of foundation models, and are interested in reviewing, let me know by March 14th!
forms.gle/Ra1dDwmzjxyT...
If you're a researcher working on efficiency of foundation models, and are interested in reviewing, let me know by March 14th!
forms.gle/Ra1dDwmzjxyT...
We're organizing the 2nd workshop on Efficient Large Vision Models at CVPR 2025.
Welcoming all submissions on efficient (visual) foundation models!
⏰ Submission deadline: March 21
🌐 sites.google.com/view/elvm/
We're organizing the 2nd workshop on Efficient Large Vision Models at CVPR 2025.
Welcoming all submissions on efficient (visual) foundation models!
⏰ Submission deadline: March 21
🌐 sites.google.com/view/elvm/
qualcomm-ai-research.github.io/mobile-video...
go check out its demo at Qualcomm booth if you are attending NeurIPS!
youtu.be/aARq9YXEQEQ