Skip to main content
Question

M.2 with more RAM or a smarter off-load

  • August 22, 2025
  • 3 replies
  • 65 views

I'm evaluating your devices. I really love both the European heart and the great TOPS/Watt ratio.

It would be amazing for edge computing.

On the other side, our company is waiting for a GX10 from ASUS/NVidia with 128Gb of shared memory.

The reason why I think that the Metis form factor is more interesting than the pci-ex would need a very long document :-)

I read that your SDK is supporting off-load but reading on the forum I think that your off-loading support is limited only to data and not the core of the model.

You know that when you use CUDA+torch you're able to off-load some part of the model out of the GPU memory and execute it running a small part of the model at once swapping coefficients from the GPU memory to the host RAM and vice versa.

Are you planning to improve your off-loading system in your SDK?

Are you planning to introduce a Metis with more RAM?

 

3 replies

Spanner
Axelera Team
Forum|alt.badge.img+2
  • Axelera Team
  • August 28, 2025

Hi there ​@dario.finardi ! Great to see you here, and that you’re finding Metis interesting for edge use!

Yeah, you’re right, our focus has been more on optimising full-model execution on the AIPU, but we’re always tracking use cases like yours. In terms of increased memory, there’s thing to share in that respect at the moment, but again, it’s useful to hear that there’s potential demand and value in it 👍 I’ll pass that along.

What kind of models/applications are you targeting?


  • Author
  • Cadet
  • August 28, 2025

Actually we’re running EuroLLM-9b, really large for Metis but also EuroLLM-1.7b with some fine tuning could fit, otherwise new Gemma3N-e2b seems to be promising.
 

 


Spanner
Axelera Team
Forum|alt.badge.img+2
  • Axelera Team
  • August 29, 2025

Ah, nice! It’s always great to hear about this different models people are running, and this sounds awesome. Keep us posted on how it goes! Would love to see it in action some time, too.