• 0 Posts
  • 7 Comments
Joined 3 months ago
cake
Cake day: January 24th, 2025

help-circle
rss
  • @llama@lemmy.dbzer0.com Depends on the inference engine. Some of them will try to load the model until it blows up and runs out of memory. Which can cause its own problems. But it won’t overheat the phone, no. But if you DO use a model that the phone can run, like any intense computation, it can cause the phone to heat up. Best not run a long inference prompt while the phone is in your pocket, I think.




  • @hendrik@palaver.p3x.de no. I use the app service one. It works well, but it’s basically for bridging public channels. The Mautrix bridges all work very well. I’ve used the Facebook one in the past. It’s just the limits those platforms put on the bridge (e.g. banning or locking account) that can be a problem. If your bridge is connecting from the same place as you normally connect to Discord from, you should be fine.