{"type":"video","version":"1.0","html":"<iframe width=\"560\" height=\"315\" sandbox=\"allow-same-origin allow-scripts allow-popups allow-forms\" title=\"Running LLMs locally: Practical LLM Performance on DGX Spark — Mozhgan Kabiri chimeh, NVIDIA\" src=\"https://video.ut0pia.org/videos/embed/9Bnh3TzXCVBmi7YwrwMrzf\" style=\"border: none\" allow=\"fullscreen\"></iframe>","width":560,"height":315,"title":"Running LLMs locally: Practical LLM Performance on DGX Spark — Mozhgan Kabiri chimeh, NVIDIA","author_name":"AI","author_url":"https://video.ut0pia.org/video-channels/ai","provider_name":"PeerTube","provider_url":"https://video.ut0pia.org","thumbnail_url":"https://video.ut0pia.org/lazy-static/thumbnails/ad310b44-c767-4197-8eca-5c1b7ed6f39e.jpg","thumbnail_width":1280,"thumbnail_height":720}