all AI news
Fast groq-hosted LLMs vs browser jank
May 19, 2024, 1:35 p.m. |
Simon Willison's Weblog simonwillison.net
Fast groq-hosted LLMs vs browser jank
Groq is now serving LLMs such as Llama 3 so quickly that JavaScript which attempts to render Markdown strings on every new token can cause performance issues in browsers.
Taras Glek's solution was to move the rendering to a requestAnimationFrame() callback, effectively buffering the rendering to the fastest rate the browser can support.
Via lobste.rs
browser browsers callback every groq javascript llama llama 3 llms markdown performance render rendering solution strings token
More from simonwillison.net / Simon Willison's Weblog
Quoting Andrej Karpathy
17 hours ago |
simonwillison.net
How (some) good corporate engineering blogs are written
2 days, 6 hours ago |
simonwillison.net
Jobs in AI, ML, Big Data
Senior Machine Learning Engineer
@ GPTZero | Toronto, Canada
ML/AI Engineer / NLP Expert - Custom LLM Development (x/f/m)
@ HelloBetter | Remote
Doctoral Researcher (m/f/div) in Automated Processing of Bioimages
@ Leibniz Institute for Natural Product Research and Infection Biology (Leibniz-HKI) | Jena
Seeking Developers and Engineers for AI T-Shirt Generator Project
@ Chevon Hicks | Remote
Principal Autonomy Applications
@ BHP | Chile
Quant Analytics Associate - Data Visualization
@ JPMorgan Chase & Co. | Bengaluru, Karnataka, India