AI’s impact on PC memory and storage

AI abounds Nobody goes a day sans hearing or seeing AI AI is transforming our interactions, from smart devices to self-driving automobiles

AI relies on big language models (LLMs) created from humans’ massive unlabeled text. Artificial brains with billions of variables and often many networks create material for spoken queries that mimic human replies

AI may happen anywhere outside the cloud Moving some AI processing to consumer devices might be helpful for numerous reasons Edge AI can optimise latency, privacy, network costs, and offline functionality

PCs feature huge screens that show more information and improve user experience Second, PCs’ big batteries can sustain longer, more demanding AI work

Intel, AMD, Qualcomm, Mediatek, and Nvidia are incorporating strong computational engines and/or integrated graphics in PC CPUs and chipsets to give tens of TOPS of AI performance

That’s hardly surprising given Microsoft’s drive on Copilot, an AI-powered service that helps users create code, troubleshoot issues, and recommend improvements

Model size is a major issue when running AI’s impact on PC. AI models, particularly LLMs, need a lot of data storage and memory to store and load billions or trillions of parameters

A local on-device personal assistant may also require a huge parameter model. A model with less than 10B parameters may be utilized on common devices