OpenAI are advancing work on large language models (LLMs) that power mainstream generative AI (artificial intelligence), a few others are gearing up to deploy small AI models, which are likely to reach the consumers sometime in 2024. Industry insiders see this as a natural progression to monetizing generative AI technologies, making them more widely accessible, instead of targeting only enterprises through very expensive cloud platforms. For instance, on 23 February, Qualcomm demonstrated an image-generating AI model running on a device, with Stable Diffusion showcased on an Android smartphone.
Subsequently, on 18 July, Meta and Qualcomm jointly introduced Llama-2, an LLM which will be available for real-world use by next year. Experts said this approach is a way forward to address concerns over costs, environmental impact and practicality, and would pave the way for further adoption of AI technology. Speaking at a media roundtable, Ziad Asghar, senior vice-president of product management, Qualcomm, said: “When an AI model is being trained, it takes all the information from the cloud.
This training aspect will happen on the cloud—but now, it is extremely intensive to use it for inference. This is why we’re looking at making AI models smaller and more compressed for running them on-device." This evolution in generative AI is natural, Jaspreet Bindra, founder of Tech Whisperer, a technology consultancy firm, said. “This should be the way for our future, and we can be very confident about it.
Read more on livemint.com