Google's Locallm: Developing AI applications without a GPU

Teacher

Professional
Messages
2,673
Reputation
9
Reaction score
692
Points
113
Now you can create innovative solutions directly on local processors.

Google introduced an open source toolset called localllm, which allows developers to take advantage of large language models (LLMs) directly on local processors or within Workstations, a fully managed development environment in the Google Cloud. The innovation opens up new opportunities for specialists in the field of machine learning and artificial intelligence to create and test their projects, while ensuring a high level of data protection and confidentiality.

The language models are hosted on the Hugging Face platform and are located in The Bloke repository. One of the key features of these models is their compatibility with the quantization method, which makes them suitable for running on processors or GPUs with low power consumption.

Quantized models are artificial intelligence models that are optimized to run on local devices with limited computing resources. These models are designed to be more efficient in terms of memory usage and processing power, allowing them to run smoothly on devices such as smartphones, laptops, and other peripherals. Google suggested deploying such models on cloud workstations.

The models are optimized to perform calculations using lower-precision data types, such as 8-bit integers, instead of standard 32-bit floating-point numbers. By representing weights and activations with fewer bits, the overall size of the model is reduced, which makes it easier to place on devices with limited memory capacity. Quantized models can perform calculations faster due to lower accuracy and smaller size.

Google's approach, based on combining quantum models and cloud workstations, allows developers to take full advantage of the flexibility, scalability, and cost-effectiveness of this platform.

localllm provides a set of tools and libraries for easy access to quantized models from HuggingFace via a command-line utility. This repository provides a comprehensive framework and tools for running LLM locally on the CPU and in memory directly on Google Cloud Workstation or on a computer or other device. localllm integrates with various Google Cloud services, including data warehousing, machine learning APIs, and so on.

To get started with localllm, developers need to go to the GitHub repository , where they will find detailed documentation, code samples, and instructions for configuring and using LLM locally on the processor and in the Google Cloud environment. The process involves installing the toolset, loading and running the model from HuggingFace, and executing an initial health check request.
 
Top