部署GPT不用显卡,只需8G以上的内存!!!!!

大家好,今天给大家带来的项目是部署GPT,而且不需要显卡,只需要8G以上的内存即可实现

项目简介

项目地址

nomic-ai/gpt4all: gpt4all: run open-source LLMs anywhere (github.com)

简介

gpt4all是由nomicai发布的开源项目,以下是项目的自述文件

GPT4All

Open-source large language models that run locally on your CPU and nearly any GPU

GPT4All Website and Models

GPT4All Documentation

Discord

🦜️🔗 Official Langchain Backend

GPT4All is made possible by our compute partner Paperspace.

Run on an M1 macOS Device (not sped up!)

GPT4All: An ecosystem of open-source on-edge large language models.

Important

GPT4All v2.5.0 and newer only supports models in GGUF format (.gguf). Models used with a previous version of GPT4All (.bin extension) will no longer work.

GPT4All is an ecosystem to run powerful and customized large language models that work locally on consumer grade CPUs and any GPU. Note that your CPU needs to support AVX or AVX2 instructions.

Learn more in the documentation.

A GPT4All model is a 3GB - 8GB file that you can download and plug into the GPT4All open-source ecosystem software. Nomic AI supports and maintains this software ecosystem to enforce quality and security alongside spearheading the effort to allow any person or enterprise to easily train and deploy their own on-edge large language models.

What's New (Issue Tracker)

  • October 19th, 2023: GGUF Support Launches with Support for:

    • Mistral 7b base model, an updated model gallery on gpt4all.io, several new local code models including Rift Coder v1.5

    • Nomic Vulkan support for Q4_0, Q6 quantizations in GGUF.

    • Offline build support for running old versions of the GPT4All Local LLM Chat Client.

  • September 18th, 2023: Nomic Vulkan launches supporting local LLM inference on AMD, Intel, Samsung, Qualcomm and NVIDIA GPUs.

  • August 15th, 2023: GPT4All API launches allowing inference of local LLMs from docker containers.

  • July 2023: Stable support for LocalDocs, a GPT4All Plugin that allows you to privately and locally chat with your data.

Chat Client

Run any GPT4All model natively on your home desktop with the auto-updating desktop chat client. See GPT4All Website for a full list of open-source models you can run with this powerful desktop application.

Direct Installer Links:

Find the most up-to-date information on the GPT4All Website

Chat Client building and running

  • Follow the visual instructions on the chat client build_and_run page

Bindings

Integrations

Contributing

GPT4All welcomes contributions, involvement, and discussion from the open source community! Please see CONTRIBUTING.md and follow the issues, bug reports, and PR markdown templates.

Check project discord, with project owners, or through existing issues/PRs to avoid duplicate work. Please make sure to tag all of the above with relevant project identifiers or your contribution could potentially get lost. Example tags: , , , , etc.backendbindingspython-bindingsdocumentation

Technical Reports

📗 Technical Report 3: GPT4All Snoozy and Groovy

📗 Technical Report 2: GPT4All-J

📗 Technical Report 1: GPT4All

总结

该项目能在本地无需GPU跑gpt以及部分大模型,正如他们的官网说的一样

A free-to-use, locally running, privacy-aware chatbot. No GPU or internet required.

部署教程

购买服务器或者使用本地电脑进行项目

服务器推荐自家浪浪云,后续的镜像会同步跟新到官网上,首台服务器75折浪浪云:提供高性能、可靠稳定的云服务器解决方案 (langlangy.com)

内存推荐8G及以上,最优24g

文件下载

来到git地址或者官网下载可执行文件

选择对应的系统进行文件下载,此时我拿windows进行示范

模型下载

模型的下载有两种方式

软件内置下载

安装完软件之后来到首页,即可选择对应的软件模型

需要查看内存或者其他是否达标

内存或者其他不达标的情况

官方网站下载

官网选择对应的模型下载

选择对应的模型下载,下方有模型的简介

本实例使用gpt4all-falcon-newbpe-q4_0.gguf

模型简介:

模型速度较快而且质量较高

响应速度较快

基于based

由TII训练

由 Nomic AI 微调

许可用于商业用途

将下载好的文件和模型上传到服务器或者本机

在windows下直接点击安装即可,其余操作系统安装也较为相似,或者进群进行讨论也可

一直点击下一步安装即可,如果不想安装在浪浪云购买对应的产品即可(对应模型文件已经下载完毕)

模型切换

点击箭头所指位置进行模型的切换

产品使用

我现在用的微调模型只支持英文,使用英文和其进行对话,问它下逻辑题和其他问题

问题回复展示

内存占用情况

模型微调

后续跟新教程

现在存在的问题

对于中文问题的回应没有优化,需要一些英文水平,随着项目的推进会逐渐解决问题,但是无需要GPU进行本地部署,真的好香,