Forget RAG, Embrace Large Models

Forget RAG, Embrace Large Models

“ Happy New Year! This article should be the last piece in the collection of algorithm insights regarding RAG. Throughout the past year, most of my work has focused on some aspects of large model applications, and I would like to briefly discuss two points that left a significant impression on me regarding RAG system … Read more

The Path to AGI: From Alchemy to Energy Equation

The Path to AGI: From Alchemy to Energy Equation

“Years later, facing the firing squad, Colonel Aureliano Buendía will remember that distant afternoon when his father took him to see ice blocks.” “One Hundred Years of Solitude” opens with a “future past tense” that sets the stage for the century-long fate of the Buendía family. José Buendía leads his kin away from their hometown … Read more

How Mianbi Intelligent Surpasses Large Models with MiniCPM

How Mianbi Intelligent Surpasses Large Models with MiniCPM

Cost is the invisible competitive advantage of large models. Author|Liu Yangnan Editor|Zhao Jian Today, the Tsinghua University-affiliated large model company “Mianbi Intelligent” released its first flagship large model “Mianbi MiniCPM”, which has been aptly named “Little Cannon”. According to Mianbi Intelligent’s co-founder and CEO Li Dahai, the parameter scale of Mianbi MiniCPM is 2B, using … Read more

Performance of 2B Parameters Surpasses Mistral-7B: Wall Intelligence Multimodal Edge Model Open Source

Performance of 2B Parameters Surpasses Mistral-7B: Wall Intelligence Multimodal Edge Model Open Source

Machine Heart reports Editor: Zenan Low-cost devices can run locally. As large models continue to evolve towards larger scales, recent developments have also been made in optimization and deployment. On February 1, Wall Intelligence, in collaboration with Tsinghua NLP Laboratory, officially launched its flagship edge large model “Wall MiniCPM” in Beijing. The new generation large … Read more

Four Lines of Code to Triple Large Model Context Length

Four Lines of Code to Triple Large Model Context Length

Crecy from Aofeisi Quantum Bit | WeChat Official Account QbitAI No fine-tuning is required; just four lines of code can triple the context length of large models! Moreover, it is “plug-and-play” and theoretically adaptable to any large model, successfully tested on Mistral and Llama2. With this technology, large models (LargeLM) can transform into LongLM. Recently, … Read more

Ollama: Local Large Model Running Guide

Ollama: Local Large Model Running Guide

Foreword If your hard drive is running low on space, check this out first for a thrill. Running models has become as easy as changing packages. This article introduces the Ollama framework developed in Go language, which allows users to run large models locally. Through Ollama, users can download and run different models, and generate … Read more

Getting Started with Meta Llama3-8B Using Ollama and OpenWebUI

Getting Started with Meta Llama3-8B Using Ollama and OpenWebUI

On April 18, 2024, Meta open-sourced the Llama 3 large models[1]. Although there are only 8B[2] and 70B[3] versions, the powerful capabilities demonstrated by Llama 3 have shocked the AI large model community. I personally tested the inference capabilities of the Llama3-70B version, which are very close to OpenAI’s GPT-4[4]. Moreover, a 400B super large … Read more

Ollama: Local Large Model Running Guide

Ollama: Local Large Model Running Guide

The author of this article is a front-end developer at 360 Qiwutuan. Introduction to Ollama Ollama is an open-source framework developed in Go that can run large models locally. Official website: https://ollama.com/ GitHub repository: https://github.com/ollama/ollama Installing Ollama Download and Install Ollama Choose the appropriate installation package based on your operating system type from the Ollama … Read more

Guide to Large Model Training: Faith Brings Success

Guide to Large Model Training: Faith Brings Success

Source|YuanChuan Institute (ID: YuanChuanInstitution) Author|Chen Bin Editor|Dai Boss Cover Image|Movie “Blade Runner 2049″ Every large model is an expensive “money printer”, which has become a common knowledge among various AI observers. The training cost of large models has a simple ratio: the training cost is proportional to the square of the number of parameters. For … Read more