This is a Phi Family of SLMs book for getting started with Phi Models. Phi a family of open sourced AI models developed by Microsoft. Phi models are the most capable and cost-effective small language ...
We view Large Language Models as stochastic language layers in a network, where the learnable parameters are the natural language prompts at each layer. We stack two such layers, feeding the output of...
#计算机科学#Train to 94% on CIFAR-10 in <6.3 seconds on a single A100. Or ~95.79% in ~110 seconds (or less!)
#大语言模型#Commanding robots using only Language Models' prompts
An n-gram model is a type of probabilistic language model for predicting the next item in such a sequence. In n gram language modelling, words are modeled such that each n-gram is composed of n words
Gesture recognition helps computers to understand human body language. This helps to build a more potent link between humans and machines, rather than just the basic text user interfaces or graphical ...
ChatGPT built with NEXT.JS, awesome gpt prompts, and more inspired by: Large Language Models are Zero-Shot Reasoners
This repository explores Context-Aware Generation (CAG) as a next-generation alternative or complement to Retrieval-Augmented Generation (RAG). Both are powerful paradigms designed to improve the fact...
How does the keyboard on your phone know what you would like to type next? Language prediction is a Natural Language Processing - NLP application concerned with predicting the text given in the preced...
GPT-3: Language Models are Few-Shot Learners
《Evaluating Large Language Models Trained on Code》论文代码
Experiments and data for the paper "When and why vision-language models behave like bags-of-words, and what to do about it?" Oral @ ICLR 2023
Liquid: Language Models are Scalable and Unified Multi-modal Generators
#大语言模型#Next-token prediction in JavaScript — build fast language and diffusion models.
gpt-oss-120b and gpt-oss-20b are two open-weight language models by OpenAI
#大语言模型#Code and models for ICML 2024 paper, NExT-GPT: Any-to-Any Multimodal Large Language Model
Code for the Paper "Zero-Shot Next-Item Recommendation using Large Pretrained Language Models"
A guidance language for controlling large language models.
#大语言模型#[ECIR'24] Implementation of "Large Language Models are Zero-Shot Rankers for Recommender Systems"
pre-trained Language Models
The papers are organized according to our survey: Evaluating Large Language Models: A Comprehensive Survey.
Next-gen language engineering / DSL framework
StableLM: Stability AI Language Models
Reading notes about Multimodal Large Language Models, Large Language Models, and Diffusion Models