人気の記事一覧

GPT-2を読む⑪関連研究

GPT-2を読む⑦実験概要

GPT-2を読む⑧各タスクの結果

In-Context Retrieval-Augmented Language Models

8か月前

You Only Cache Once: Decoder-Decoder Architectures for Language Models

9か月前

MemLLM: Finetuning LLMs to Use An Explicit Read-Write Memory

10か月前

FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness

9か月前

State-Free Inference of State-Space Models: The Transfer Function Approach

9か月前

Ferret-v2: An Improved Baseline for Referring and Grounding with Large Language Models

10か月前

言語AIの進化史⑧埋め込みベクトル

MoEUT: Mixture-of-Experts Universal Transformers

8か月前

Lessons from the Trenches on Reproducible Evaluation of Language Models

8か月前

Scaling Transformer to 1M tokens and beyond with RMT

9か月前

Thinking Tokens for Language Modeling

9か月前

Memory Mosaics

9か月前

Granite Code Models: A Family of Open Foundation Models for Code Intelligence

9か月前

Text summarization with ChatGPT for drug labeling documents

9か月前

Infini-gram: Scaling Unbounded n-gram Language Models to a Trillion Tokens

9か月前

On the Long Range Abilities of Transformers

9か月前

Towards Graph Foundation Models: A Survey and Beyond

9か月前

Transformers are Multi-State RNNs

9か月前

Fewer Truncations Improve Language Modeling

10か月前

X-LoRA: Mixture of Low-Rank Adapter Experts, a Flexible Framework for Large Language Models with Applications in Protein Mechanics and Design