文献阅读
文章平均质量分 91
文献内容整理
Espresso Macchiato
这个作者很懒,什么都没留下…
展开
-
文献阅读:Solving olympiad geometry without human demonstrations
文献阅读:Solving olympiad geometry without human demonstrations原创 2024-06-09 20:46:57 · 836 阅读 · 0 评论 -
文献阅读:Large Language Models Are Neurosymbolic Reasoners
文献阅读:Large Language Models Are Neurosymbolic Reasoners原创 2024-05-19 14:26:36 · 1112 阅读 · 0 评论 -
文献阅读:LESS: Selecting Influential Data for Targeted Instruction Tuning
文献阅读:LESS: Selecting Influential Data for Targeted Instruction Tuning原创 2024-04-14 19:28:27 · 1494 阅读 · 0 评论 -
文献阅读:APAR: LLMs Can Do Auto-Parallel Auto-Regressive Decoding
文献阅读:APAR: LLMs Can Do Auto-Parallel Auto-Regressive Decoding原创 2024-03-24 17:04:29 · 879 阅读 · 0 评论 -
文献阅读:Mixtral of Experts
文献阅读:Mixtral of Experts原创 2024-03-17 18:59:50 · 922 阅读 · 0 评论 -
文献阅读:Mistral 7B
文献阅读:Mistral 7B原创 2024-03-10 20:44:23 · 982 阅读 · 0 评论 -
文献阅读:The Unreasonable Effectiveness of Easy Training Data for Hard Tasks
文献阅读:The Unreasonable Effectiveness of Easy Training Data for Hard Tasks原创 2024-03-03 14:59:17 · 879 阅读 · 0 评论 -
文献阅读:Large Language Models are Null-Shot Learners
文献阅读:Large Language Models are Null-Shot Learners原创 2024-02-25 18:28:27 · 987 阅读 · 0 评论 -
文献阅读:Transformers are Multi-State RNNs
文献阅读:Transformers are Multi-State RNNs原创 2024-02-25 18:24:07 · 1104 阅读 · 0 评论 -
文献阅读:Mamba: Linear-Time Sequence Modeling with Selective State Spaces
文献阅读:Mamba: Linear-Time Sequence Modeling with Selective State Spaces原创 2024-02-11 16:24:29 · 1952 阅读 · 0 评论 -
文献阅读:Large Language Models as Optimizers
文献阅读:Large Language Models as Optimizers原创 2024-01-14 19:17:02 · 1334 阅读 · 1 评论 -
文献阅读:Sparse Low-rank Adaptation of Pre-trained Language Models
文献阅读:Sparse Low-rank Adaptation of Pre-trained Language Models原创 2024-01-07 19:36:55 · 1234 阅读 · 2 评论 -
文献阅读:PromptAgent: Strategic Planning with Language Models Enables Expert-level Prompt Optimization
文献阅读:PromptAgent: Strategic Planning with Language Models Enables Expert-level Prompt Optimization原创 2023-12-31 19:01:55 · 1030 阅读 · 0 评论 -
文献阅读:Exchange-of-Thought: Enhancing Large Language Model Capabilities through Cross-Model Communicat
文献阅读:Exchange-of-Thought: Enhancing Large Language Model Capabilities through Cross-Model Communication原创 2023-12-31 18:56:48 · 958 阅读 · 0 评论 -
文献阅读:LoRA: Low-Rank Adaptation of Large Language Models
文献阅读:LoRA: Low-Rank Adaptation of Large Language Models原创 2023-12-31 18:34:54 · 1255 阅读 · 0 评论 -
文献阅读:LLMLingua: Compressing Prompts for Accelerated Inference of Large Language Models
文献阅读:LLMLingua: Compressing Prompts for Accelerated Inference of Large Language Models原创 2023-12-24 20:05:22 · 1906 阅读 · 0 评论 -
文献阅读:StopWasting My Time! Saving Days of ImageNet and BERT Training with LatestWeight Averaging
文献阅读:StopWasting My Time! Saving Days of ImageNet and BERT Training with LatestWeight Averaging原创 2023-12-17 20:01:41 · 383 阅读 · 0 评论 -
文献阅读:Take a Step Back:Evoking Reasoning via Abstraction in Large Language Models
文献阅读:Take a Step Back:Evoking Reasoning via Abstraction in Large Language Models原创 2023-12-10 19:26:29 · 1070 阅读 · 0 评论 -
文献阅读:LONGNET: Scaling Transformers to 1,000,000,000 Tokens
文献阅读:LONGNET: Scaling Transformers to 1,000,000,000 Tokens原创 2023-11-05 16:04:06 · 379 阅读 · 0 评论 -
文献阅读:The Reversal Curse: LLMs trained on “A is B” fail to learn “B is A”
文献阅读:The Reversal Curse: LLMs trained on “A is B” fail to learn “B is A”原创 2023-10-22 20:39:52 · 139 阅读 · 0 评论 -
文献阅读:RLAIF: Scaling Reinforcement Learning from Human Feedback with AI Feedback
文献阅读:RLAIF: Scaling Reinforcement Learning from Human Feedback with AI Feedback原创 2023-10-01 16:58:19 · 880 阅读 · 0 评论 -
文献阅读:LIMA: Less Is More for Alignment
文献阅读:LIMA: Less Is More for Alignment原创 2023-09-24 19:52:28 · 222 阅读 · 0 评论 -
文献阅读:Chain-of-Thought Prompting Elicits Reasoning in Large Language Models
文献阅读:Chain-of-Thought Prompting Elicits Reasoning in Large Language Models原创 2023-09-03 20:15:17 · 730 阅读 · 0 评论 -
文献阅读:Is ChatGPT Good at Search? Investigating Large Language Models as Re-Ranking Agent
文献阅读:Is ChatGPT Good at Search? Investigating Large Language Models as Re-Ranking Agent原创 2023-08-27 18:46:16 · 834 阅读 · 1 评论 -
文献阅读:AnnoLLM: Making Large Language Models to Be Better Crowdsourced Annotators
文献阅读:AnnoLLM: Making Large Language Models to Be Better Crowdsourced Annotators原创 2023-08-13 19:59:03 · 516 阅读 · 0 评论 -
文献阅读:MathPrompter: Mathematical Reasoning using Large Language Models
文献阅读:MathPrompter: Mathematical Reasoning using Large Language Models原创 2023-07-16 18:11:13 · 1531 阅读 · 0 评论 -
文献阅读:Foundation Transformers
文献阅读:Foundation Transformers原创 2023-06-18 18:19:43 · 1255 阅读 · 0 评论 -
文献阅读:LLaMA: Open and Efficient Foundation Language Models
文献阅读:LLaMA: Open and Efficient Foundation Language Models原创 2023-05-07 17:03:19 · 787 阅读 · 0 评论 -
文献阅读:LaMDA: Language Models for Dialog Applications
文献阅读:LaMDA: Language Models for Dialog Applications原创 2023-05-07 16:57:41 · 723 阅读 · 0 评论 -
文献阅读:Large Language Models Encode Clinical Knowledge
文献阅读:Large Language Models Encode Clinical Knowledge原创 2023-03-12 18:27:48 · 1261 阅读 · 2 评论 -
文献阅读:Training language models to follow instructions with human feedback
文献阅读:Training language models to follow instructions with human feedback原创 2023-02-26 18:24:14 · 3411 阅读 · 2 评论 -
文献阅读:Scaling Instruction-Finetuned Language Models
文献阅读:Scaling Instruction-Finetuned Language Models原创 2023-02-12 14:39:58 · 2009 阅读 · 0 评论 -
文献阅读:Finetuned Language Models Are Zero-Shot Learners
文献阅读:Finetuned Language Models Are Zero-Shot Learners原创 2023-02-12 14:35:30 · 608 阅读 · 0 评论 -
文献阅读:Language Models are Few-Shot Learners
文献阅读:Language Models are Few-Shot Learners原创 2023-02-05 16:51:54 · 1216 阅读 · 1 评论 -
文献阅读:Language Models are Unsupervised Multitask Learners
文献阅读:Language Models are Unsupervised Multitask Learners原创 2023-01-29 19:45:04 · 826 阅读 · 0 评论 -
文献阅读:Improving Language Understanding by Generative Pre-Training
文献阅读:Improving Language Understanding by Generative Pre-Training原创 2023-01-29 19:41:42 · 1115 阅读 · 0 评论 -
文献阅读:Towards Out-of-Distribution Sequential Event Prediction:A Causal Treatment
文献阅读:Towards Out-of-Distribution Sequential Event Prediction:A Causal Treatment原创 2023-01-15 16:03:10 · 541 阅读 · 0 评论 -
文献阅读:Image as a Foreign Language: BEIT Pretraining for All Vision and Vision-Language Tasks
文献阅读:Image as a Foreign Language: BEIT Pretraining for All Vision and Vision-Language Tasks原创 2022-10-23 16:39:06 · 1249 阅读 · 0 评论 -
文献阅读:RoFormer: Enhanced Transformer with Rotary Position Embedding
文献阅读:RoFormer: Enhanced Transformer with Rotary Position Embedding1. 工作简介2. 常见位置编码方式1. 绝对位置编码1. Bert2. Attention Is All You Need2. 相对位置编码1. 经典相对位置编码2. XLNet3. T54. DeBerta3. RoPE方法介绍4. 实验效果考察1. 翻译任务上的表现2. 预训练任务上的表现3. GLUE下游Finet..原创 2022-05-08 18:55:07 · 3377 阅读 · 2 评论 -
文献阅读:Linformer: Self-Attention with Linear Complexity
文献阅读:Linformer: Self-Attention with Linear Complexity1. 问题描述2. 核心方法1. vanilla attention layer2. attention优化3. 分析 & 证明1. self-attention是低阶的2. linear self-attention效果与vanilla self-attention相仿3. 实验1. 预训练效果考察2. 下游任务效果3. 时间优化考察4. 结论 ..原创 2022-04-17 18:18:49 · 4660 阅读 · 0 评论