-
Attention Is All You Need
Paper • 1706.03762 • Published • 120 -
Language Models are Few-Shot Learners
Paper • 2005.14165 • Published • 20 -
GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints
Paper • 2305.13245 • Published • 6 -
Llama 2: Open Foundation and Fine-Tuned Chat Models
Paper • 2307.09288 • Published • 251
Eli Chen
elichen3051
AI & ML interests
Learning Algorithm, Reinforcement Learning, Data Synthesize, Benchmarking
Organizations
LLM Fundamental papers
-
Attention Is All You Need
Paper • 1706.03762 • Published • 120 -
Language Models are Few-Shot Learners
Paper • 2005.14165 • Published • 20 -
GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints
Paper • 2305.13245 • Published • 6 -
Llama 2: Open Foundation and Fine-Tuned Chat Models
Paper • 2307.09288 • Published • 251
My interested papers - LLM
models 6
elichen3051/Llama-3.1-8B-GGUF
8B • Updated • 21
elichen3051/Llama-3.2-1B-GGUF-fp16
1B • Updated • 20
elichen3051/Llama-3.1-8B-Instruct-GGUF
8B • Updated • 1
elichen3051/Llama-3.2-1B-Q4_K_M-GGUF
Text Generation • 1B • Updated • 5
elichen3051/Llama-3.1-8B-Q4_K_M-GGUF
Text Generation • 8B • Updated • 14
elichen3051/mistral-7B-v0.1-open-orca
Updated