Byron Gibson
bgibson
·
AI & ML interests
None yet
Recent Activity
updated a collection 16 days ago
llm-models updated a collection about 1 year ago
llm-models liked a model about 1 year ago
unsloth/DeepSeek-R1-GGUFOrganizations
None yet
papers
-
Vision Mamba: Efficient Visual Representation Learning with Bidirectional State Space Model
Paper • 2401.09417 • Published • 62 -
MoE-Mamba: Efficient Selective State Space Models with Mixture of Experts
Paper • 2401.04081 • Published • 74 -
SwitchHead: Accelerating Transformers with Mixture-of-Experts Attention
Paper • 2312.07987 • Published • 41 -
DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models
Paper • 2401.06066 • Published • 60
llm-models
llm-local
-
LLM in a flash: Efficient Large Language Model Inference with Limited Memory
Paper • 2312.11514 • Published • 264 - Paused98
Transformers to Core ML
⚡98Display a loading screen with a spinner
-
enterprise-explorers/Llama-2-7b-chat-coreml
Text Generation • Updated • 919 • 137 -
tiiuae/falcon-7b-instruct
Text Generation • 7B • Updated • 59.5k • 1.03k
llm-datasets
llm-analysis
llm-local
-
LLM in a flash: Efficient Large Language Model Inference with Limited Memory
Paper • 2312.11514 • Published • 264 - Paused98
Transformers to Core ML
⚡98Display a loading screen with a spinner
-
enterprise-explorers/Llama-2-7b-chat-coreml
Text Generation • Updated • 919 • 137 -
tiiuae/falcon-7b-instruct
Text Generation • 7B • Updated • 59.5k • 1.03k
papers
-
Vision Mamba: Efficient Visual Representation Learning with Bidirectional State Space Model
Paper • 2401.09417 • Published • 62 -
MoE-Mamba: Efficient Selective State Space Models with Mixture of Experts
Paper • 2401.04081 • Published • 74 -
SwitchHead: Accelerating Transformers with Mixture-of-Experts Attention
Paper • 2312.07987 • Published • 41 -
DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models
Paper • 2401.06066 • Published • 60
llm-datasets
llm-models