Logo
Explore Help
Register Sign In
sleepy/llama.cpp
1
0
Fork 0
You've already forked llama.cpp
Code Issues 8 Pull Requests 1 Actions 98 Packages Projects Releases Wiki Activity
Labels Milestones New Issue
8 Open 3 Closed
Label
Use alt + click/enter to exclude labels
All labels No label
bug

doc

feature

infra

kernel

perf

profiling

Milestone
All milestones No milestones
Project
All projects No project
Author
All users
Assignee
Assigned to nobody Assigned to anybody
sleepy
Sort
Newest Oldest Most recently updated Least recently updated Most commented Least commented Nearest due date Farthest due date
8 Open 3 Closed
Label
Clear labels
bug
doc
feature
infra
kernel
perf
profiling
Milestone
No milestone
Projects
Clear projects
Assignee
Clear assignees
No assignee
sleepy
Implement MXFP4 GGUF converter feature
#37 opened 2026-04-30 18:11:37 +02:00 by sleepy
Compare llama.cpp and MLX dispatch structure profiling
#36 opened 2026-04-30 18:11:37 +02:00 by sleepy
Profile concurrent encoding effectiveness profiling
#35 opened 2026-04-30 18:11:37 +02:00 by sleepy
Profile graph fusion effectiveness profiling
#34 opened 2026-04-30 18:11:36 +02:00 by sleepy
KV cache IO scaling with context length perf
#32 opened 2026-04-30 18:11:35 +02:00 by sleepy
Investigate CPY overhead (159 MB/tick at 9B) perf
#31 opened 2026-04-30 18:11:35 +02:00 by sleepy
Investigate GET_ROWS overhead (678 MB/tick at 9B) perf
#30 opened 2026-04-30 18:11:35 +02:00 by sleepy
Port contiguous weight reads to Q4_0 MUL_MAT kernel kernel
#29 opened 2026-04-30 18:11:34 +02:00 by sleepy
Powered by Gitea Version: 1.26.1 Page: 25ms Template: 1ms
Auto
English
Bahasa Indonesia Deutsch English Español Français Gaeilge Italiano Latviešu Magyar nyelv Nederlands Polski Português de Portugal Português do Brasil Suomi Svenska Türkçe Čeština Ελληνικά Български Русский Українська فارسی മലയാളം 日本語 简体中文 繁體中文(台灣) 繁體中文(香港) 한국어
Licenses API