🦙
🦙
I like big .vimrc and I cannot lie
- Sofia, Bulgaria
-
04:52
(UTC +02:00) - https://ggerganov.com
- @ggerganov
Sponsors
Block or Report
Block or report ggerganov
Report abuse
Contact GitHub support about this user’s behavior. Learn more about reporting abuse.
Report abusePinned
2,874 contributions in the last year
| Day of Week | November Nov | December Dec | January Jan | February Feb | March Mar | April Apr | May May | June Jun | July Jul | August Aug | September Sep | October Oct | November Nov | ||||||||||||||||||||||||||||||||||||||||
| Sunday Sun | |||||||||||||||||||||||||||||||||||||||||||||||||||||
| Monday Mon | |||||||||||||||||||||||||||||||||||||||||||||||||||||
| Tuesday Tue | |||||||||||||||||||||||||||||||||||||||||||||||||||||
| Wednesday Wed | |||||||||||||||||||||||||||||||||||||||||||||||||||||
| Thursday Thu | |||||||||||||||||||||||||||||||||||||||||||||||||||||
| Friday Fri | |||||||||||||||||||||||||||||||||||||||||||||||||||||
| Saturday Sat | |||||||||||||||||||||||||||||||||||||||||||||||||||||
Less
More
Contribution activity
November 2023
Created 64 commits in 4 repositories
Created a pull request in ggerganov/whisper.cpp that received 20 comments
whisper : add full CUDA and Metal offloading
Build with: # make WHISPER_CUBLAS=1 make -j # cmake cmake -DWHISPER_CUBLAS=1 ../ Also, the convolution ops are now offloaded both with CUDA and Met…
+1,029
−1,707
lines changed
•
20
comments
Opened 29 other pull requests in 3 repositories
ggerganov/whisper.cpp
2
open
9
merged
2
closed
-
sync : ggml (ggml-alloc + linker + gguf fixes)
This contribution was made on Nov 16
-
whisper : quantize encoder only
This contribution was made on Nov 16
-
whisper : make large version explicit + fix data size units
This contribution was made on Nov 15
-
java : fix test
This contribution was made on Nov 15
-
whisper : add batched decoding
This contribution was made on Nov 14
-
java : use tiny.en for tests
This contribution was made on Nov 13
-
whisper : try to fix the parallel whisper_state functionality
This contribution was made on Nov 11
-
whisper : support ggml_conv with CUDA and Metal
This contribution was made on Nov 10
-
whisper : migrating to ggml-backend
This contribution was made on Nov 9
-
whisper : add support for large v3
This contribution was made on Nov 7
-
ggml : try to fix the abort mechanism
This contribution was made on Nov 5
-
whisper : add support for new distilled Whisper models
This contribution was made on Nov 3
-
sync : ggml (backend v2, k-quants, CUDA opts, Metal opts, etc.)
This contribution was made on Nov 3
ggerganov/llama.cpp
1
open
10
merged
-
llama : fix data units
This contribution was made on Nov 16
-
sync : ggml (im2col, GPU conv, 32-bit arm compat)
This contribution was made on Nov 13
-
make : do not add linker flags when compiling static llava lib
This contribution was made on Nov 7
-
common : YAYF (yet another YARN fix)
This contribution was made on Nov 3
-
cuda : fix const ptrs warning causing ROCm build issues
This contribution was made on Nov 2
-
sync : ggml (backend v2)
This contribution was made on Nov 2
-
gguf : print error for GGUFv1 files
This contribution was made on Nov 2
-
gguf : remove special-case code for GGUFv1
This contribution was made on Nov 2
-
metal : fix build errors and rope kernel sig after #2268
This contribution was made on Nov 2
-
cuda : do not use batched GEMM when tensor cores are not available
This contribution was made on Nov 1
-
llm : add llm_build_context
This contribution was made on Nov 1
ggerganov/ggml
1
open
4
merged
-
sync : whisper.cpp (update whisper example + minor)
This contribution was made on Nov 16
-
sync : llama.cpp (CUDA ReLU, CPU-only with CUDA, bloom fix, etc)
This contribution was made on Nov 13
-
sync : whisper.cpp (whisper full GPU, fix warnings)
This contribution was made on Nov 12
-
sync : whisper.cpp (ARM 32-bit, abort callback, wav_writer, etc.)
This contribution was made on Nov 3
-
sync : llama.cpp (CUDA opts, ggml-quants, YARN, etc.)
This contribution was made on Nov 2
Reviewed 56 pull requests in 3 repositories
ggerganov/llama.cpp
25 pull requests
-
gguf: fix potential infinite loops while parsing
This contribution was made on Nov 16
-
StableLM support
This contribution was made on Nov 16
-
Update gpt2 preprocess and add deepseek coder preprocess
This contribution was made on Nov 16
-
Use BLAS to implement ggml_compute_forward_out_prod_f32 for matrix src0, src1 (finetuning speedup ~5x).
This contribution was made on Nov 16
-
Fix #4017
This contribution was made on Nov 16
-
ggml-cuda : increase max graph size
This contribution was made on Nov 15
-
Fix MacOS Sonoma model quantization
This contribution was made on Nov 14
-
llava : fix regression for square images in #3613
This contribution was made on Nov 13
-
Add ReLU and SQR CUDA ops to fix Persimmon offloading
This contribution was made on Nov 13
-
sync : ggml (backend v2)
This contribution was made on Nov 12
-
gguf-py: Refactor and allow reading/modifying existing GGUF files
This contribution was made on Nov 9
-
Generalize convert scripts
This contribution was made on Nov 9
-
Fix backward rope after YaRN
This contribution was made on Nov 7
-
Expose Llava as a shared library for downstream projects
This contribution was made on Nov 6
-
feat(ci): add an option to fail on compile warning
This contribution was made on Nov 6
-
supports running on CPU for GGML_USE_CUBLAS=ON build
This contribution was made on Nov 6
-
ggml-cuda : fix f16 mul mat
This contribution was made on Nov 5
-
cuda : fix disabling device with --tensor-split 1,0
This contribution was made on Nov 5
-
feat: mark LLM_ARCH_STARCODER as full offload supported
This contribution was made on Nov 5
-
server : allow continue edit on completion mode
This contribution was made on Nov 5
-
ggml-metal: round up to 16 to fix setThreadgroupMemoryLength assertion
This contribution was made on Nov 3
-
MSVC instruction detection (fixed up #809)
This contribution was made on Nov 3
-
ggml-cuda : move row numbers to x grid dim in mul mat vec kernels
This contribution was made on Nov 3
-
Enable sigint handler even when not in interactive mode
This contribution was made on Nov 3
-
Add ROCM aliases for CUDA pool stuff
This contribution was made on Nov 2
- Some pull request reviews not shown.
ggerganov/whisper.cpp
17 pull requests
-
ISSUE-1329: replace " with ' so it doesn't try to execute code in backticks
This contribution was made on Nov 16
-
whisper : add grammar-based sampling
This contribution was made on Nov 13
-
Add n_gpu_layers option to talk-llama example
This contribution was made on Nov 13
-
whisper : add full CUDA and Metal offloading
This contribution was made on Nov 11
-
whisper : support ggml_conv with CUDA and Metal
This contribution was made on Nov 10
-
examples : Enhanced compatibility with older Android versions using Java
This contribution was made on Nov 10
-
Return with error from whisper_encode_internal and whisper_decode_int…
This contribution was made on Nov 10
-
Add '-l auto' to talk-llama example
This contribution was made on Nov 9
-
whisper : migrating to ggml-backend
This contribution was made on Nov 9
-
models : Fix
n_melmismatch in convert-whisper-to-openvino.pyThis contribution was made on Nov 9 -
Reset ctx->t_start_us when calling whisper_reset_timings()
This contribution was made on Nov 7
-
Fix variable names in GitHub actions config
This contribution was made on Nov 7
-
talk-llama : fix n_gpu_layers usage again
This contribution was made on Nov 7
-
whisper : add context param for disable gpu
This contribution was made on Nov 5
-
Allow env variable to override metal resource path
This contribution was made on Nov 3
-
examples: Fix the encoding issues on Windows
This contribution was made on Nov 3
-
Use absolute paths for the converted OpenVINO model
This contribution was made on Nov 3
ggerganov/ggml
4 pull requests
-
update examples and tests to use ggml_allocr_new_measure_from_backend
This contribution was made on Nov 13
-
Passing parameters and simple prompt on SAM CLI
This contribution was made on Nov 2
-
Changing SAM documentation to provide executable example
This contribution was made on Nov 2
-
Fix Windows MSVC Builds
This contribution was made on Nov 1
Created an issue in ggerganov/llama.cpp that received 4 comments
clip : offload to GPU
With the recent support for running convolutions on the GPU (#4060) we should be able to offload CLIP to run fully on the GPU. llama.cpp/example…
4
comments
Opened 4 other issues in 1 repository
ggerganov/llama.cpp
3
open
1
closed
-
metal : compile-time kernel args and params
This contribution was made on Nov 15
-
llama : become thread-safe
This contribution was made on Nov 5
-
llama : build with AVX support on Windows by default with CMake
This contribution was made on Nov 2
-
ci : add an option to fail on compile warning
This contribution was made on Nov 2






