Commit Graph

  • 7a6e385c1b
    sync : ggml Georgi Gerganov 2024-02-25 19:59:34 +0200
  • 578e47e70c
    sync : llama.cpp (ggml/0) Georgi Gerganov 2024-02-25 19:58:06 +0200
  • fac5b43830
    code : normalize enum names (llama/5697) Georgi Gerganov 2024-02-25 12:09:09 +0200
  • 9e7c5212a1
    IQ3_S: a much better alternative to Q3_K (llama/5676) Kawrakow 2024-02-24 16:23:52 +0200
  • 1cb64f7368
    Introduce backend GUIDs (ggml/743) UEXTM.com 2024-02-24 11:27:36 -0500
  • ab10265128
    Stream.wasm: Fix invalid memory access when no segments are returned Andrew S 2024-02-24 16:55:21 -0600
  • 7469011f4f Correct syscall name SYS_get_cpu used with old glibc where getcpu() is missing. Petter Reinholdtsen 2024-02-24 05:02:46 +0100
  • f18738f247
    talk, talk-llama : pass text_to_speak as a file (#1865) Tamotsu Takahashi 2024-02-24 16:24:47 +0900
  • 0e4977aeb7 talk and talk-llama: ignore to_speak.txt Tamotsu Takahashi 2024-02-24 16:08:31 +0900
  • 4e8cce737d examples: add speak_with_file() Tamotsu Takahashi 2024-02-20 01:42:00 +0900
  • 458347b56c talk and talk-llama: add more options to eleven-labs.py Tamotsu Takahashi 2024-02-15 23:27:16 +0900
  • 650a9660b9 talk and talk-llama: get voices from api Tamotsu Takahashi 2024-02-14 09:54:07 +0900
  • fd8ee9084e talk and talk-llama: get all installed voices in speak.ps1 Tamotsu Takahashi 2024-02-14 00:45:09 +0900
  • 6584e6301b talk: sync with talk-llama Tamotsu Takahashi 2024-02-13 20:21:56 +0900
  • c72857b1c4 talk-llama: pass voice_id again Tamotsu Takahashi 2024-02-13 20:42:44 +0900
  • f5ec91c696 talk-llama: check installed commands in "speak" Tamotsu Takahashi 2024-02-13 19:26:51 +0900
  • 766c8cb6df talk-llama: let eleven-labs.py accept options Tamotsu Takahashi 2024-02-13 17:57:55 +0900
  • 83d48ee7d6 talk-llama: pass heard_ok as a file Tamotsu Takahashi 2024-02-13 17:03:11 +0900
  • df26676494 talk-llama: pass file instead of arg Tamotsu Takahashi 2024-02-12 23:17:39 +0900
  • a0ddd8392c
    whisper : add SYCL support (#1863) Abhilash Majumder 2024-02-23 12:52:24 +0530
  • 2bec4899b5
    remove cxx file check Abhilash Majumder 2024-02-23 11:39:25 +0530
  • a2506909b1
    talk-llama : sync llama.cpp Georgi Gerganov 2024-02-22 23:30:53 +0200
  • 7b1ff212d9
    sync : ggml Georgi Gerganov 2024-02-22 23:25:38 +0200
  • e5d06cfc0f
    ggml : always define ggml_fp16_t as uint16_t (llama/5666) Georgi Gerganov 2024-02-22 23:21:39 +0200
  • 31891db2e3
    ci : fix whitespace Georgi Gerganov 2024-02-22 20:20:34 +0200
  • 9989c9cdf1
    Merge branch 'ggerganov:master' into sycl_dev Abhilash Majumder 2024-02-22 23:31:45 +0530
  • 5fdb27ff80
    ggml : 32-bit arm compat (#1891) Georgi Gerganov 2024-02-22 18:31:40 +0200
  • 48f68f0ae2
    ggml : cont Georgi Gerganov 2024-02-22 17:17:56 +0200
  • cdf08ab7ad
    ggml : add ggml_vqtbl1q_s8 impl Georgi Gerganov 2024-02-22 15:44:14 +0200
  • 47876c2497
    ggml : 32-bit arm compat Georgi Gerganov 2024-02-22 15:30:40 +0200
  • 6b16927d18
    sync : ggml Georgi Gerganov 2024-02-22 15:15:38 +0200
  • ce411498f6
    sync : llama.cpp (ggml/0) Georgi Gerganov 2024-02-21 16:19:39 +0200
  • 208de95ac7
    conext add name (llama/5624) Meng, Hengyu 2024-02-21 17:52:06 +0800
  • c2ce39c795
    Update ggml_sycl_op_mul_mat_vec_q (llama/5502) AidanBeltonS 2024-02-20 07:01:25 +0000
  • 8daa534818
    Refactor validation and enumeration platform checks into functions to clean up ggml_vk_instance_init() 0cc4m 2024-02-14 20:57:17 +0100
  • 9fca69b410
    Add check for VK_KHR_portability_enumeration for MoltenVK support 0cc4m 2024-02-10 22:14:52 +0100
  • b26c645420
    Add preprocessor checks for Apple devices. Mathijs de Bruin 2024-02-06 14:39:22 +0000
  • 1879ec556e
    Resolve ErrorIncompatibleDriver with Vulkan on MacOS. Mathijs de Bruin 2024-02-03 18:00:11 +0000
  • c6e53cfc46
    Allow for Vulkan build with Accelerate. Mathijs de Bruin 2024-02-03 17:56:46 +0000
  • b19f2fb815
    cuda : ignore peer access already enabled errors (llama/5597) slaren 2024-02-19 23:40:26 +0100
  • a6b0950916
    ggml : compute forward no longer pass src tensors (ggml/729) Siddharth Ramakrishnan 2024-02-21 04:34:53 -0800
  • d352dbd163
    ggml : fix conv_2d batch mode (ggml/737) bssrdf 2024-02-20 14:17:09 -0500
  • eb23f4ef16
    openvino : fix convert-whisper-to-openvino.py (#1890) st-gr 2024-02-22 05:11:35 -0800
  • c56344b509
    main : fix file existence check in main.cpp (#1889) Davidson Francis 2024-02-22 10:01:08 -0300
  • e74cddab2d add readme abhilash1910 2024-02-22 02:45:52 -0800
  • c1a397bc6f add sample abhilash1910 2024-02-22 02:34:36 -0800
  • cbc9fe87a7
    fix quant typecast Abhilash Majumder 2024-02-22 13:28:20 +0530
  • b5a88f3b7c readd fp16 and readme abhilash1910 2024-02-21 23:51:37 -0800
  • 3d2a731fe6 fix typecast in quant.c abhilash1910 2024-02-21 23:39:20 -0800
  • f7c5d1fae6 openvino : fix convert-whisper-to-openvino.py for v2023.0.0 (#1870) Stefan Grundmann 2024-02-21 23:26:48 -0800
  • 64ec1087a9 add target abhilash1910 2024-02-21 23:21:19 -0800
  • ae7e48b028
    Merge branch 'ggerganov:master' into sycl_dev Abhilash Majumder 2024-02-22 11:29:47 +0530
  • 686de031ce main : fix file existence check in main.cpp Davidson Francis 2024-02-21 23:19:09 -0300
  • 383bf4f8cd examples: refine the Android sample code zhou.weiguo 2024-02-22 09:47:53 +0800
  • 15efeb2613 add examples abhilash1910 2024-02-20 20:10:26 -0800
  • 59119f4f20
    talk-llama : sync llama.cpp Georgi Gerganov 2024-02-20 12:09:57 +0200
  • 276615d708
    make : fix CUBLAS link with WSL (#1878) LBlue 2024-02-20 18:05:38 +0800
  • 103df0d26b Linking issue fix via Makefile when CUBLAS enabled in the WSL #1876 xmtXmtxytxyt 2024-02-20 00:15:11 +0800
  • 508489dd98
    Merge f10a7b43a5 into b602819b6e Didzis Gosko 2024-02-19 13:56:55 +0000
  • b602819b6e
    sync : ggml Georgi Gerganov 2024-02-19 15:54:25 +0200
  • c2c606f05b
    ggml : resolve merge conflicts (ggml/0) Georgi Gerganov 2024-02-19 15:33:51 +0200
  • 83afebe872
    common : add IQ1_S (ggml/0) Georgi Gerganov 2024-02-19 15:27:37 +0200
  • a4d8f9d559
    ci : enable -Werror for CUDA builds (llama/5579) Georgi Gerganov 2024-02-19 14:45:41 +0200
  • 5ec1e0edfa
    cuda, metal : fix nans in soft_max (llama/5574) slaren 2024-02-19 09:04:45 +0100
  • 30a11b1ab8
    ggml : android and old glibc NUMA incompatibility bugfixes (llama/5557) bmwl 2024-02-18 23:38:32 -0800
  • f04e6b87d7
    ggml : restore vec dot stride arg names (llama/5453) Georgi Gerganov 2024-02-18 22:58:57 +0200
  • 0c33928b55
    ci : fix wikitext url + compile warnings (llama/5569) Georgi Gerganov 2024-02-18 22:39:30 +0200
  • 0775374750
    metal : fix unused warnings (llama/0) Georgi Gerganov 2024-02-18 21:39:58 +0200
  • 7d90bb035b
    ggml, common, examples, tests : fixed type arguments in printf (llama/5528) Herman Semenov 2024-02-18 16:20:12 +0000
  • 2c1ad21ba8
    1.5 bit quantization (llama/5453) Kawrakow 2024-02-18 18:16:55 +0200
  • eca5ff9868
    ggml : add ALiBi support for ggml_soft_max_ext (llama/5488) Georgi Gerganov 2024-02-19 15:18:09 +0200
  • 1b25d2fa0a
    ci : add an option to fail on compile warning (llama/3952) Ananta Bastola 2024-02-17 16:03:14 -0500
  • 74a6acc999
    cmake : fix VULKAN and ROCm builds (llama/5525) Georgi Gerganov 2024-02-16 19:05:56 +0200
  • a4ed8a0821
    ggml : add numa options (llama/5377) bmwl 2024-02-16 01:31:07 -0800
  • 9f675e021c
    cuda : print message when initialization fails (llama/5512) slaren 2024-02-15 16:49:01 +0100
  • a38efcb9fd
    vulkan: Find optimal memory type but with fallback (llama/5381) Neuman Vong 2024-02-15 17:11:15 +1100
  • 31591649a0
    Early return for zero size calls to get_tensor. (llama/5482) AT 2024-02-13 15:44:25 -0600
  • 4f5c46a84f
    ggml-quants : fix compiler warnings (shadow variable) (llama/5472) Kawrakow 2024-02-13 09:07:57 +0200
  • 462ffc58db
    ggml-sycl: Replace 3d ops with macro (llama/5458) Abhilash Majumder 2024-02-12 20:22:05 +0530
  • 65faae0b6a
    build : update CBLAS flags + fix unused var warning (#0) Georgi Gerganov 2024-02-19 14:44:46 +0200
  • 5b2876607d use sycl header abhilash1910 2024-02-19 02:45:43 -0800
  • dda4b0ed06
    main : check if input files exist before proceeding (#1872) Davidson Francis 2024-02-19 05:51:26 -0300
  • 07d04280be
    examples : clean up common code (#1871) Felix 2024-02-19 09:50:15 +0100
  • fdcd73c958
    Merge 98d4b23baf into 917c56ded4 shanelenagh 2024-02-19 00:39:06 -0800
  • 917c56ded4
    models : fix openvino setup info (#1874) Jumper775 2024-02-18 21:19:47 -0500
  • eb93a6319e
    fix openvino setup info Jumper775 2024-02-17 20:44:38 -0500
  • e33773fa3c main : check if input files exist before proceeding Davidson Francis 2024-02-16 17:35:49 -0300
  • ef2e31bb9b clean up common code in examples felrock 2024-02-16 17:41:32 +0100
  • 66c5b2611e
    call build Abhilash Majumder 2024-02-16 14:10:39 +0530
  • 125b1321ea
    update build Abhilash Majumder 2024-02-16 12:06:24 +0530
  • 33fbf8ddfc fix bug abhilash1910 2024-02-15 05:34:16 -0800
  • ec79010589 refactor build abhilash1910 2024-02-15 03:45:29 -0800
  • 9d3d1d23f2
    Add files via upload bobqianic 2024-02-14 01:29:33 +0000
  • b7bc969d65
    Merge pull request #9 from bobqianic/fix bobqianic 2024-02-14 00:31:57 +0000
  • 53d58fb149
    rewrite bpe_gpt2_preprocess bobqianic 2024-02-14 00:31:05 +0000
  • 6eb97e9114
    Add files via upload bobqianic 2024-02-14 00:24:54 +0000
  • 0143c9d83d
    fix bug Abhilash Majumder 2024-02-13 17:42:30 +0530
  • 35e04fbd62
    fix bug Abhilash Majumder 2024-02-13 17:41:50 +0530
  • 96ac62ddf1
    add sycl build config Abhilash Majumder 2024-02-13 17:35:23 +0530
  • 7386c906a9
    update cmake Abhilash Majumder 2024-02-13 17:29:14 +0530