Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

sync master #7

Merged
merged 47 commits into from
May 28, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
47 commits
Select commit Hold shift + click to select a range
dacfceb
readme : add GPT-NeoX + Pythia to the list of supported models (#7491)
felladrin May 23, 2024
55ac3b7
ci : use Pythia models instead of OpenLlama (#7470)
ggerganov May 23, 2024
3015851
llama : add getters for n_threads/n_threads_batch (#7464)
danbev May 23, 2024
8b94e79
readme : add Bunny in supported models [no ci] (#7469)
criminact May 23, 2024
007489e
Fix phi3 chat template confusion with zephyr (#7449)
tristandruyen May 23, 2024
1debe72
ggml : silence UB sanitizer error during iq2_xxs quantization (#0)
ggerganov May 23, 2024
74f33ad
readme : remove trailing space (#7469)
ggerganov May 23, 2024
0df0aa8
add build shared lib in win release package (#7438)
arthw May 24, 2024
fbca2f2
Add support for ArcticForCausalLM (#7020)
fairydreaming May 24, 2024
27891f6
docker.yml: disable light-intel and server-intel test (#7515)
mofosyne May 24, 2024
d041d2c
flake.lock: Update (#7232)
ggerganov May 24, 2024
b83bab1
gguf-py : fix and simplify quantized shape round-trip (#7483)
compilade May 25, 2024
5768433
Make tokenize CLI tool have nicer command line arguments. (#6188)
Noeda May 25, 2024
902184d
fix missing slash in `fs_get_cache_directory()` (#7503)
ngxson May 25, 2024
9791f40
android : module (#7502)
eltonkola May 25, 2024
faa0e69
ggml: aarch64: SVE kernels for q8_0_q8_0, q4_0_q8_0 vector dot (#7433)
msy-kato May 25, 2024
00c6390
main : don't print special tokens with --grammar (#6923)
jart May 25, 2024
3cbd23e
labeler: added Apple Metal detector (+Kompute) (#7529)
mofosyne May 25, 2024
9588f19
train : change default FA argument (#7528)
ggerganov May 25, 2024
b9adcbb
SimpleChat Completion Mode flexibility and cleanup, Settings gMe, Opt…
hanishkvc May 26, 2024
9146d36
Readme: add akx/ggify to tools (#1484)
akx May 26, 2024
c429b33
llama : add Smaug 70B support (#7402)
bartowski1182 May 26, 2024
32a2821
Fix aya-23 conversion scripts (#7539)
Galunid May 26, 2024
d298382
main: replace --no-special with --special (#7534)
mofosyne May 26, 2024
dff451c
flake.lock: Update (#7540)
ggerganov May 26, 2024
d6ef0e7
github: add self sorted issue ticket forms (#7543)
mofosyne May 27, 2024
eaf6e03
llama : add comments about experimental flags (#7544)
ggerganov May 27, 2024
62bfef5
metal : disable FA kernel for HS=256 (#7556)
ggerganov May 27, 2024
1d8fca7
metal : add GGML_OP_REPEAT kernels (#7557)
ggerganov May 27, 2024
5487593
Add freq factors (#7495)
AidanBeltonS May 27, 2024
95f84d5
Fix q_xxs using mul_mat_q (#7459)
AidanBeltonS May 27, 2024
197c006
Allow multiple copy function pointers for CUDA graph kernel param upd…
agray3 May 27, 2024
10b1e45
make: add --device-debug to NVCC debug flags (#7542)
JohannesGaessler May 27, 2024
0136966
adding in x64 targets to cmake presets (#7574)
kunnis May 27, 2024
852aafb
update HIP_UMA #7399 (#7414)
Djip007 May 27, 2024
74b239b
llava : update clip.h (#7580)
eltociear May 28, 2024
c417671
Markdownish code block fix (#7571)
nathan-sixnines May 28, 2024
9335b96
server: do not remove whitespace at the start of a completion chunk (…
mgroeber9110 May 28, 2024
0548a41
ggml : generalize GGML_OP_CONCAT (#7563)
ggerganov May 28, 2024
e2b0650
[SYCL]fix ggml_sycl_mul_mat_id() to match the change of api (#7436)
arthw May 28, 2024
271ff3f
github: add refactor to issue template (#7561)
mofosyne May 28, 2024
8b99e2a
llama : handle unknown utf8 bytes (#7588)
ggerganov May 28, 2024
edc2943
tests : fix test-tokenizer-0.sh
ggerganov May 28, 2024
ee3dff6
Add support for DeepseekV2ForCausalLM (#7519)
fairydreaming May 28, 2024
2b737ca
rpc : resource management rework (#7562)
rgerganov May 28, 2024
56411a9
vulkan: properly initialize vulkan devices for LLAMA_SPLIT_MODE_NONE …
Adriankhl May 28, 2024
8767ce2
Merge branch 'prepare-PR-of-minicpm-v2.5' into prepare-PR
tc-mb May 28, 2024
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
50 changes: 50 additions & 0 deletions .github/ISSUE_TEMPLATE/01-bug-low.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,50 @@
name: Low Severity Bugs
description: Used to report low severity bugs in llama.cpp (e.g. cosmetic issues, non critical UI glitches)
title: "Bug: "
labels: ["bug-unconfirmed", "low severity"]
body:
- type: markdown
attributes:
value: |
Thanks for taking the time to fill out this bug report!
Please include information about your system, the steps to reproduce the bug,
and the version of llama.cpp that you are using.
If possible, please provide a minimal code example that reproduces the bug.
- type: textarea
id: what-happened
attributes:
label: What happened?
description: Also tell us, what did you expect to happen?
placeholder: Tell us what you see!
validations:
required: true
- type: textarea
id: version
attributes:
label: Name and Version
description: Which executable and which version of our software are you running? (use `--version` to get a version string)
placeholder: |
$./main --version
version: 2999 (42b4109e)
built with cc (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0 for x86_64-linux-gnu
validations:
required: true
- type: dropdown
id: operating-system
attributes:
label: What operating system are you seeing the problem on?
multiple: true
options:
- Linux
- Mac
- Windows
- BSD
- Other? (Please let us know in description)
validations:
required: false
- type: textarea
id: logs
attributes:
label: Relevant log output
description: Please copy and paste any relevant log output. This will be automatically formatted into code, so no need for backticks.
render: shell
50 changes: 50 additions & 0 deletions .github/ISSUE_TEMPLATE/02-bug-medium.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,50 @@
name: Medium Severity Bug
description: Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but generally still useable)
title: "Bug: "
labels: ["bug-unconfirmed", "medium severity"]
body:
- type: markdown
attributes:
value: |
Thanks for taking the time to fill out this bug report!
Please include information about your system, the steps to reproduce the bug,
and the version of llama.cpp that you are using.
If possible, please provide a minimal code example that reproduces the bug.
- type: textarea
id: what-happened
attributes:
label: What happened?
description: Also tell us, what did you expect to happen?
placeholder: Tell us what you see!
validations:
required: true
- type: textarea
id: version
attributes:
label: Name and Version
description: Which executable and which version of our software are you running? (use `--version` to get a version string)
placeholder: |
$./main --version
version: 2999 (42b4109e)
built with cc (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0 for x86_64-linux-gnu
validations:
required: true
- type: dropdown
id: operating-system
attributes:
label: What operating system are you seeing the problem on?
multiple: true
options:
- Linux
- Mac
- Windows
- BSD
- Other? (Please let us know in description)
validations:
required: false
- type: textarea
id: logs
attributes:
label: Relevant log output
description: Please copy and paste any relevant log output. This will be automatically formatted into code, so no need for backticks.
render: shell
50 changes: 50 additions & 0 deletions .github/ISSUE_TEMPLATE/03-bug-high.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,50 @@
name: High Severity Bug
description: Used to report high severity bugs in llama.cpp (e.g. Malfunctioning features hindering important common workflow)
title: "Bug: "
labels: ["bug-unconfirmed", "high severity"]
body:
- type: markdown
attributes:
value: |
Thanks for taking the time to fill out this bug report!
Please include information about your system, the steps to reproduce the bug,
and the version of llama.cpp that you are using.
If possible, please provide a minimal code example that reproduces the bug.
- type: textarea
id: what-happened
attributes:
label: What happened?
description: Also tell us, what did you expect to happen?
placeholder: Tell us what you see!
validations:
required: true
- type: textarea
id: version
attributes:
label: Name and Version
description: Which executable and which version of our software are you running? (use `--version` to get a version string)
placeholder: |
$./main --version
version: 2999 (42b4109e)
built with cc (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0 for x86_64-linux-gnu
validations:
required: true
- type: dropdown
id: operating-system
attributes:
label: What operating system are you seeing the problem on?
multiple: true
options:
- Linux
- Mac
- Windows
- BSD
- Other? (Please let us know in description)
validations:
required: false
- type: textarea
id: logs
attributes:
label: Relevant log output
description: Please copy and paste any relevant log output. This will be automatically formatted into code, so no need for backticks.
render: shell
50 changes: 50 additions & 0 deletions .github/ISSUE_TEMPLATE/04-bug-critical.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,50 @@
name: Critical Severity Bug
description: Used to report critical severity bugs in llama.cpp (e.g. Crashing, Corrupted, Dataloss)
title: "Bug: "
labels: ["bug-unconfirmed", "critical severity"]
body:
- type: markdown
attributes:
value: |
Thanks for taking the time to fill out this bug report!
Please include information about your system, the steps to reproduce the bug,
and the version of llama.cpp that you are using.
If possible, please provide a minimal code example that reproduces the bug.
- type: textarea
id: what-happened
attributes:
label: What happened?
description: Also tell us, what did you expect to happen?
placeholder: Tell us what you see!
validations:
required: true
- type: textarea
id: version
attributes:
label: Name and Version
description: Which executable and which version of our software are you running? (use `--version` to get a version string)
placeholder: |
$./main --version
version: 2999 (42b4109e)
built with cc (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0 for x86_64-linux-gnu
validations:
required: true
- type: dropdown
id: operating-system
attributes:
label: What operating system are you seeing the problem on?
multiple: true
options:
- Linux
- Mac
- Windows
- BSD
- Other? (Please let us know in description)
validations:
required: false
- type: textarea
id: logs
attributes:
label: Relevant log output
description: Please copy and paste any relevant log output. This will be automatically formatted into code, so no need for backticks.
render: shell
51 changes: 51 additions & 0 deletions .github/ISSUE_TEMPLATE/05-enhancement.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,51 @@
name: Enhancement
description: Used to request enhancements for llama.cpp
title: "Feature Request: "
labels: ["enhancement"]
body:
- type: markdown
attributes:
value: |
[Please post your idea first in Discussion if there is not yet a consensus for this enhancement request. This will help to keep this issue tracker focused on enhancements that the community has agreed needs to be implemented.](https://github.com/ggerganov/llama.cpp/discussions/categories/ideas)

- type: checkboxes
id: prerequisites
attributes:
label: Prerequisites
description: Please confirm the following before submitting your enhancement request.
options:
- label: I am running the latest code. Mention the version if possible as well.
required: true
- label: I carefully followed the [README.md](https://github.com/ggerganov/llama.cpp/blob/master/README.md).
required: true
- label: I searched using keywords relevant to my issue to make sure that I am creating a new issue that is not already open (or closed).
required: true
- label: I reviewed the [Discussions](https://github.com/ggerganov/llama.cpp/discussions), and have a new and useful enhancement to share.
required: true

- type: textarea
id: feature-description
attributes:
label: Feature Description
description: Please provide a detailed written description of what you were trying to do, and what you expected `llama.cpp` to do as an enhancement.
placeholder: Detailed description of the enhancement
validations:
required: true

- type: textarea
id: motivation
attributes:
label: Motivation
description: Please provide a detailed written description of reasons why this feature is necessary and how it is useful to `llama.cpp` users.
placeholder: Explanation of why this feature is needed and its benefits
validations:
required: true

- type: textarea
id: possible-implementation
attributes:
label: Possible Implementation
description: If you have an idea as to how it can be implemented, please write a detailed description. Feel free to give links to external sources or share visuals that might be helpful to understand the details better.
placeholder: Detailed description of potential implementation
validations:
required: false
38 changes: 38 additions & 0 deletions .github/ISSUE_TEMPLATE/06-question.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,38 @@
name: Question
description: Used to ask questions about llama.cpp
title: "Question: "
labels: ["question"]
body:
- type: markdown
attributes:
value: |
[Please search your question first in Discussion if you got a common general question.](https://github.com/ggerganov/llama.cpp/discussions/categories/q-a)

- type: checkboxes
id: prerequisites
attributes:
label: Prerequisites
description: Please confirm the following before submitting your question.
options:
- label: I searched using keywords relevant to my issue to make sure that I am creating a new issue that is not already open (or closed).
required: true
- label: I reviewed the [Discussions](https://github.com/ggerganov/llama.cpp/discussions), and have a new useful question to share that cannot be answered within Discussions.
required: true

- type: textarea
id: background-description
attributes:
label: Background Description
description: Please provide a detailed written description of what you were trying to do, and what you expected `llama.cpp` to do as an question.
placeholder: Detailed description of your question
validations:
required: true

- type: textarea
id: possible-answer
attributes:
label: Possible Answer
description: If you have some idea of possible answers you want to confirm, that would also be appreciated.
placeholder: Your idea of possible answers
validations:
required: false
28 changes: 28 additions & 0 deletions .github/ISSUE_TEMPLATE/07-refactor.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,28 @@
name: Refactor (Maintainers)
description: Used to track refactoring opportunities
title: "Refactor: "
labels: ["refactor"]
body:
- type: markdown
attributes:
value: |
Don't forget to [check for existing refactor issue tickets](https://github.com/ggerganov/llama.cpp/issues?q=is%3Aopen+is%3Aissue+label%3Arefactoring) in case it's already covered.
Also you may want to check [Pull request refactor label as well](https://github.com/ggerganov/llama.cpp/pulls?q=is%3Aopen+is%3Apr+label%3Arefactoring) for duplicates too.

- type: textarea
id: background-description
attributes:
label: Background Description
description: Please provide a detailed written description of the pain points you are trying to solve.
placeholder: Detailed description behind your motivation to request refactor
validations:
required: true

- type: textarea
id: possible-approaches
attributes:
label: Possible Refactor Approaches
description: If you have some idea of possible approaches to solve this problem. You may want to make it a todo list.
placeholder: Your idea of possible refactoring opportunity/approaches
validations:
required: false
11 changes: 0 additions & 11 deletions .github/ISSUE_TEMPLATE/bug.md

This file was deleted.

28 changes: 0 additions & 28 deletions .github/ISSUE_TEMPLATE/enhancement.md

This file was deleted.

14 changes: 13 additions & 1 deletion .github/labeler.yml
Original file line number Diff line number Diff line change
@@ -1,5 +1,16 @@
# https://github.com/actions/labeler

Kompute:
- changed-files:
- any-glob-to-any-file:
- ggml-kompute.h
- ggml-kompute.cpp
- README-kompute.md
Apple Metal:
- changed-files:
- any-glob-to-any-file:
- ggml-metal.h
- ggml-metal.cpp
- README-metal.md
SYCL:
- changed-files:
- any-glob-to-any-file:
Expand All @@ -9,6 +20,7 @@ SYCL:
Nvidia GPU:
- changed-files:
- any-glob-to-any-file:
- ggml-cuda.h
- ggml-cuda/**
Vulkan:
- changed-files:
Expand Down
Loading
Loading