Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merge from upstream #31

Merged
merged 38 commits into from
Aug 8, 2024
Merged
Changes from 1 commit
Commits
Show all changes
38 commits
Select commit Hold shift + click to select a range
b72c20b
Fix conversion of unnormalized BF16->BF16 weights (#7843)
CISC Aug 2, 2024
76614f3
ggml : reading the runtime sve config of the cpu (#8709)
jdomke Aug 3, 2024
4b77ea9
flake.lock: Update (#8847)
ggerganov Aug 4, 2024
01aae2b
baby-llama : remove duplicate vector include
danbev Aug 3, 2024
ecf6b7f
batched-bench : handle empty `-npl` (#8839)
cunnie Aug 4, 2024
978ba3d
Server: Don't ignore llama.cpp params (#8754)
ardfork Aug 4, 2024
0d6fb52
Install curl in runtime layer (#8693)
bsquizz Aug 4, 2024
c02b0a8
cann: support q4_0 model (#8822)
wangshuai09 Aug 5, 2024
655858a
ggml : move c parameter comment to ggml_rope_ext (ggml/901)
danbev Jul 29, 2024
a3738b2
vulkan : implement Stable Diffusion operators (ggml/904)
0cc4m Aug 4, 2024
5587e57
sync : ggml
ggerganov Aug 4, 2024
064cdc2
vulkan : fix Qantized Mat-Vec Mul on AMD GPUs for ncols < 64 (#8855)
0cc4m Aug 5, 2024
f1ea514
llama : better replace_all (#8852)
ggerganov Aug 5, 2024
400ae6f
readme : update model list (#8851)
BarfingLemurs Aug 5, 2024
e31a4f6
cmake: fix paths for vulkan shaders compilation on Windows (#8573)
stduhpf Aug 5, 2024
d3f0c71
Stop the generation when <|eom_id|> token is encountered - needed for…
fairydreaming Aug 5, 2024
1ef14b3
py: Add more authorship metadata from model card (#8810)
mofosyne Aug 5, 2024
b9dfc25
ggml : fix overflows in elu function (#8866)
jart Aug 5, 2024
b42978e
readme : add ramalama to the availables UI (#8811)
ericcurtin Aug 5, 2024
bc0f887
cann: fix buffer_num and runtime speed slowly error (#8865)
wangshuai09 Aug 5, 2024
0a4ce78
common : Changed tuple to struct (TODO fix) (#8823)
Septa2112 Aug 5, 2024
d4ff847
[SYCL] correct cmd name (#8877)
arthw Aug 6, 2024
c21a896
[CANN]: Fix ggml_backend_cann_buffer_get_tensor (#8871)
MengqingCao Aug 6, 2024
cdd1889
convert : add support for XLMRoberta embedding models (#8658)
iamlemec Aug 6, 2024
2d5dd7b
ggml : add epsilon as a parameter for group_norm (#8818)
MollySophia Aug 6, 2024
0bf16de
contributing : add note about write access
ggerganov Aug 6, 2024
efda90c
[Vulkan] Fix compilation of `vulkan-shaders-gen` on w64devkit after `…
MaggotHATE Aug 6, 2024
db20f50
cmake : Link vulkan-shaders-gen with pthreads (#8835)
Patater Aug 6, 2024
5f4dcb1
simple : update name of executable to llama-simple (#8885)
danbev Aug 6, 2024
641f5dd
CUDA: fix padding logic for FP16/FP32 (#8884)
JohannesGaessler Aug 6, 2024
1e6f655
server : add lora hotswap endpoint (WIP) (#8857)
ngxson Aug 6, 2024
3195854
typo correction (#8891)
Nexesenex Aug 6, 2024
725e3d9
quantize : update usage comment in quantize.cpp (#8889)
danbev Aug 6, 2024
506122d
llama-bench : add support for getting cpu info on Windows (#8824)
kylo5aby Aug 7, 2024
a8dbc6f
CUDA/HIP: fix tests/test-backend-ops (#8896)
JohannesGaessler Aug 7, 2024
0478174
[SYCL] Updated SYCL device filtering (#8901)
OuadiElfarouki Aug 7, 2024
be55695
ggml-backend : fix async copy from CPU (#8897)
slaren Aug 7, 2024
15fa07a
make : use C compiler to build metal embed object (#8899)
slaren Aug 7, 2024
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Prev Previous commit
Next Next commit
ggml : fix overflows in elu function (ggml-org#8866)
It's helpful to use expm1f(x), because expf(x)-1 will result in overflow
for 25% of single-precision floating point numbers.
  • Loading branch information
jart authored Aug 5, 2024
commit b9dfc25ca385a83bde9e9456c4d4fae15377bc7b
2 changes: 1 addition & 1 deletion ggml/src/ggml.c
Original file line number Diff line number Diff line change
@@ -2312,7 +2312,7 @@ inline static void ggml_vec_abs_f32 (const int n, float * y, const float * x) {
inline static void ggml_vec_sgn_f32 (const int n, float * y, const float * x) { for (int i = 0; i < n; ++i) y[i] = (x[i] > 0.f) ? 1.f : ((x[i] < 0.f) ? -1.f : 0.f); }
inline static void ggml_vec_step_f32 (const int n, float * y, const float * x) { for (int i = 0; i < n; ++i) y[i] = (x[i] > 0.f) ? 1.f : 0.f; }
inline static void ggml_vec_tanh_f32 (const int n, float * y, const float * x) { for (int i = 0; i < n; ++i) y[i] = tanhf(x[i]); }
inline static void ggml_vec_elu_f32 (const int n, float * y, const float * x) { for (int i = 0; i < n; ++i) y[i] = (x[i] > 0.f) ? x[i] : expf(x[i])-1; }
inline static void ggml_vec_elu_f32 (const int n, float * y, const float * x) { for (int i = 0; i < n; ++i) y[i] = (x[i] > 0.f) ? x[i] : expm1f(x[i]); }
inline static void ggml_vec_relu_f32 (const int n, float * y, const float * x) { for (int i = 0; i < n; ++i) y[i] = (x[i] > 0.f) ? x[i] : 0.f; }
inline static void ggml_vec_leaky_relu_f32 (const int n, float * y, const float * x, const float ns) { for (int i = 0; i < n; ++i) y[i] = ((x[i] > 0.f) ? x[i] : 0.f) + ns * ((x[i] < 0.0f) ? x[i] : 0.f); }
inline static void ggml_vec_sigmoid_f32 (const int n, float * y, const float * x) { for (int i = 0; i < n; ++i) y[i] = 1.f / (1.f + expf(-x[i])); }