Skip to content

Add RMS norm and use it#187

Merged
ggerganov merged 2 commits intomasterfrom
rms_norm
Mar 15, 2023
Merged

Add RMS norm and use it#187
ggerganov merged 2 commits intomasterfrom
rms_norm

Conversation

@hoangmit
Copy link
Contributor

To resolve issue #173

@hoangmit hoangmit changed the title Add Rms norm Add RMS norm and use it Mar 15, 2023
Copy link
Member

@ggerganov ggerganov left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

🦙

@ggerganov ggerganov merged commit 6eac39b into master Mar 15, 2023
@ggerganov ggerganov deleted the rms_norm branch March 15, 2023 22:42
blackhole89 pushed a commit that referenced this pull request Mar 15, 2023
* add ggml_rms_norm

* update op num
Deadsg pushed a commit to Deadsg/llama.cpp that referenced this pull request Dec 19, 2023
SamuelOliveirads pushed a commit to SamuelOliveirads/llama.cpp that referenced this pull request Dec 29, 2025
* iq1_m_r4: basics (quantize/dequantize)

* iq1_m_r4: Zen4 gemm

* iq1_m_r4: neon gemm

* iq1_m_r4: switch to q8_0_x4 also on AVX2/Zen4

With the deltas being per group of 8, we cannot make use
of the q8 sums stored in q8_1, so we get a tiny gain by
using q8_0_x4.

* iq1_m_r4: rename mul_mat_iq1_m_r4_q8_1 to mul_mat_iq1_m_r4_q8_0

---------

Co-authored-by: Iwan Kawrakow <iwan.kawrakow@gmail.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants