Viet-Anh on Software Logo

Viet-Anh on Software

AI & Software Engineering

HomeAboutProjectsBlogNotes

What is: SwiGLU?

SourceGLU Variants Improve Transformer
Year2000
Data SourceCC BY-SA - https://paperswithcode.com

SwiGLU is an activation function which is a variant of GLU. The definition is as follows:

SwiGLU(x,W,V,b,c,β)=Swish_β(xW+b)⊗(xV+c)\text{SwiGLU}\left(x, W, V, b, c, \beta\right) = \text{Swish}\_{\beta}\left(xW + b\right) \otimes \left(xV + c\right)SwiGLU(x,W,V,b,c,β)=Swish_β(xW+b)⊗(xV+c)

Collections

Activation-Functions

Previous Term

RESCAL

Next Term

MLFPN
← Back to the glossary list
Viet-Anh on Software Logo

Viet-Anh on Software

AI & Software Engineering

Sharing insights, tutorials, and experiences in artificial intelligence, machine learning, and software development.

Navigation

  • Home
  • About
  • Projects
  • Blog
  • Notes

Content

  • Videos
  • Contact
  • Glossary

Resources

  • PageSpeed
© 2025 Viet-Anh Nguyen. All rights reserved.
Privacy PolicyTerms of ServicePageSpeed