• Thumbnail for Transformer (deep learning architecture)
    A transformer is a deep learning architecture developed by Google and based on the multi-head attention mechanism, proposed in a 2017 paper "Attention...
    65 KB (8,139 words) - 00:09, 1 June 2024
  • Mamba is a deep learning architecture focused on sequence modeling. It was developed by researchers from Carnegie Mellon University and Princeton University...
    12 KB (1,254 words) - 10:13, 25 April 2024
  • paper "Attention Is All You Need" which introduced the Transformer model, a novel architecture that uses a self-attention mechanism and has since become...
    7 KB (542 words) - 15:45, 31 March 2024
  • Thumbnail for Deep learning
    supervised, semi-supervised or unsupervised. Deep-learning architectures such as deep neural networks, deep belief networks, recurrent neural networks,...
    177 KB (17,587 words) - 05:50, 27 May 2024
  • a {\displaystyle W_{a}} is a learnable weight matrix. Transformer (deep learning architecture) § Efficient implementation Rumelhart, David E.; Mcclelland...
    28 KB (2,207 words) - 17:49, 13 May 2024
  • Thumbnail for Generative pre-trained transformer
    used in natural language processing tasks. GPTs are based on the transformer architecture, pre-trained on large data sets of unlabelled text, and able to...
    46 KB (4,098 words) - 18:01, 26 May 2024
  • Multimodal learning, in the context of machine learning, is a type of deep learning using multiple modalities of data, such as text, audio, or images....
    7 KB (1,696 words) - 14:24, 1 June 2024
  • An AI accelerator, deep learning processor, or neural processing unit (NPU) is a class of specialized hardware accelerator or computer system designed...
    53 KB (5,142 words) - 07:23, 24 May 2024
  • Thumbnail for Federated learning
    Things, and pharmaceuticals. Federated learning aims at training a machine learning algorithm, for instance deep neural networks, on multiple local datasets...
    51 KB (5,963 words) - 19:17, 13 May 2024
  • present in a photo that a human could easily spot. The transformer deep learning architecture was invented by Google Brain researchers in 2017, and explained...
    35 KB (3,833 words) - 11:04, 18 April 2024