Open in app

Sign In

Write

Sign In

Mastodon
Naoki
Naoki

2.6K Followers

Home

Lists

About

Aug 23

VAE: Variational Auto-Encoder (2013)

Understanding the Auto-Encoding Variational Bayes Paper — Ever stumbled upon the Auto-Encoding Variational Bayes paper and felt lost amid mathematical equations and concepts like Auto-Encoder, Bayesian Theorem, Variational Inference, and Deep Learning? Understanding this groundbreaking work is a challenge but a rewarding journey that unveils new perspectives in machine learning. Variational Auto-Encoders, at the core of this…

Artificial Intelligence

29 min read

VAE: Variational Auto-Encoder (2013)
VAE: Variational Auto-Encoder (2013)
Artificial Intelligence

29 min read


Aug 13

CLIP: Learning Transferable Visual Models From Natural Language Supervision (2021)

Bridging the Gap Between Vision and Language — A Look at OpenAI’s CLIP Model — Deep learning vision models traditionally relied on vast collections of labeled images, each tailored to recognize objects in a specific category or class. OpenAI’s approach, using images and natural language, offers an alternative that doesn’t necessitate such tailored examples. They developed a CLIP model that can recognize objects without needing…

Artificial Intelligence

9 min read

Learning Transferable Visual Models From Natural Language Supervision (2021)
Learning Transferable Visual Models From Natural Language Supervision (2021)
Artificial Intelligence

9 min read


Apr 30

ICL: Why Can GPT Learn In-Context? (2022)

Why Can GPT Learn In-Context? Language Models Secretly Perform Gradient Descent as Meta-Optimizers — The paper Why Can GPT Learn In-Context? Language Models Secretly Perform Gradient Descent as Meta-Optimizers provides insights into how GPT-3 can learn from a few demonstrations and predict labels for unseen inputs. This ability is known as In-Context Learning (ICL). The paper explains how GPT-3 performs ICL as follows:

Artificial Intelligence

2 min read

ICL: Why Can GPT Learn In-Context? (2022)
ICL: Why Can GPT Learn In-Context? (2022)
Artificial Intelligence

2 min read


Jan 12

Hands-on Deep Learning with PyTorch

A Series of Video Lectures on YouTube — I’ve created a short course on deep learning with PyTorch. It contains a series of videos to cover various topics. I keep each video short and straight to the point. I hope you enjoy it! An overview of the lecture series Prerequisites to studying deep learning PyTorch Installation Tensors NumPy arrays (hands-on) PyTorch tensors (hands-on) Torchvision datasets (hands-on)

Pytorch

1 min read

Hands-on Deep Learning with PyTorch
Hands-on Deep Learning with PyTorch
Pytorch

1 min read


Jan 4

GPT-3: In-Context Few-Shot Learner (2020)

Language Models are Few-Shot Learners — #GPT #Transformer In 2020, OpenAI announced GPT-3, a generative language model with 175 billion parameters, 10x more than any previous language model, and published its performance on NLP benchmarks. However, it wasn’t just another size upgrade. GPT-3 showed the improved capability to handle tasks purely via text interaction. Those tasks…

Artificial Intelligence

7 min read

GPT-3: In-Context Few-Shot Learner (2020)
GPT-3: In-Context Few-Shot Learner (2020)
Artificial Intelligence

7 min read


Dec 30, 2022

GPT-2: Too Dangerous To Release (2019)

Language Models are Unsupervised Multitask Learners — #GPT #Transformer GPT-2 is a direct scale-up of GPT-1, with more parameters and trained on more data. However, it was deemed to be too dangerous to release by OpenAI: Due to our concerns about malicious applications of the technology, we are not releasing the trained model. As an experiment in…

Artificial Intelligence

4 min read

GPT-2: Too Dangerous To Release (2019)
GPT-2: Too Dangerous To Release (2019)
Artificial Intelligence

4 min read


Dec 26, 2022

GPT (2018)

Generative Pre-Trained Transformer — #GPT #Transformer In 2018, OpenAI released the first version of GPT (Generative Pre-Trained Transformer) for generating texts as if humans wrote. The architecture of GPT is based on the original transformer’s decoder. They trained GPT in two stages: Unsupervised Pre-training pre-trains GPT on unlabeled text, which taps into abundant text…

Artificial Intelligence

6 min read

GPT (2018)
GPT (2018)
Artificial Intelligence

6 min read


Nov 27, 2022

Longformer (2020)

The Long-Document Transformer — In 2020, researchers at Allen Institute for Artificial Intelligence (AI2) published “Longformer: The Long-Document Transformer”. AI2 is a non-profit research organization that hosts the Semantic Scholar website providing AI-driven search and discovery tools for research publications. …

Artificial Intelligence

6 min read

Longformer(2020)
Longformer(2020)
Artificial Intelligence

6 min read


Nov 4, 2022

Swin Transformer (2021)

Hierarchical Vision Transformer using Shifted Windows — In 2021, Microsoft announced a new Vision Transformer called Swin Transformer, which can act as a backbone for computer vision tasks like image classification, object detection, and semantic segmentation. The word Swin stands for Shifted windows that provide the Transformer with hierarchical vision, which is the main topic of this…

Artificial Intelligence

8 min read

Swin Transformer (2021)
Swin Transformer (2021)
Artificial Intelligence

8 min read


Nov 2, 2022

ViT: Vision Transformer (2020)

An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale — In 2020, the Google Brain team developed Vision Transformer (ViT), an image classification model without a CNN (convolutional neural network). ViT directly applies a Transformer Encoder to sequences of image patches for classification. This article explains how ViT works. Vision Transformer Architecture Vision Transformer Overview

Artificialintelligenceai

5 min read

ViT: Vision Transformer (2020)
ViT: Vision Transformer (2020)
Artificialintelligenceai

5 min read

Naoki

Naoki

2.6K Followers

Solopreneur @ kikaben.com

Following
  • Synced

    Synced

  • Sik-Ho Tsang

    Sik-Ho Tsang

  • Jesus Rodriguez

    Jesus Rodriguez

  • Qiskit

    Qiskit

  • Ms Aerin

    Ms Aerin

See all (10)

Help

Status

Writers

Blog

Careers

Privacy

Terms

About

Text to speech

Teams