Team GoGL M.S. in Statistics & Data Science, Yonsei Trying to blend in with the A.I. madness

[Paper Review] Sequence to Sequence Learning with Neural Networks

Briefly introduce what Seq2seq is

[Paper Review] Text Understanding with the Attention Sum Reader Network

There are several large cloze-style context-question-answer datasets which was introduced 4~5 years ago - the CNN, Daily Mail news data, and the Children's Book Test. Thanks to the introduction to these large datasets, it became easier to associate text comprehension task to deep-learning techniques that seem to outperform all alternative approaches.

[CV] Few Shot Learning

Intro to Few Shot Learning

Pose Estimation - AlphaPose and its application

[AlphaPose](http://www.mvig.org/research/alphapose.html) is an accurate multi-person pose estimator, which is the first open-source system that achieves 70+ mAP (75 mAP) on COCO dataset and 80+ mAP (82.1 mAP) on MPII dataset.

[Paper Review] Neural Translation by Jointly Learning to Align and Translate (2015)

In this post, I aim to address potential problems with RNN based _neural machine translation (NMT)_ models and introduce a solution this paper proposed.

[DL 101] Understanding Bi-RNN/LSTM (Pytorch)

[Paper Review] Distributed Representations of Words and Phrases and their Compositionality

To improve the Vector Representation Quality of Skip-gram (one of the Word2Vec method). There are 4 ways to improve representation quality and computational efficiency.

[Paper Review] Efficient Estimation of Word Representations in Vector Space

How to make text to be input in deep learning?

[NLP] Google Duplex - 'Jarvis' comes true?

Google Duplex is a name for the technology supporting Google Assistant. This service was first introduced in 2018 and has been mainly used in _booking_ with human-like phone calls.

[Paper Review] U-Net: Convolutional Networks for Biomedical Image Segmentation

In Biomedical field, the instance segmentation are frequently used such as detecting tumors based on radiography, lesion segmentation, etc. What is important here, in biomedical data, is that the output should include localization. Let's look at what is U-Net and how it works

Web-crawling using Python

Today, we are going to know how to crawl iherb using Python, especially information about supplements.

[Paper Review] YOLO (You Look Only Once)

See what is YOLO and how it works

[DL 101] Activation Functions

Today we will look at different activation functions, especially the family of **ReLU (Rectified Linear Unit)** activation function. The role of activation functions in neural networks is taking the input and mapping it into output that goes into next layer. Deciding which activation function to use heavily depends on the target.

[Paper Review] Faster R-CNN

Fast R-CNN starts from the idea: What if using convolutional feature maps for generating region proposals?

[Paper Review] Fast R-CNN

Briefly introduce one of the most important object detection papers

[NLP] Big Bird

In this posting, we will deal with the NLP model named Big Bird. As you can see in the title, it is the model for _longer_ sequences. Let's take a brief look at the concept of Big Bird.

[NLP] Multi-Task Learning & MT-DNN

In this posting, I will talk about **Multi-Task Learning (MTL)** briefly and then deal with the NLP model named MT-DNN.

[Paper Review] Sparse Transformer

I will focus on the core part of this paper (=Factorized self-attention) and briefly mention rest of the stuffs.

[CV] One Shot learning

What is One Shot Learning? examples with face recognition

[AD] Introducing Anomaly Detection

Today, we will look at the basic approaches to anomaly detection.

[DL 101] OpenCV python tutorial

Take baby steps towards the Computer Vision master

[Paper Review] R-CNN

Today we are going to cover some important papers about object detection using deep learning architecture.

[NLP] NLP tasks

In this posting, we will take a quick look at the NLP tasks that I picked for explanation.

[NLP] Code - BERT + KMeans

Text Clustering - BERT + Dimension reduction + KMeans

Useful methods for CV competition

From SWA to TTA

[NLP] How does BERT work?

BERT is a pre-trained model released by Google in 2018, and has been used a lot so far, showing the highest performance in many NLP tasks. In this post, let's learn more about the detailed structure of BERT.

[DL 101] RAdam - a novel variant of Adam

Rectified Adaptive Learning Rate (RAdam)

[DL 101] Learning Rate Scheduler

What is learning rate scheduler?

[RecBasics] Negative samples in Top-K recommender task (Part 1)

In this post, we will dive into negative sampling for test / train instances in recommendation system.

[Paper Review] Attention is all you need

In this posting, we will review a paper titled: Attention is all you need, which introduces the ttention mechanism and Transformer structure that are still widely used in NLP and other fields. BERT, which was covered in the last posting, is the typical NLP model using this attention mechanism and Transformer. Although Attention and Transformer are actively used in NLP, they are also used in many areas where recurrent methods were used. From now on, let's take a closer look at what Attention and Transformer are.

[NLP] Difficulties in applying NLP models to Korean data

Although most of the NLP models already offered a pre-trained model for multilingual data, it is still difficult to put it directly into Korean. Korean is a complex language, so there are many aspects that the Tokenizer used here does not fit well. No matter how well pre-training has done, the performance will be terrible if you don't make subword vocab well.

[DL 101] Maxpooling? Convolutions?

Can we replace maxpooling with convolutional layers?

[DL 101] Upsampling

In many cases including image segmentation, a model consists of downsampling and upsampling parts and the latter restore the feature map to the input sized image. There are two types of upsampling using torch: UpSampling, ConvTranspose2D.

[DL 101] Autoencoder Tutorial (Pytorch)

Today we are going to build a simple autoencoder model using pytorch. We'll flatten CIFAR-10 dataset vectors then train the autoencoder with these flattened data.

[Jump into ML] Overview of Cross-validation

Cross-validation is one of the most popular methods to evaluate model performance and tune model parameters. Like the bootstrap, it belongs to a family of Monte Carlo methods. Today, we will go over several types of CV methods.

[NLP] Overview of NLP

There are countless ways to perform NLP, and the flow of methodologies is changing very quickly. Therefore, it is important to understand the latest trends in NLP. Since it is difficult to handle each method in a single posting, this posting will cover only the overview of NLP, and then individual models will be covered in the subsequent postings in more detail.

[DL 101] Transfer Learning vs. Fine Tuning vs. Training from scratch

What is transfer learning and why?

[DL 101] List comprehension

Let's get familiar with list comprehension, instead of for loops

[DL 101] Early Stopping, Weight Decay

regularization methods: early stopping, weight decay

[DL 101] Object Recognition Terminology

What is object localization, object detection, semantic segmentation and instance segmentation?

[DL 101] Global Average Pooling

Alternatives to the Fully Connected Layer(FC layer)

Fundamentals of the Recommendation System

Fundamental recommendation system models, content-based, collaborative-filtering and knowledge-based

[Paper Review] Hybrid Recommender Systems: A Systematic Literature Review (2017)

Hybrid method, as the name suggests, is a mixture of methodologies such as CBF and CF.

[Paper Review] Wide & Deep Learning for Recommender Systems (2016)

Today, I will review a paper that first integrated the idea of latent variables with deep learning.

Welcome to GoGL's blog :)

Let me tell you who we are