Data Science by ODS.ai 🦜
46.1K subscribers
663 photos
77 videos
7 files
1.75K links
First Telegram Data Science channel. Covering all technical and popular staff about anything related to Data Science: AI, Big Data, Machine Learning, Statistics, general Math and the applications of former. To reach editors contact: @malev
加入频道
Neural nets are terrible at arithmetic & counting. If you train one in 1 to 10, it will do okay on 3 + 5 but fail miserably for 1000 + 3000. Resolving this, «Neural Arithmetic Logic Units» can track time, do arithmetic on images of numbers, & extrapolate, providing better results than other architectures.

https://arxiv.org/pdf/1808.00508.pdf

#nn #architecture #concept #deepmind #arithmetic
The Conversational Intelligence Challenge 2 (ConvAI second part) got announced today.

The aim of our competition is to establish a concrete scenario for testing chatbots that aim to engage humans, and become a standard evaluation tool in order to make such systems directly comparable.

The most promising team attending will receive a travel grant to attend #NIPS2018

#nlp #dl #dialoguesystem #competition
First 1e6 integers, represented as binary vectors indicating their prime factors, and laid out using the sparse matrix support in leland_mcinnes's UMAP dimensionality reduction algorithm. This is from a 1000000x78628 (!) binary matrix. Very pretty structure emerges.
"Learning Hierarchical Semantic Image Manipulation through Structured Representations":

https://arxiv.org/abs/1808.07535

#cv #dl
Online ad demand prediction #kaggle competition 1st place summary:
https://www.kaggle.com/c/avito-demand-prediction/discussion/59880

Winner explains how to combine categorical, numerical, image and text features into a single #NN that gets you into top 10 without stacking.
Tickets for one of the key conferences #NIPS2018 were sold out in 15 minutes.
NIPS Conference Registrations 2002 thru 2019.
[2018] War erupts for tickets
[2019] AI researchers discover time travel.

#NIPS2018
1st place solution in the recent Home Credit Default Risk #Kaggle competition

- extensive feature engineering, with ~700 of features total used
- XGBoost, LightGBM, CatBoost, FastRGF, DAE+NN, Lin Reg
- 3-level ensembling (stacking x2 + blending)

Link: https://www.kaggle.com/c/home-credit-default-risk/discussion/64821
#Google introduced Conceptual Captions, a new dataset and challenge for image captioning consisting of ~3.3 million image/caption pairs for the machine learning community to train and evaluate their own image captioning models.

Link: https://ai.googleblog.com/2018/09/conceptual-captions-new-dataset-and.html

#dataset