NLP News
New NLP News - Conversational AI tutorial, RNNs for particle physics, InfoGAN, NLP Coursera, NLP book, killer robots, Code2pix, Google AI principles, relational reasoning https://t.co/5zT3QpjYAT
— Sebastian Ruder (@seb_ruder) June 11, 2018
NLP Transfer Learning
OpenAI
What I've been working on for the past year! https://t.co/CAQMYS1rR7
— Alec Radford (@AlecRad) June 11, 2018
Inspired by CoVE, ELMo, and ULMFiT we show that a single transformer language model can be finetuned to a wide variety of NLP tasks and performs very well with little tuning/tweaking.
This is exactly where we were hoping our ULMFit work would head - really great work from @OpenAI! 😊
— Jeremy Howard (@jeremyphoward) June 11, 2018
If you're doing NLP and haven't tried language model transfer learning yet, then jump in now, because it's a Really Big Deal. https://t.co/0Dj8ChCxvu
Great work! Language models serving as the basis for transfer learning in task agnostic NLP is really going to feed the next generation of tools and transformations =]
— Smerity (@Smerity) June 11, 2018
Google Brain
We show large language models trained on massive text corpora (LM1b, CommonCrawl, Gutenberg) can be used for commonsense reasoning and obtain SOTA on Winograd Schema Challenge. Paper at https://t.co/aRndlByWfj, results reproducible at https://t.co/jFOmUYf03O pic.twitter.com/s3uyrksAQz
— Trieu H. Trinh (@thtrieu_) June 11, 2018
Terminology
(Bonus)
did terminology move and i didn't notice? do you consider language modeling (training a probabilistic model p(w1 w2 w3 ... wN)) to be unsupervised learning?
— Hal Daumé III (@haldaume3) June 12, 2018
Probabilistic FastText for Multi-Sense Word Embeddings
Our new paper, Probabilistic FastText for Multi-Sense Word Embeddings, is appearing as an oral at #ACL2018, with code! https://t.co/WoYzYLdd5d
— Andrew Gordon Wilson (@andrewgwils) June 11, 2018
We learn density embeddings that account for sub-word structure and multiple senses. Joint with Ben Athiwaratkun and @AnimaAnandkumar! pic.twitter.com/Xl1MkzJIbt
Graph CNN for Web-Scale Recommender
Graph Convolutional Neural Networks for Web-Scale Recommender Systems. Our new #KDD2018 paper with @PinterestEng. https://t.co/rvFhja3PvO
— Jure Leskovec (@jure) June 11, 2018
"We deploy PinSage at @Pinterest and train it on 7.5 billion examples on a graph with 3 billion nodes representing pins and boards, and 18 billion edges" https://t.co/woN0ZY769t
— Monica Rogati (@mrogati) June 11, 2018
Photo Cartoonizing
CartoonGAN: Generative Adversarial Networks for Photo Cartoonization (CVPR 2018) https://t.co/xISVUHNbha pic.twitter.com/NfEsbMiQzZ
— hardmaru (@hardmaru) June 12, 2018
Notable Research
Looking for a piece about autonomous weapon systems that manages to link Shakespeare, war dogs, and swarming drones? Looks no further - "Autonomous Weapon Systems and the Limits of Analogy" was published today!https://t.co/zfMqKIypeL
— Rebecca Crootof (@RebeccaCrootof) June 11, 2018
Code, pre-trained models and paper: "QuaterNet: A Quaternion-based Recurrent Model for Human Motion" by Dario Pavllo, David Grangier and Michael Auli.
— Yann LeCun (@ylecun) June 12, 2018
Code: https://t.co/8pEGUAKNeZ
Paper: https://t.co/YA5etbCJdW https://t.co/J9MZ3NvyTE
Excited to share our recent work showing that pooling is neither necessary nor sufficient for appropriate deformation stability in CNNs! Rather, filter smoothness most directly modulates deformation stability in networks both *with* and *without* pooling. https://t.co/nJvvyGenV0 pic.twitter.com/UAmuXbVXbZ
— Ari Morcos (@arimorcos) June 6, 2018
Very interesting paper that uses the Frobenius norm of the difference of the trained weights and their init in a Rademacher complexity bound. Can explain the counter intuitive improved generalization of wider nets. https://t.co/vWWh3M4lit
— Olivier Grisel (@ogrisel) June 11, 2018
Tutorials and Resources
Deep Learning Research Review: Natural Language Processing https://t.co/aoBz43ZSMQ pic.twitter.com/bv4gYaBzlh
— KDnuggets (@kdnuggets) June 11, 2018
Slides from my talk earlier using Atari/Go as case studies of recent AI progress. See our recent paper (my pinned tweet) and my blog posts for more on these themes. https://t.co/3nHjq5QtJn
— Miles Brundage (@Miles_Brundage) June 11, 2018
Video here: https://t.co/He9FxnJtYr (orange slides)
Thx @azeem et al. for hosting!
I'm laughing so hard at this slide a friend sent me from one of Geoff Hinton's courses;
— Robbie Barrat (@DrBeef_) June 10, 2018
"To deal with hyper-planes in a 14-dimensional space, visualize a 3-D space and say 'fourteen' to yourself very loudly. Everyone does it." pic.twitter.com/nTakZArbsD
👍: the updated @rawgraphs comes with bee swarm and contour plots https://t.co/HC0jzilvZv pic.twitter.com/diB5dM7rsh
— Maarten Lambrechts (@maartenzam) June 11, 2018
Kernel Trick https://t.co/eZ2bbpDzwV pic.twitter.com/eqAJ8IRGna
— Chris Albon (@chrisalbon) June 11, 2018
Description of 4th place and fastest solution in @Lyft Perception Challenge with #PyTorch code https://t.co/kLsZoV2CUg
— Alexandr Kalinin (@alxndrkalinin) June 11, 2018
- simulated data for semantic segmentation
- hardware limitations
- LinkNet34
- postprocessing on GPU
- speed-up tricks to get >20FPS on single K80 pic.twitter.com/7XaMzu5uZv
rstats
Always super helpful: "Basic Regular Expressions in R" by Ian Kopacka https://t.co/WyfezkISPh #RegEx #rstats (#SoDS18 clutch @kierisi) pic.twitter.com/6UY96B8rVu
— Mara Averick (@dataandme) June 11, 2018
Key difference between rbind and dplyr's bind_rows: One refuses to add when columns aren't matching, the other one looks for where the match is (with warnings for missing columns). Depending on the scenario, either one can come in handy #rstats #DSLearning pic.twitter.com/PQyNsT2JeY
— Suzan Baert (@SuzanBaert) June 11, 2018
ICYMI, 👍 code-through:
— Mara Averick (@dataandme) June 11, 2018
"Supervised vs. Unsupervised Learning: Exploring Brexit w/ PLS & PCA" by @gokhan_ciflikli https://t.co/iGEZCxFLIu #rstats #caret #tmap pic.twitter.com/jx902RUjBl
Tensorflow Serving 1.8
TensorFlow Serving has released version 1.8.0 & now supports a RESTful API out of the box! You can now easily make classification, regression & prediction requests to your TensorFlow models using JSON objects.
— TensorFlow (@TensorFlow) June 11, 2018
Learn more here ↓ https://t.co/BtvgZLi2bb
Datasets
New #dataset of 5.5 million tweets covers the suicides of Anthony Bourdain and Kate Spade and covers a period of a few days before Kate's suicide to a few days after Bourdain's. Description file ends with .txt and gives details. 30 GB.https://t.co/jaF7rxXGSJ#datascience
— Jason Baumgartner (@jasonbaumgartne) June 11, 2018
Description file located here: https://t.co/vVUZTWYMQY
— Jason Baumgartner (@jasonbaumgartne) June 11, 2018
ICYMI last week we released the largest git dataset in the history of #MLonCode!
— source{d} (@sourcedtech) June 11, 2018
Repos contain objects and metadata for all existing revisionshttps://t.co/Aa5Qq8p6bf
Miscellaneous
The General Video Game AI (@gvgai) competition lets you compete not only in AI for playing games, but also for making them!
— Julian Togelius (@togelius) June 11, 2018
The Level Generation track challenges you to create level generators:https://t.co/DLrWA1yJZy
The Rule Generation track:https://t.co/sAD6vkgMdW pic.twitter.com/PgUqY2G3qB
Nice talk from @karpathy about the challenges of building the Software 2.0 stack. https://t.co/n4Z4bNhtuG pic.twitter.com/S8nPitGlWt
— hardmaru (@hardmaru) June 11, 2018
my thoughts on the trouble with #d3js https://t.co/1JfLFHl1Lt
— Ian Johnson (@enjalot) June 12, 2018
to me it's about setting our expectations for learning, and guiding folks on different paths in different ways pic.twitter.com/pZ67Krs0CX
I'm happy to report that hyper-intelligent AI isn't that close: currently Amazon, Microsoft, and Google's cutting-edge APIs all believe "My leg was eaten by a great white shark" is a sentence with a highly positive sentiment. pic.twitter.com/rWSvyek9zs
— Jonathan Nolis (@skyetetra) June 11, 2018
Prioritize AI dangers:
— Aurélien Geron (@aureliengeron) June 11, 2018
- Important and urgent: fight mass manipulation
- Important but not urgent: fight Terminator
- Not important but urgent: stop using stock pictures of robots in every &$!#% article about AI
- Not important, not urgent: stop calling 3-layer neural nets "deep"
When the AI bubble bursts by John Langfordhttps://t.co/8qIKLUyo4K
— Giorgio Patrini (@GiorgioPatrini) June 11, 2018
"What fraction of the field is currently in units at companies which are very expensive without yet justifying that expense? It’s no longer a small fraction so there is a chance for something traumatic"
In 2012 my wife and I published a paper using text-mining of neuroscience papers to generate hypotheses.
— Brad Voytek (@bradleyvoytek) June 11, 2018
One prediction was a migraine/striatum link. 3-fold increase in publications since that prediction.
Small-n... need to extend. But neat first-pass!https://t.co/IVNv757trp pic.twitter.com/1norX4OidM
'Luck is statistics taken personally.' -- Penn Jillette
— Data Science Fact (@DataSciFact) June 11, 2018