Leo Boytsov (@srchvrs) 's Twitter Profile
Leo Boytsov

@srchvrs

Sr. Research Scientist @AWS Labs (ph-D @LTIatCMU) working on (un)natural language processing, speaking πtorch & C++. Opinions sampled from MY OWN 100T param LM.

ID: 87473622

linkhttp://searchivarius.org/about calendar_today04-11-2009 16:22:33

24,24K Tweet

8,8K Followers

1,1K Following

Leo Boytsov (@srchvrs) 's Twitter Profile Photo

""Whenever someone asks me if RL works, I say it doesn't and 70% of the time I am right." alexirpan.com/2018/02/14/rl-…

Leo Boytsov (@srchvrs) 's Twitter Profile Photo

This is a rather blockbuster piece of news: the Hugging Face library is dropping support for both Jax and Tensorflow. linkedin.com/posts/lysandre…

Leo Boytsov (@srchvrs) 's Twitter Profile Photo

I find it quite upsetting that people use the terms LLM, or PLM (pretrained LM) to denote every possible pre-trained Transformer model including BERT, T5, etc... This reminds me good old days when DNN was often used to denote a multilayer feedforward network (or at least this was

ACLRollingReview (@reviewacl) 's Twitter Profile Photo

Dear ACL community, We are seeking emergency reviewers for the May cycle. Please indicate your availability (ASAP) if you can help review extra papers urgently (by the 24th of June AOE). Many thanks!

Leo Boytsov (@srchvrs) 's Twitter Profile Photo

This is something I have been suspecting for a while, but I do not think we have a lot of supporting evidence though. Anyways, it looks like efficiency of generation plays an important role in user's perception of quality.