Deep Learning is an interesting new branch of machine learning where neural networks consisting of multiple layers have shown new generalization capabilities. The seminar will look at advances in both general deep learning approaches, and at the specific case of Neural Machine Translation (NMT). NMT is a new paradigm in data-driven machine translation. In Neural Machine Translation, the entire translation process is posed as an end-to-end supervised classification problem, where the training data is pairs of sentences and the full sequence to sequence task is handled in one model.
Here is a link to last semester's seminar.
There is a Munich interest group for Deep Learning, which has an associated mailing list, the paper announcements are sent out on this list. See the link here.
Email Address: SubstituteLastName@cis.uni-muenchen.de
CIS, LMU Munich
CIS, LMU Munich
Thursdays 14:45 (s.t.), location ZOOM ONLINE (was room 115)
You can install the zoom client or click cancel and use browser support (might not work for all browsers).
Contact Alexander Fraser if you need the zoom link.
New attendees are welcome. Read the paper and bring a paper or electronic copy with you, you will need to refer to it during the discussion.
Click here for directions to CIS.
If this page appears to be out of date, use the refresh button of your browser
|May 7th||Jonathan Clark, Eunsol Choi, Michael Collins et al. (2020). TyDi QA: A Benchmark for Information-Seeking Question Answering in Typologically Diverse Languages. TACL 2020.||paper||Denis Peskov|
|May 14th||Alex Tamkin, Trisha Singh, Davide Giovanardi, Noah Goodman (2020). Investigating Transferability in Pretrained Language Models. arXiv.||paper||Antonis Maronikolakis|
|June 18th||Automated Machine Learning (AutoML). See abstract for reading.||abstract||Pranav Rai|
|June 25th||Tom B. Brown, Benjamin Mann, [Open AI Team] (2020). Language Models are Few-Shot Learners. arXiv 2020.||paper||Hinrich Schütze|
|July 16th||Zirui Wang, Jiateng Xie, Ruochen Xu, Yiming Yang, Graham Neubig, Jaime Carbonell (2020). Cross-lingual Alignment vs Joint Training: A Comparative Study and A Simple Unified Framework. ICLR 2020.||paper||Viktor Hangya|
|July 23rd||ACL trip reports||Hinrich Schütze|
|July 30th||Francis Willett, Donald Avansino, Leigh Hochberg, Jaimie Henderson, Krishna Shenoy (2020). High-performance brain-to-text communication via imagined handwriting. bioRxiv 2020.||paper||Alex Fraser|
|August 6th||Kelvin Guu, Kenton Lee, Zora Tung, Panupong Pasupat, Ming-Wei Chang (2020). REALM: Retrieval-Augmented Language Model Pre-Training. arXiv 2020.||paper||Timo Schick|
You can go back through the previous semesters by clicking on the link near the top of the page.