Autor der Publikation

Embers of Autoregression: Understanding Large Language Models Through the Problem They are Trained to Solve

, , , , und . (2023)cite arxiv:2309.13638Comment: 50 pages plus 11 page of references and 23 pages of appendices.

Bitte wählen Sie eine Person um die Publikation zuzuordnen

Um zwischen Personen mit demselben Namen zu unterscheiden, wird der akademische Grad und der Titel einer wichtigen Publikation angezeigt. Zudem lassen sich über den Button neben dem Namen einige der Person bereits zugeordnete Publikationen anzeigen.

 

Weitere Publikationen von Autoren mit dem selben Namen

Embers of Autoregression: Understanding Large Language Models Through the Problem They are Trained to Solve., , , , und . CoRR, (2023)How poor is the stimulus? Evaluating hierarchical generalization in neural networks trained on child-directed speech., , , und . ACL (1), Seite 9370-9393. Association for Computational Linguistics, (2023)What do you learn from context? Probing for sentence structure in contextualized word representations., , , , , , , , , und 1 andere Autor(en). ICLR (Poster), OpenReview.net, (2019)Neurocompositional computing: From the Central Paradox of Cognition to a new generation of AI systems., , , , und . CoRR, (2022)Structural Biases for Improving Transformers on Translation into Morphologically Rich Languages., , , , , , , , , und 2 andere Autor(en). CoRR, (2022)Does syntax need to grow on trees? Sources of hierarchical inductive bias in sequence-to-sequence networks., , und . CoRR, (2020)BERTs of a feather do not generalize together: Large variability in generalization across models with similar test set performance., , und . CoRR, (2019)RNNs Implicitly Implement Tensor Product Representations., , , und . CoRR, (2018)RNNs implicitly implement tensor-product representations., , , und . ICLR (Poster), OpenReview.net, (2019)BERTs of a feather do not generalize together: Large variability in generalization across models with similar test set performance., , und . BlackboxNLP@EMNLP, Seite 217-227. Association for Computational Linguistics, (2020)