Multilingual LAMA: Investigating Knowledge in Multilingual Pretrained Language Models

Data

Data from this paper can be downloaded here.

Code

Code is available here.

Downloads

mlama1.1.zip 53 languages [40MB]

mlama1.1-all.zip all data [122MB]

Cite

If you use the provided data in your work, please cite the following paper.

@inproceedings{kassner2021multilingual,
    title = "Multilingual {LAMA}: Investigating Knowledge in Multilingual Pretrained Language Models",
    author = {Kassner, Nora  and
      Dufter, Philipp  and
      Sch{\"u}tze, Hinrich},
    booktitle = "to appear in Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics",
    year = "2021",
    address = "Online",
    publisher = "Association for Computational Linguistics",
}

Contact: Nora Kassner, Philipp Dufter