User profiles for Alena Fenogenova
Alena FenogenovaSaluteDevices, HSE Verified email at hse.ru Cited by 525 |
mgpt: Few-shot learners go multilingual
Recent studies report that autoregressive language models can successfully solve many
NLP tasks via zero- and few-shot learning paradigms, which opens up new possibilities for …
NLP tasks via zero- and few-shot learning paradigms, which opens up new possibilities for …
A family of pretrained transformer language models for Russian
Transformer language models (LMs) are fundamental to NLP research methodologies and
applications in various languages. However, developing such models specifically for the …
applications in various languages. However, developing such models specifically for the …
RussianSuperGLUE: A Russian language understanding evaluation benchmark
T Shavrina, A Fenogenova, A Emelyanov… - arXiv preprint arXiv …, 2020 - arxiv.org
In this paper, we introduce an advanced Russian general language understanding evaluation
benchmark -- RussianGLUE. Recent advances in the field of universal language models …
benchmark -- RussianGLUE. Recent advances in the field of universal language models …
Russian paraphrasers: Paraphrase with transformers
A Fenogenova - Proceedings of the 8th workshop on balto-slavic …, 2021 - aclanthology.org
This paper studies the generation methods for paraphrasing in the Russian language. There
are several transformer-based models (Russian and multilingual) trained on a collected …
are several transformer-based models (Russian and multilingual) trained on a collected …
mGPT: Few-Shot Learners Go Multilingual
This paper introduces mGPT, a multilingual variant of GPT-3, pretrained on 61 languages
from 25 linguistically diverse language families using Wikipedia and the C4 Corpus. We detail …
from 25 linguistically diverse language families using Wikipedia and the C4 Corpus. We detail …
Findings of the the ruatd shared task 2022 on artificial text detection in russian
We present the shared task on artificial text detection in Russian, which is organized as a
part of the Dialogue Evaluation initiative, held in 2022. The shared task dataset includes texts …
part of the Dialogue Evaluation initiative, held in 2022. The shared task dataset includes texts …
The Russian-focused embedders' exploration: ruMTEB benchmark and Russian embedding model design
…, M Tikhonova, A Maksimova, A Fenogenova… - arXiv preprint arXiv …, 2024 - arxiv.org
Embedding models play a crucial role in Natural Language Processing (NLP) by creating
text embeddings used in various tasks such as information retrieval and assessing semantic …
text embeddings used in various tasks such as information retrieval and assessing semantic …
RuBLiMP: Russian benchmark of linguistic minimal pairs
Minimal pairs are a well-established approach to evaluating the grammatical knowledge of
language models. However, existing resources for minimal pairs address a limited number of …
language models. However, existing resources for minimal pairs address a limited number of …
TAPE: Assessing few-shot Russian language understanding
Recent advances in zero-shot and few-shot learning have shown promise for a scope of
research and practical purposes. However, this fast-growing area lacks standardized evaluation …
research and practical purposes. However, this fast-growing area lacks standardized evaluation …
Mera: A comprehensive llm evaluation in russian
A Fenogenova, A Chervyakov, N Martynov… - arXiv preprint arXiv …, 2024 - arxiv.org
Over the past few years, one of the most notable advancements in AI research has been in
foundation models (FMs), headlined by the rise of language models (LMs). As the models' …
foundation models (FMs), headlined by the rise of language models (LMs). As the models' …