B Bhattarai MultiModal Learning Lab (MMLL)

B Bhattarai MultiModal Learning Lab (MMLL) is a research group within NAAMII that focuses on theoretical and applied research in Machine learning (ML) where the researches process information from heterogeneous sources such as vision, text, and speech to make computers understand, interpret and reason. Our applications include but are not limited to computer vision, medical image analysis and low-resource language processing.

Latest Related Publications

View All Publications
Sophia Bano, Alessandro Casella, Francisco Vasconcelos, Abdul Qayyum, Abdesslam Benzinou, Moona Mazher, Fabrice Meriaudeau, Chiara Lena, Ilaria Anita Cintorrino, Gaia Romana De Paolis, Jessica Biagioli, Daria Grechishnikova, Jing Jiao, Bizhe Bai, Yanyan Qiao, Binod Bhattarai, Rebati Raman Gaire, Ronast Subedi, Eduard Vazquez, Szymon Płotka, Aneta Lisowska, Arkadiusz Sitek, George Attilakos, Ruwan Wimalasundera, Anna L David, Dario Paladini, Jan Deprest, Elena De Momi, Leonardo S Mattos, Sara Moccia, Danail Stoyanov
Placental Vessel Segmentation and Registration in Fetoscopy: Literature Review and MICCAI FetReg2021 Challenge Findings
Bano, Sophia, et al. "Fetreg2021: a challenge on placental vessel segmentation and registration in fetoscopy.", (2022)., 2023
Bibtex

@article{bano2023placental,
  title={Placental Vessel Segmentation and Registration in Fetoscopy: Literature Review and MICCAI FetReg2021 Challenge Findings},
  author={Bano, Sophia and Casella, Alessandro and Vasconcelos, Francisco and Qayyum, Abdul and Benzinou, Abdesslam and Mazher, Moona and Meriaudeau, Fabrice and Lena, Chiara and Cintorrino, Ilaria Anita and De Paolis, Gaia Romana and others},
  journal={Medical Image Analysis},
  year={2023}
}

Binod Bhattarai, Ronast Subedi, Rebati Raman Gaire, Eduard Vazquez, Danail Stoyanov
Histogram of Oriented Gradients Meet Deep Learning: A Novel Multi-task Deep Network for Medical Image Semantic Segmentation
Medical Image Analysis, 2023
Bibtex

@article{bhattarai2023histogram,
  title={Histogram of Oriented Gradients meet deep learning: A novel multi-task deep network for 2D surgical image semantic segmentation},
  author={Bhattarai, Binod and Subedi, Ronast and Gaire, Rebati Raman and Vazquez, Eduard and Stoyanov, Danail},
  journal={Medical Image Analysis},
  pages={102747},
  year={2023},
  publisher={Elsevier}
}

Sulav Timilsina, Milan Gautam, Binod Bhattarai
NepBERTa: Nepali Language Model Trained in a Large Corpus
Proceedings of the 2nd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 12th International Joint Conference on Natural Language Processing, (2022)., 2022
Bibtex

@inproceedings{timilsina-etal-2022-nepberta,
    title = "{N}ep{BERT}a: {N}epali Language Model Trained in a Large Corpus",
    author = "Timilsina, Sulav  and
      Gautam, Milan  and
      Bhattarai, Binod",
    booktitle = "Proceedings of the 2nd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 12th International Joint Conference on Natural Language Processing (Volume 2: Short Papers)",
    month = nov,
    year = "2022",
    address = "Online only",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2022.aacl-short.34",
    pages = "273--284",
    abstract = "Nepali is a low-resource language with more than 40 million speakers worldwide. It is written in Devnagari script and has rich semantics and complex grammatical structure. To this date, multilingual models such as Multilingual BERT, XLM and XLM-RoBERTa haven{'}t been able to achieve promising results in Nepali NLP tasks, and there does not exist any such a large-scale monolingual corpus. This study presents NepBERTa, a BERT-based Natural Language Understanding (NLU) model trained on the most extensive monolingual Nepali corpus ever. We collected a dataset of 0.8B words from 36 different popular news sites in Nepal and introduced the model. This data set is 3 folds times larger than the previous publicly available corpus. We evaluated the performance of NepBERTa in multiple Nepali-specific NLP tasks, including Named-Entity Recognition, Content Classification, POS Tagging, and Sequence Pair Similarity. We also introduce two different datasets for two new downstream tasks and benchmark four diverse NLU tasks altogether. We bring all these four tasks under the first-ever Nepali Language Understanding Evaluation (Nep-gLUE) benchmark. We will make Nep-gLUE along with the pre-trained model and data sets publicly available for research.",
}

Latest Related News

View All News

Nothing Found

There are no content in this category for this theme. Please check again later , or try a different theme


Latest Related Blogs

View All Blogs

Nothing Found

There are no content in this category for this theme. Please check again later , or try a different theme