Profils utilisateurs correspondant à "Siddhant Arora"

Siddhant Arora

Graduate Student, Carnegie Mellon University
Adresse e-mail validée de andrew.cmu.edu
Cité 632 fois

Espnet-slu: Advancing spoken language understanding through espnet

S Arora, S Dalmia, P Denisov, X Chang… - ICASSP 2022-2022 …, 2022 - ieeexplore.ieee.org
As Automatic Speech Processing (ASR) systems are getting better, there is an increasing
interest of using the ASR output to do downstream Natural Language Processing (NLP) tasks. …

A survey on graph neural networks for knowledge graph completion

S Arora - arXiv preprint arXiv:2007.12374, 2020 - arxiv.org
Knowledge Graphs are increasingly becoming popular for a variety of downstream tasks like
Question Answering and Information Retrieval. However, the Knowledge Graphs are often …

High-dose versus low-dose prednisolone in symptomatic patients with post-COVID-19 diffuse parenchymal lung abnormalities: an open-label, randomised trial (the …

…, IS Sehgal, R Agarwal, S Arora… - European …, 2022 - Eur Respiratory Soc
In some patients, respiratory symptoms and imaging abnormalities persist after acute coronavirus
disease 2019 (COVID-19) pneumonia [1–3]. Chest computed tomography (CT) scans …

Reproducing whisper-style training using an open-source toolkit and publicly available data

…, X Chang, X Li, J Shi, S Arora… - 2023 IEEE Automatic …, 2023 - ieeexplore.ieee.org
Pre-training speech models on large volumes of data has achieved remarkable success.
OpenAI Whisper is a multilingual multitask model trained on 680k hours of supervised speech …

SLUE phase-2: A benchmark suite of diverse spoken language understanding tasks

S Shon, S Arora, CJ Lin, A Pasad, F Wu… - arXiv preprint arXiv …, 2022 - arxiv.org
Spoken language understanding (SLU) tasks have been studied for many decades in the
speech research community, but have not received as much attention as lower-level tasks like …

Explain, edit, and understand: Rethinking user study design for evaluating model explanations

S Arora, D Pruthi, N Sadeh, WW Cohen… - Proceedings of the …, 2022 - ojs.aaai.org
In attempts to" explain" predictions of machine learning models, researchers have proposed
hundreds of techniques for attributing predictions to features that are deemed important. …

BERT meets CTC: New formulation of end-to-end speech recognition with pre-trained masked language model

Y Higuchi, B Yan, S Arora, T Ogawa… - arXiv preprint arXiv …, 2022 - arxiv.org
This paper presents BERT-CTC, a novel formulation of end-to-end speech recognition that
adapts BERT for connectionist temporal classification (CTC). Our formulation relaxes the …

OWSM v3. 1: Better and faster open whisper-style speech models based on e-branchformer

Y Peng, J Tian, W Chen, S Arora, B Yan, Y Sudo… - arXiv preprint arXiv …, 2024 - arxiv.org
Recent studies have advocated for fully open foundation models to promote transparency
and open science. As an initial step, the Open Whisper-style Speech Model (OWSM) …

A study on the integration of pre-trained ssl, asr, lm and slu models for spoken language understanding

Y Peng, S Arora, Y Higuchi, Y Ueda… - 2022 IEEE Spoken …, 2023 - ieeexplore.ieee.org
Collecting sufficient labeled data for spoken language understanding (SLU) is expensive
and time-consuming. Recent studies achieved promising results by using pre-trained models …

Decoder-only architecture for streaming end-to-end speech recognition

E Tsunoo, H Futami, Y Kashiwagi, S Arora… - arXiv preprint arXiv …, 2024 - arxiv.org
Decoder-only language models (LMs) have been successfully adopted for speech-processing
tasks including automatic speech recognition (ASR). The LMs have ample …