Publications
publications by categories in reversed chronological order. generated by jekyll-scholar.
2024
- ACM SIGCHI 24Do People Mirror Emotion Differently with a Human or TTS Voice? Comparing Listener Ratings and Word EmbeddingsMichelle Cohn, Grisha Bandodkar, Raj Bharat Sangani, and 2 more authors2024
Speakers increasingly communicate with technology using spoken language and display behaviors from human-human interaction, such as mirroring the pronunciation patterns of text-to-speech (TTS) voices. Yet, the magnitude and directionality of mirroring varies across prior studies comparing human and TTS addressees. The current study uses two approaches to assess mirroring of emotionally expressive speech produced by a human and a TTS voice. We compare AXB perceptual similarity (holistic human listener judgment, n=109 raters), and distance in latent space (wav2vec 2.0) on a dataset of single word shadowing in a cohort of 36 United States English speakers. Results show that both AXB and wav2vec 2.0 capture mirroring toward emotional prosody for both human and TTS voices. We discuss these findings in terms of theories of computer personification and emotional mirroring, as well as their contributions toward methodological advancements in phonetic entrainment.
2022
- IEEE Big DataLarge-Scale Knowledge Synthesis and Complex Information Retrieval from Biomedical DocumentsShreya Saxena, Raj Sangani, Siva Prasad, and 4 more authorsDec 2022
Recent advances in the healthcare industry have led to an abundance of unstructured data, making it challenging to perform tasks such as efficient and accurate information retrieval at scale. Our work offers an all-in-one scalable solution for extracting and exploring complex information from large-scale research documents, which would otherwise be tedious. First, we briefly explain our knowledge synthesis process to extract helpful information from unstructured text data of research documents. Then, on top of the knowledge extracted from the documents, we perform complex information retrieval using three major components- Paragraph Retrieval, Triplet Retrieval from Knowledge Graphs, and Complex Question Answering (QA). These components combine lexical and semantic-based methods to retrieve paragraphs and triplets and perform faceted refinement for filtering these search results. The complexity of biomedical queries and documents necessitates using a QA system capable of handling queries more complex than factoid queries, which we evaluate qualitatively on the COVID-19 Open Research Dataset (CORD-19) to demonstrate the effectiveness and value- add.
2021
- IEEE XploreComparing Deep Sentiment Models using Quantified Local ExplanationsRaj Bharat Sangani, Apratim Shukla, and Radhika B. SelvamaniNov 2021
Online judge systems that enable the automatic evaluation of predictive models face the problem of choosing the best model among those with near equal accuracy measures. In such scenarios, interpreting competitive models may provide a better insight towards robustness. Sentiment analysis is a well- explored domain with many available approaches and libraries. Despite this, it is context-sensitive and poses many challenges to the research community and thus deems fit to our analysis. Selecting a model with the highest accuracy may not be satisfactory, especially when the decision margins are too narrow. Our comparative study on models with similar accuracies and f1-scores but distinct underlying architectures incorporate custom metrics and evaluation methods to assess their performance on a sentiment analysis task. We have proposed to include human judgment in an online judge system using a feedback acquisition mechanism, presenting explanations for model decisions on selected test cases. Our initial experimental findings indicate that evaluating model robustness by incorporating a well-defined human feedback mechanism, using model agnostic approaches encourages online judge systems to make explainable decisions. The p-value associated with a paired t-Test on the feedback collected for model preference indicates a significant preference for the model supported by our metrics.