I'm Vishwas Kothari, a CS graduate student and AI researcher focused on building machine learning systems that are not only accurate, but understandable and trustworthy. My work spans explainable AI, data engineering, vision-language models, and ML research, with experience at ISRO's Space Applications Centre and projects across financial AI, optimizer behavior, and robustness. I care about turning complex ideas into systems people can inspect, question, and rely on.
|
Research Intern Space Applications Centre - ISRO |
Built Python pipelines for satellite sensor data processing, NetCDF merging, optimization, and cross-sensor validation. |
|
Professional Master's in CS University of Colorado Boulder |
Graduate work focused on machine learning, interpretability, and real-world deployment. |
|
Publication & Reviewing Springer · Elsevier |
Springer publication on ethical AI and peer-review work with Elsevier journals. |
|
Empowering Survivors Ethical Artificial Intelligence for Countering Violence Against Women DOI: 10.1007/978-981-96-6046-9_26 |
Springer book chapter in Data Mining and Information Security, focused on ethical AI systems for survivor support, privacy-aware intervention, and risk identification.
AbstractThe maltreatment of women is a problem that affects all countries. Prevention, intervention, and support are among the many uses that can be provided through artificial intelligence. This paper discusses how artificial intelligence can be used ethically in the fight against VAWs. One way is through chatbots which could guide victims toward resources and legal remedies without having to reveal their identity. Furthermore, by sifting through information collected from different sources, such systems may also help identify risk factors or predict where violence might occur next. In any case, we must always remember about the privacy concerns when it comes to handling sensitive data thus while designing such approaches, fairness issues should also be taken into consideration so that some degree of human control remains. AI should not only support our global efforts to end violence against women but also empower survivors. |
|
XAI Credit Lens Explainable credit risk |
SHAP, LIME, DiCE, fairness auditing, and regulatory mapping for transparent credit-risk decisions. |
|
Financial Document VQA Vision-language models |
Benchmarked models on SEC 10-K filings and improved domain performance with LoRA. |
|
Implicit Bias of Adam vs. SGD Optimizer behavior |
Studied max-margin solutions, optimizer geometry, and patched-MNIST spurious correlations. |