Zobrazeno 1 - 10
of 131
pro vyhledávání: '"Jaiswal, Ajay A."'
Autor:
Naeem, Awais, Li, Tianhao, Liao, Huang-Ru, Xu, Jiawei, Mathew, Aby M., Zhu, Zehao, Tan, Zhen, Jaiswal, Ajay Kumar, Salibian, Raffi A., Hu, Ziniu, Chen, Tianlong, Ding, Ying
Accurate diagnosis and prognosis assisted by pathology images are essential for cancer treatment selection and planning. Despite the recent trend of adopting deep-learning approaches for analyzing complex pathology images, they fall short as they oft
Externí odkaz:
http://arxiv.org/abs/2411.17073
Autor:
Bandari, Abhinav, Yin, Lu, Hsieh, Cheng-Yu, Jaiswal, Ajay Kumar, Chen, Tianlong, Shen, Li, Krishna, Ranjay, Liu, Shiwei
Network pruning has emerged as a potential solution to make LLMs cheaper to deploy. However, existing LLM pruning approaches universally rely on the C4 dataset as the calibration data for calculating pruning scores, leaving its optimality unexplored.
Externí odkaz:
http://arxiv.org/abs/2410.07461
Autor:
Jaiswal, Ajay, Choudhary, Nurendra, Adkathimar, Ravinarayana, Alagappan, Muthu P., Hiranandani, Gaurush, Ding, Ying, Wang, Zhangyang, Huang, Edward W, Subbian, Karthik
Graph Neural Networks (GNNs) have attracted immense attention in the past decade due to their numerous real-world applications built around graph-structured data. On the other hand, Large Language Models (LLMs) with extensive pretrained knowledge and
Externí odkaz:
http://arxiv.org/abs/2407.14996
Autor:
Jaiswal, Ajay, Yin, Lu, Zhang, Zhenyu, Liu, Shiwei, Zhao, Jiawei, Tian, Yuandong, Wang, Zhangyang
Modern Large Language Models (LLMs) are composed of matrices with billions of elements, making their storage and processing quite demanding in terms of computational resources and memory usage. Being significantly large, such matrices can often be ex
Externí odkaz:
http://arxiv.org/abs/2407.11239
Autor:
Zhang, Zhenyu, Jaiswal, Ajay, Yin, Lu, Liu, Shiwei, Zhao, Jiawei, Tian, Yuandong, Wang, Zhangyang
Training Large Language Models (LLMs) is memory-intensive due to the large number of parameters and associated optimization states. GaLore, a recent method, reduces memory usage by projecting weight gradients into a low-rank subspace without compromi
Externí odkaz:
http://arxiv.org/abs/2407.08296
Autoregressive Large Language Models (e.g., LLaMa, GPTs) are omnipresent achieving remarkable success in language understanding and generation. However, such impressive capability typically comes with a substantial model size, which presents signific
Externí odkaz:
http://arxiv.org/abs/2404.03865
Autor:
Hong, Junyuan, Duan, Jinhao, Zhang, Chenhui, Li, Zhangheng, Xie, Chulin, Lieberman, Kelsey, Diffenderfer, James, Bartoldson, Brian, Jaiswal, Ajay, Xu, Kaidi, Kailkhura, Bhavya, Hendrycks, Dan, Song, Dawn, Wang, Zhangyang, Li, Bo
Compressing high-capability Large Language Models (LLMs) has emerged as a favored strategy for resource-efficient inferences. While state-of-the-art (SoTA) compression methods boast impressive advancements in preserving benign task performance, the p
Externí odkaz:
http://arxiv.org/abs/2403.15447
Graph Neural Networks (GNNs) have empowered the advance in graph-structured data analysis. Recently, the rise of Large Language Models (LLMs) like GPT-4 has heralded a new era in deep learning. However, their application to graph data poses distinct
Externí odkaz:
http://arxiv.org/abs/2402.08170
Autor:
Holste, Gregory, Zhou, Yiliang, Wang, Song, Jaiswal, Ajay, Lin, Mingquan, Zhuge, Sherry, Yang, Yuzhe, Kim, Dongkyun, Nguyen-Mau, Trong-Hieu, Tran, Minh-Triet, Jeong, Jaehyup, Park, Wongi, Ryu, Jongbin, Hong, Feng, Verma, Arsh, Yamagishi, Yosuke, Kim, Changhyun, Seo, Hyeryeong, Kang, Myungjoo, Celi, Leo Anthony, Lu, Zhiyong, Summers, Ronald M., Shih, George, Wang, Zhangyang, Peng, Yifan
Many real-world image recognition problems, such as diagnostic medical imaging exams, are "long-tailed" $\unicode{x2013}$ there are a few common findings followed by many more relatively rare conditions. In chest radiography, diagnosis is both a long
Externí odkaz:
http://arxiv.org/abs/2310.16112
Autor:
Yin, Lu, Wu, You, Zhang, Zhenyu, Hsieh, Cheng-Yu, Wang, Yaqing, Jia, Yiling, Li, Gen, Jaiswal, Ajay, Pechenizkiy, Mykola, Liang, Yi, Bendersky, Michael, Wang, Zhangyang, Liu, Shiwei
Large Language Models (LLMs), renowned for their remarkable performance across diverse domains, present a challenge when it comes to practical deployment due to their colossal model size. In response to this challenge, efforts have been directed towa
Externí odkaz:
http://arxiv.org/abs/2310.05175