Anima Anandkumar
Applied Filters
- Anima Anandkumar
- AuthorRemove filter
People
Colleagues
- Anima Anandkumar (116)
- Kamyar Azizzadenesheli (12)
- Daniel Joseph Hsu (10)
- Sham M Kakade (10)
- Zhiding Yu (10)
- Alan S Willsky (9)
- Jean Kossaifi (8)
- Lang Tong (8)
- Vincent Yan F Tan (8)
- Arvind Ramanathan (6)
- Majid Janzamin (6)
- Nikola Borislavov Kovachki (5)
- C. Bisdikian (4)
- Dakshi Agrawal (4)
- Ian Foster (4)
- Venkatram Vishwanath (4)
- Andrew M Stuart (3)
- Jan Kautz (3)
- Rick L Stevens (3)
- Zhangyang Wang (3)
Roles
Publication
Journal/Magazine Names
- The Journal of Machine Learning Research (13)
- IEEE Transactions on Information Theory (4)
- IEEE Transactions on Signal Processing (4)
- ACM SIGMETRICS Performance Evaluation Review (3)
- International Journal of High Performance Computing Applications (3)
- IEEE Journal on Selected Areas in Communications (2)
- ACM / IMS Journal of Data Science (1)
- Algorithmica (1)
- Foundations and Trends® in Machine Learning (1)
- International Journal of Robotics Research (1)
- Performance Evaluation (1)
- SIAM Journal on Optimization (1)
Proceedings/Book Names
- NIPS '22: Proceedings of the 36th International Conference on Neural Information Processing Systems (8)
- NIPS '20: Proceedings of the 34th International Conference on Neural Information Processing Systems (7)
- NIPS '21: Proceedings of the 35th International Conference on Neural Information Processing Systems (7)
- ICML'24: Proceedings of the 41st International Conference on Machine Learning (6)
- ICML'20: Proceedings of the 37th International Conference on Machine Learning (4)
- ICML'23: Proceedings of the 40th International Conference on Machine Learning (4)
- NIPS '23: Proceedings of the 37th International Conference on Neural Information Processing Systems (4)
- NIPS'12: Proceedings of the 26th International Conference on Neural Information Processing Systems - Volume 1 (3)
- 2021 IEEE International Conference on Robotics and Automation (ICRA) (2)
- NIPS'11: Proceedings of the 25th International Conference on Neural Information Processing Systems (2)
- Computer Vision – ECCV 2018 (1)
- Computer Vision – ECCV 2022 (1)
- Computer Vision – ECCV 2022 (1)
- DAC '22: Proceedings of the 59th ACM/IEEE Design Automation Conference (1)
- ISPD '23: Proceedings of the 2023 International Symposium on Physical Design (1)
- KDD '21: Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining (1)
- MM '21: Proceedings of the 29th ACM International Conference on Multimedia (1)
- PASC '23: Proceedings of the Platform for Advanced Scientific Computing Conference (1)
- SIGMETRICS '08: Proceedings of the 2008 ACM SIGMETRICS international conference on Measurement and modeling of computer systems (1)
- SIGMETRICS '11: Proceedings of the ACM SIGMETRICS joint international conference on Measurement and modeling of computer systems (1)
Publisher
- Curran Associates Inc. (34)
- JMLR.org (29)
- IEEE Press (22)
- Association for Computing Machinery (12)
- Springer-Verlag (5)
- Sage Publications, Inc. (4)
- MIT Press (3)
- Cornell University (1)
- Elsevier Science Publishers B. V. (1)
- IEEE Computer Society (1)
- Now Publishers Inc. (1)
- Omnipress (1)
- Society for Industrial and Applied Mathematics (1)
- Unknown publishers (1)
Publication Date
Export Citations
Publications
Save this search
Please login to be able to save your searches and receive alerts for new content matching your search criteria.
- research-article
MProt-DPO: Breaking the ExaFLOPS Barrier for Multimodal Protein Design Workflows with Direct Preference Optimization
Gautham Dharuman
Argonne National Laboratory and Joint first authors
,Kyle Hippe
Argonne National Laboratory, University of Chicago and Joint first authors
,Alexander Brace
Argonne National Laboratory, University of Chicago and Joint first authors
,Sam Foreman
Argonne National Laboratory and Joint first authors
,Väinö Hatanpää
Argonne National Laboratory
,Varuni K. Sastry
Argonne National Laboratory
,Huihuo Zheng
Argonne National Laboratory
,Logan Ward
Argonne National Laboratory
,Servesh Muralidharan
Argonne National Laboratory
,Archit Vasan
Argonne National Laboratory
,Bharat Kale
Argonne National Laboratory
,Carla M. Mann
Argonne National Laboratory and University of Chicago
,Heng Ma
Argonne National Laboratory
,Yun-Hsuan Cheng
NVIDIA Inc.
,Yuliana Zamora
NVIDIA Inc.
,Shengchao Liu
University of California, Berkeley
,Chaowei Xiao
University of Wisconsin-Madison, Madison
,Murali Emani
Argonne National Laboratory
,Tom Gibbs
NVIDIA Inc.
,Mahidhar Tatineni
San Diego Supercomputing Center
,Deepak Canchi
Intel Corporation
,Jerome Mitchell
Intel Corporation
,Koichi Yamada
Intel Corporation
,Maria Garzaran
Intel Corporation
,Michael E. Papka
Argonne National Laboratory and University of Illinois Chicago
,Ian Foster
Argonne National Laboratory and University of Chicago
,Rick Stevens
Argonne National Laboratory and University of Chicago
,Anima Anandkumar
California Institute of Technology
,Venkatram Vishwanath
Argonne National Laboratory and University of Illinois Chicago
,Arvind Ramanathan
Argonne National Laboratory and University of Chicago
SC '24: Proceedings of the International Conference for High Performance Computing, Networking, Storage, and Analysis•November 2024, Article No.: 7, pp 1-13• https://doi.org/10.1109/SC41406.2024.00013We present a scalable, end-to-end workflow for protein design. By augmenting protein sequences with natural language descriptions of their biochemical properties, we train generative models that can be preferentially aligned with protein fitness ...
- 0Citation
- 213
- Downloads
MetricsTotal Citations0Total Downloads213Last 12 Months213Last 6 weeks32- 1
- research-article
GaLore: memory-efficient LLM training by gradient low-rank projection
Jiawei Zhao
California Institute of Technology
,Zhenyu Zhang
University of Texas at Austin
,Beidi Chen
Meta AI and Carnegie Mellon University
,Zhangyang Wang
University of Texas at Austin
,Anima Anandkumar
California Institute of Technology
,Yuandong Tian
Meta AI
ICML'24: Proceedings of the 41st International Conference on Machine Learning•July 2024, Article No.: 2528, pp 61121-61143Training Large Language Models (LLMs) presents significant memory challenges, predominantly due to the growing size of weights and optimizer states. Common memory-reduction approaches, such as low-rank adaptation (LoRA), add a trainable low-rank matrix ...
- 0Citation
MetricsTotal Citations0
- research-article
Equivariant graph neural operator for modeling 3D dynamics
Minkai Xu
Stanford University
,Jiaqi Han
Stanford University
,Aaron Lou
Stanford University
,Jean Kossaifi
NVIDIA
,Arvind Ramanathan
Argonne National Laboratory
,Kamyar Azizzadenesheli
NVIDIA
,Jure Leskovec
Stanford University
,Stefano Ermon
Stanford University
,Anima Anandkumar
California Institute of Technology
ICML'24: Proceedings of the 41st International Conference on Machine Learning•July 2024, Article No.: 2265, pp 55015-55032Modeling the complex three-dimensional (3D) dynamics of relational systems is an important problem in the natural sciences, with applications ranging from molecular simulations to particle mechanics. Machine learning methods have achieved good success by ...
- 0Citation
MetricsTotal Citations0
- research-article
Solving poisson equations using neural walk-on-spheres
Hong Chul Nam
ETH Zurich
,Julius Berner
Caltech
,Anima Anandkumar
Caltech
ICML'24: Proceedings of the 41st International Conference on Machine Learning•July 2024, Article No.: 1513, pp 37277-37292We propose Neural Walk-on-Spheres (NWoS), a novel neural PDE solver for the efficient solution of high-dimensional Poisson equations. Leveraging stochastic representations and Walk-on-Spheres methods, we develop novel losses for neural networks based on ...
- 0Citation
MetricsTotal Citations0
- research-article
Autoformalizing euclidean geometry
Logan Murphy
University of Toronto
,Kaiyu Yang
Caltech
,Jialiang Sun
University of Toronto
,Zhaoyu Li
University of Toronto
,Anima Anandkumar
Caltech
,Xujie Si
University of Toronto
ICML'24: Proceedings of the 41st International Conference on Machine Learning•July 2024, Article No.: 1497, pp 36847-36893Autoformalization involves automatically translating informal math into formal theorems and proofs that are machine-verifiable. Euclidean geometry provides an interesting and controllable domain for studying autoformalization. In this paper, we introduce ...
- 0Citation
MetricsTotal Citations0
- research-article
Neural operators with localized integral and differential kernels
Miguel Liu-Schiaffini
Department of Computing and Mathematical Sciences, California Institute of Technology, Pasadena CA
,Julius Berner
Department of Computing and Mathematical Sciences, California Institute of Technology, Pasadena CA
,Boris Bonev
NVIDIA, Santa Clara, CA
,Thorsten Kurth
NVIDIA, Santa Clara, CA
,Kamyar Azizzadenesheli
NVIDIA, Santa Clara, CA
,Anima Anandkumar
Department of Computing and Mathematical Sciences, California Institute of Technology, Pasadena CA
ICML'24: Proceedings of the 41st International Conference on Machine Learning•July 2024, Article No.: 1321, pp 32576-32594Neural operators learn mappings between function spaces, which is practical for learning solution operators of PDEs and other scientific modeling applications. Among them, the Fourier neural operator (FNO) is a popular architecture that performs global ...
- 0Citation
MetricsTotal Citations0
- research-article
DPOT: auto-regressive denoising operator transformer for large-scale PDE pre-training
Zhongkai Hao
Dept. of Comp. Sci. & Techn., Institute for AI, BNRist Center, Tsinghua-Bosch Joint ML Center and Dept. of EE, Tsinghua University
,Chang Su
Dept. of Comp. Sci. & Techn., Institute for AI, BNRist Center, Tsinghua-Bosch Joint ML Center, Tsinghua University
,Songming Liu
Dept. of Comp. Sci. & Techn., Institute for AI, BNRist Center, Tsinghua-Bosch Joint ML Center, Tsinghua University
,Julius Berner
Caltech
,Chengyang Ying
Dept. of Comp. Sci. & Techn., Institute for AI, BNRist Center, Tsinghua-Bosch Joint ML Center, Tsinghua University
,Hang Su
Dept. of Comp. Sci. & Techn., Institute for AI, BNRist Center, Tsinghua-Bosch Joint ML Center, Tsinghua University
,Anima Anandkumar
Caltech
,Jian Song
Dept. of EE, Tsinghua University
,Jun Zhu
Dept. of Comp. Sci. & Techn., Institute for AI, BNRist Center, Tsinghua-Bosch Joint ML Center, Tsinghua University and RealAI
ICML'24: Proceedings of the 41st International Conference on Machine Learning•July 2024, Article No.: 703, pp 17616-17635Pre-training has been investigated to improve the efficiency and performance of training neural operators in data-scarce settings. However, it is largely in its infancy due to the inherent complexity and diversity, such as long trajectories, multiple ...
- 0Citation
MetricsTotal Citations0
- research-articleOpen Access
Published By ACM
Published By ACM
Physics-Informed Neural Operator for Learning Partial Differential Equations
Zongyi Li
Computing and mathematical science, California Institute of Technology, Pasadena, USA
,Hongkai Zheng
Computing and mathematical science, California Institute of Technology, Pasadena, USA
,Nikola Kovachki
Computing and mathematical science, California Institute of Technology, Pasadena, USA
,David Jin
Computing and mathematical science, California Institute of Technology, Pasadena, USA
,Haoxuan Chen
Computing and mathematical science, California Institute of Technology, Pasadena, USA
,Burigede Liu
Computing and mathematical science, California Institute of Technology, Pasadena, USA
,Kamyar Azizzadenesheli
Computing and mathematical science, California Institute of Technology, Pasadena, USA
,Anima Anandkumar
Computing and mathematical science, California Institute of Technology, Pasadena, USA
ACM / IMS Journal of Data Science, Volume 1, Issue 3•September 2024, Article No.: 9, pp 1-27 • https://doi.org/10.1145/3648506In this article, we propose physics-informed neural operators (PINO) that combine training data and physics constraints to learn the solution operator of a given family of parametric Partial Differential Equations (PDE). PINO is the first hybrid approach ...
HighlightsPROBLEM STATEMENT
Machine learning methods have recently shown promise in solving partial differential equations (PDEs) raised in science and engineering. They can be classified into two broad categories: approximating the solution function ...
- 35Citation
- 15,341
- Downloads
MetricsTotal Citations35Total Downloads15,341Last 12 Months15,341Last 6 weeks2,043
- research-article
Action-conditional implicit visual dynamics for deformable object manipulation
Bokui Shen
Computer Science Department, 6429Stanford University, Stanford, CA, USA
,Zhenyu Jiang
Computer Science Department, 12330The University of Texas at Austin, Austin, TX, USA
,Christopher Choy
AI Algorithm Research, 196328Nvidia Inc., Santa Clara, CA, USA
,Silvio Savarese
Computer Science Department, 6429Stanford University, Stanford, CA, USA
,Leonidas J. Guibas
Computer Science Department, 12330The University of Texas at Austin, Austin, TX, USA
,Anima Anandkumar
AI Algorithm Research, 196328Nvidia Inc., Santa Clara, CA, USA
Computer Science Department , California Institute of Technology, Pasadena, CA, USA
,Yuke Zhu
Computer Science Department, 12330The University of Texas at Austin, Austin, TX, USA
AI Algorithm Research, 196328Nvidia Inc., Santa Clara, CA, USA
,Shoudong Huang,
Kris Hauser,
Dylan A. Shell
International Journal of Robotics Research, Volume 43, Issue 4•Apr 2024, pp 437-455 • https://doi.org/10.1177/02783649231191222Manipulating volumetric deformable objects in the real world, like plush toys and pizza dough, brings substantial challenges due to infinite shape variations, non-rigid motions, and partial observability. We introduce ACID, an action-conditional visual ...
- 0Citation
MetricsTotal Citations0
- research-article
Symmetry-informed geometric representation for molecules, proteins, and crystalline materials
Shengchao Liu
Mila - Québec Artificial Intelligence Institute and Université de Montréal
,Weitao Du
University of Chinese Academy of Sciences
,Yanjing Li
Carnegie Mellon University
,Zhuoxinran Li
University of Toronto
,Zhiling Zheng
University of California, Berkeley
,Chenru Duan
Massachusetts Institute of Technology
,Zhiming Ma
University of Chinese Academy of Sciences
,Omar Yaghi
University of California, Berkeley
,Anima Anandkumar
California Institute of Technology
,Christian Borgs
University of California, Berkeley
,Jennifer Chayes
University of California, Berkeley
,Hongyu Guo
National Research Council Canada
,Jian Tang
Mila - Québec Artificial Intelligence Institute and HEC Montréal and CIFAR AI Chair
NIPS '23: Proceedings of the 37th International Conference on Neural Information Processing Systems•December 2023, Article No.: 2886, pp 66084-66101Artificial intelligence for scientific discovery has recently generated significant interest within the machine learning and scientific communities, particularly in the domains of chemistry, biology, and material discovery. For these scientific problems, ...
- 0Citation
MetricsTotal Citations0
- research-article
Geometry-informed neural operator for large-scale 3D PDEs
Zongyi Li
NVIDIA
,Nikola Borislavov Kovachki
NVIDIA
,Chris Choy
NVIDIA
,Boyi Li
NVIDIA
,Jean Kossaifi
NVIDIA
,Shourya Prakash Otta
NVIDIA
,Mohammad Amin Nabian
NVIDIA
,Maximilian Stadler
NVIDIA
,Christian Hundt
NVIDIA
,Kamyar Azizzadenesheli
NVIDIA
,Anima Anandkumar
NVIDIA
NIPS '23: Proceedings of the 37th International Conference on Neural Information Processing Systems•December 2023, Article No.: 1556, pp 35836-35854We propose the geometry-informed neural operator (GINO), a highly efficient approach for learning the solution operator of large-scale partial differential equations with varying geometries. GINO uses a signed distance function (SDF) and point-cloud ...
- 0Citation
MetricsTotal Citations0
- research-article
ClimSim: a large multi-scale dataset for hybrid physics-ML climate emulation
Sungduk Yu
UCI
,Walter M. Hannah
LLNL
,Liran Peng
UCI
,Jerry Lin
UCI
,Mohamed Aziz Bhouri
Columbia
,Ritwik Gupta
UCB
,Björn Lütjens
MIT
,Justus C. Will
UCI
,Gunnar Behrens
DLR
,Julius J. M. Busecke
Columbia
,Nora Loose
Princeton
,Charles Stern
Columbia
,Tom Beucler
UNIL
,Bryce E. Harrop
PNNL
,Benjamin R. Hillman
SNL
,Andrea M. Jenney
UCI and OSU
,Savannah L. Ferretti
UCI
,Nana Liu
UCI
,Anima Anandkumar
NVIDIA
,Noah D. Brenowitz
NVIDIA
,Veronika Eyring
DLR
,Nicholas Geneva
NVIDIA
,Pierre Gentine
Columbia
,Stephan Mandt
UCI
,Jaideep Pathak
NVIDIA
,Akshay Subramaniam
NVIDIA
,Carl Vondrick
Columbia
,Rose Yu
UCSD
,Laure Zanna
NYU
,Tian Zheng
Columbia
,Ryan P. Abernathey
Columbia
,Fiaz Ahmed
UCLA
,David C. Bader
LLNL
,Pierre Baldi
UCI
,Elizabeth A. Barnes
CSU
,Christopher S. Bretherton
Allen AI
,Peter M. Caldwell
LLNL
,Wayne Chuang
Columbia
,Yilun Han
Tsinghua
,Yu Huang
Columbia
,Fernando Iglesias-Suarez
DLR
,Sanket Jantre
BNL
,Karthik Kashinath
NVIDIA
,Marat Khairoutdinov
SUNY
,Thorsten Kurth
NVIDIA
,Nicholas J. Lutsko
UCSD
,Po-Lun Ma
PNNL
,Griffin Mooers
UCI
,J. David Neelin
UCLA
,David A. Randall
CSU
,Sara Shamekh
Columbia
,Mark A. Taylor
SNL
,Nathan M. Urban
BNL
,Janni Yuval
MIT
,Guang J. Zhang
UCSD
,Michael S. Pritchard
UCI an NVIDIA
NIPS '23: Proceedings of the 37th International Conference on Neural Information Processing Systems•December 2023, Article No.: 968, pp 22070-22084Modern climate projections lack adequate spatial and temporal resolution due to computational constraints. A consequence is inaccurate and imprecise predictions of critical processes such as storms. Hybrid methods that combine physics with machine ...
- 0Citation
MetricsTotal Citations0
- research-article
LeanDojo: theorem proving with retrieval-augmented language models
Kaiyu Yang
Caltech
,Aidan M. Swope
NVIDIA
,Alex Gu
MIT
,Rahul Chalamala
Caltech
,Peiyang Song
UC Santa Barbara
,Shixing Yu
UT Austin
,Saad Godil
NVIDIA
,Ryan Prenger
NVIDIA
,Anima Anandkumar
Caltech and NVIDIA
NIPS '23: Proceedings of the 37th International Conference on Neural Information Processing Systems•December 2023, Article No.: 944, pp 21573-21612Large language models (LLMs) have shown promise in proving formal theorems using proof assistants such as Lean. However, existing methods are difficult to reproduce or build on, due to private code, data, and large compute requirements. This has created ...
- 0Citation
MetricsTotal Citations0
- research-article
Published By ACM
Published By ACM
Protein Generation via Genome-scale Language Models with Bio-physical Scoring
Gautham Dharuman
Argonne National Laboratory, USA
,Logan Ward
Argonne National Laboratory, USA
,Heng Ma
Argonne National Laboratory, USA
,Priyanka V. Setty
Argonne National Laboratory, USA
,Ozan Gokdemir
University of Chicago, USA
,Sam Foreman
Argonne National Laboratory, USA
,Murali Emani
Argonne National Laboratory, USA
,Kyle Hippe
Argonne National Laboratory, USA
,Alexander Brace
Argonne National Laboratory, USA
,Kristopher Keipert
Nvidia Inc.
,Thomas Gibbs
Nvidia Inc.
,Ian Foster
Argonne National Laboratory, USA
,Anima Anandkumar
California Institute of Technology, USA
,Venkatram Vishwanath
Argonne National Laboratory, USA
,Arvind Ramanathan
Argonne National Laboratory, USA
SC-W '23: Proceedings of the SC '23 Workshops of the International Conference on High Performance Computing, Network, Storage, and Analysis•November 2023, pp 95-101• https://doi.org/10.1145/3624062.3626087Large language models (LLMs) trained on vast biological datasets can learn biological motifs and correlations across the evolutionary landscape of natural proteins. LLMs can then be used for de novo design of novel proteins with specific structures, ...
- 2Citation
- 271
- Downloads
MetricsTotal Citations2Total Downloads271Last 12 Months203Last 6 weeks7- 1
- research-article
GenSLMs: Genome-scale language models reveal SARS-CoV-2 evolutionary dynamics
Maxim Zvyagin
Data Science and Learning Division,1291 Argonne National Laboratory, Lemont, IL, USA
,Alexander Brace
Data Science and Learning Division,1291 Argonne National Laboratory, Lemont, IL, USA
Department of Computer Science,2462 University of Chicago, Hyde Park, IL, USA
,Kyle Hippe
Data Science and Learning Division,1291 Argonne National Laboratory, Lemont, IL, USA
,Yuntian Deng
NVIDIA Inc., Santa Clara, CA, USA
Harvard University, Cambridge, MA, USA
,Bin Zhang
Cerebras Inc., San Jose, CA, USA
,Cindy Orozco Bohorquez
Cerebras Inc., San Jose, CA, USA
,Austin Clyde
Data Science and Learning Division,1291 Argonne National Laboratory, Lemont, IL, USA
Department of Computer Science,2462 University of Chicago, Hyde Park, IL, USA
,Bharat Kale
Computer Science Department, 2848Northern Illinois University, DeKalb, IL, USA
,Danilo Perez-Rivera
Data Science and Learning Division,1291 Argonne National Laboratory, Lemont, IL, USA
5894New York University, New York, NY, USA
,Heng Ma
Data Science and Learning Division,1291 Argonne National Laboratory, Lemont, IL, USA
,Carla M. Mann
Data Science and Learning Division,1291 Argonne National Laboratory, Lemont, IL, USA
Department of Computer Science,2462 University of Chicago, Hyde Park, IL, USA
,Michael Irvin
Data Science and Learning Division,1291 Argonne National Laboratory, Lemont, IL, USA
,Defne G. Ozgulbas
Department of Biochemistry, University of Illinois-Urbana Champaign, Champaign, IL, USA
,Natalia Vassilieva
Cerebras Inc., San Jose, CA, USA
,James Gregory Pauloski
Department of Computer Science,2462 University of Chicago, Hyde Park, IL, USA
,Logan Ward
Data Science and Learning Division,1291 Argonne National Laboratory, Lemont, IL, USA
,Valerie Hayot-Sasson
Data Science and Learning Division,1291 Argonne National Laboratory, Lemont, IL, USA
Department of Computer Science,2462 University of Chicago, Hyde Park, IL, USA
Argonne Leadership Computing Facility, Argonne National Laboratory, Lemont, IL, USA
,Murali Emani
Data Science and Learning Division,1291 Argonne National Laboratory, Lemont, IL, USA
Argonne Leadership Computing Facility, Argonne National Laboratory, Lemont, IL, USA
,Sam Foreman
Data Science and Learning Division,1291 Argonne National Laboratory, Lemont, IL, USA
Argonne Leadership Computing Facility, Argonne National Laboratory, Lemont, IL, USA
,Zhen Xie
Data Science and Learning Division,1291 Argonne National Laboratory, Lemont, IL, USA
,Diangen Lin
Data Science and Learning Division,1291 Argonne National Laboratory, Lemont, IL, USA
Department of Computer Science,2462 University of Chicago, Hyde Park, IL, USA
,Maulik Shukla
Data Science and Learning Division,1291 Argonne National Laboratory, Lemont, IL, USA
Department of Computer Science,2462 University of Chicago, Hyde Park, IL, USA
,Weili Nie
NVIDIA Inc., Santa Clara, CA, USA
,Josh Romero
NVIDIA Inc., Santa Clara, CA, USA
,Christian Dallago
NVIDIA Inc., Santa Clara, CA, USA
Computer Science Department, Technical University of Munich, Munich,Germany
,Arash Vahdat
NVIDIA Inc., Santa Clara, CA, USA
,Chaowei Xiao
NVIDIA Inc., Santa Clara, CA, USA
Department of Biochemistry, University of Illinois-Urbana Champaign, Champaign, IL, USA
,Thomas Gibbs
NVIDIA Inc., Santa Clara, CA, USA
,Ian Foster
Data Science and Learning Division,1291 Argonne National Laboratory, Lemont, IL, USA
Department of Computer Science,2462 University of Chicago, Hyde Park, IL, USA
,James J. Davis
Data Science and Learning Division,1291 Argonne National Laboratory, Lemont, IL, USA
Department of Computer Science,2462 University of Chicago, Hyde Park, IL, USA
,Michael E. Papka
Data Science and Learning Division,1291 Argonne National Laboratory, Lemont, IL, USA
Argonne Leadership Computing Facility, Argonne National Laboratory, Lemont, IL, USA
Computer Science Department, University of Illinois Chicago, Chicago, IL, USA
,Thomas Brettin
Data Science and Learning Division,1291 Argonne National Laboratory, Lemont, IL, USA
Computing, Environment and Life Sciences Directorate, Argonne National Laboratory, Lemont, IL, USA
,Rick Stevens
Data Science and Learning Division,1291 Argonne National Laboratory, Lemont, IL, USA
Department of Computer Science,2462 University of Chicago, Hyde Park, IL, USA
Computing, Environment and Life Sciences Directorate, Argonne National Laboratory, Lemont, IL, USA
,Anima Anandkumar
NVIDIA Inc., Santa Clara, CA, USA
Computer Science Department, 6469California Institute of Technology, Pasadena, CA, USA
,Venkatram Vishwanath
Data Science and Learning Division,1291 Argonne National Laboratory, Lemont, IL, USA
Argonne Leadership Computing Facility, Argonne National Laboratory, Lemont, IL, USA
,Arvind Ramanathan
Data Science and Learning Division,1291 Argonne National Laboratory, Lemont, IL, USA
,Edmond Chow
International Journal of High Performance Computing Applications, Volume 37, Issue 6•Nov 2023, pp 683-705 • https://doi.org/10.1177/10943420231201154We seek to transform how new and emergent variants of pandemic-causing viruses, specifically SARS-CoV-2, are identified and classified. By adapting large language models (LLMs) for genomic data, we build genome-scale language models (GenSLMs) which can ...
- 5Citation
MetricsTotal Citations5
- research-article
Learning calibrated uncertainties for domain shift: a distributionally robust learning approach
Haoxuan Wang
Department of Computer Science and Engineering and MoE Key Lab of Artificial Intelligence, Shanghai Jiao Tong University
,Zhiding Yu
NVIDIA
,Yisong Yue
Department of Computing and Mathematical Sciences, California Institute of Technology
,Animashree Anandkumar
Department of Computing and Mathematical Sciences, California Institute of Technology
,Anqi Liu
Department of Computer Science, Johns Hopkins University
,Junchi Yan
Department of Computer Science and Engineering and MoE Key Lab of Artificial Intelligence, Shanghai Jiao Tong University
IJCAI '23: Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence•August 2023, Article No.: 162, pp 1460-1469• https://doi.org/10.24963/ijcai.2023/162We propose a framework for learning calibrated uncertainties under domain shifts, where the source (training) distribution differs from the target (test) distribution. We detect such domain shifts via a differentiable density ratio estimator and train it ...
- 0Citation
MetricsTotal Citations0
- research-article
Fast sampling of diffusion models via operator learning
Hongkai Zheng
Caltech
,Weilie Nie
NVIDIA
,Arash Vahdat
NVIDIA
,Kamyar Azizzadenesheli
NVIDIA
,Anima Anandkumar
Caltech and NVIDIA
ICML'23: Proceedings of the 40th International Conference on Machine Learning•July 2023, Article No.: 1784, pp 42390-42402Diffusion models have found widespread adoption in various areas. However, their sampling process is slow because it requires hundreds to thousands of network evaluations to emulate a continuous process defined by differential equations. In this work, we ...
- 0Citation
MetricsTotal Citations0
- research-article
I2SB: image-to-image Schrödinger bridge
Guan-Horng Liu
Georgia Institute of Technology
,Arash Vahdat
NVIDIA
,De-An Huang
NVIDIA
,Evangelos A. Theodorou
Georgia Institute of Technology
,Weili Nie
NVIDIA
,Anima Anandkumar
NVIDIA and California Institute of Technology
ICML'23: Proceedings of the 40th International Conference on Machine Learning•July 2023, Article No.: 915, pp 22042-22062We propose Image-to-Image Schrödinger Bridge (I2SB), a new class of conditional diffusion models that directly learn the nonlinear diffusion processes between two given distributions. These diffusion bridges are particularly useful for image restoration, ...
- 0Citation
MetricsTotal Citations0
- research-article
VIMA: robot manipulation with multimodal prompts
Yunfan Jiang
Stanford University
,Agrim Gupta
Stanford University
,Zichen Zhang
Macalester College and Allen Institute for AI
,Guanzhi Wang
NVIDIA and Caltech
,Yongqiang Dou
Tsinghua
,Yanjun Chen
Stanford University
,Li Fei-Fei
Stanford University
,Anima Anandkumar
NVIDIA and Caltech
,Yuke Zhu
NVIDIA and UT Austin
,Linxi Fan
NVIDIA
ICML'23: Proceedings of the 40th International Conference on Machine Learning•July 2023, Article No.: 611, pp 14975-15022Prompt-based learning has emerged as a successful paradigm in natural language processing, where a single general-purpose language model can be instructed to perform any task specified by input prompts. Yet task specification in robotics comes in various ...
- 0Citation
MetricsTotal Citations0
- research-article
Spherical Fourier neural operators: learning stable dynamics on the sphere
Boris Bonev
NVIDIA Corp., Santa Clara
,Thorsten Kurth
NVIDIA Corp., Santa Clara
,Christian Hundt
NVIDIA Corp., Santa Clara
,Jaideep Pathak
NVIDIA Corp., Santa Clara
,Maximilian Baust
NVIDIA Corp., Santa Clara
,Karthik Kashinath
NVIDIA Corp., Santa Clara
,Anima Anandkumar
NVIDIA Corp., Santa Clara and Caltech, Pasadena
ICML'23: Proceedings of the 40th International Conference on Machine Learning•July 2023, Article No.: 117, pp 2806-2823Fourier Neural Operators (FNOs) have proven to be an efficient and effective method for resolution-independent operator learning in a broad variety of application areas across scientific machine learning. A key reason for their success is their ability to ...
- 0Citation
MetricsTotal Citations0
Author Profile Pages
- Description: The Author Profile Page initially collects all the professional information known about authors from the publications record as known by the ACM bibliographic database, the Guide. Coverage of ACM publications is comprehensive from the 1950's. Coverage of other publishers generally starts in the mid 1980's. The Author Profile Page supplies a quick snapshot of an author's contribution to the field and some rudimentary measures of influence upon it. Over time, the contents of the Author Profile page may expand at the direction of the community.
Please see the following 2007 Turing Award winners' profiles as examples: - History: Disambiguation of author names is of course required for precise identification of all the works, and only those works, by a unique individual. Of equal importance to ACM, author name normalization is also one critical prerequisite to building accurate citation and download statistics. For the past several years, ACM has worked to normalize author names, expand reference capture, and gather detailed usage statistics, all intended to provide the community with a robust set of publication metrics. The Author Profile Pages reveal the first result of these efforts.
- Normalization: ACM uses normalization algorithms to weigh several types of evidence for merging and splitting names.
These include:- co-authors: if we have two names and cannot disambiguate them based on name alone, then we see if they have a co-author in common. If so, this weighs towards the two names being the same person.
- affiliations: names in common with same affiliation weighs toward the two names being the same person.
- publication title: names in common whose works are published in same journal weighs toward the two names being the same person.
- keywords: names in common whose works address the same subject matter as determined from title and keywords, weigh toward being the same person.
The more conservative the merging algorithms, the more bits of evidence are required before a merge is made, resulting in greater precision but lower recall of works for a given Author Profile. Many bibliographic records have only author initials. Many names lack affiliations. With very common family names, typical in Asia, more liberal algorithms result in mistaken merges.
Automatic normalization of author names is not exact. Hence it is clear that manual intervention based on human knowledge is required to perfect algorithmic results. ACM is meeting this challenge, continuing to work to improve the automated merges by tweaking the weighting of the evidence in light of experience.
- Bibliometrics: In 1926, Alfred Lotka formulated his power law (known as Lotka's Law) describing the frequency of publication by authors in a given field. According to this bibliometric law of scientific productivity, only a very small percentage (~6%) of authors in a field will produce more than 10 articles while the majority (perhaps 60%) will have but a single article published. With ACM's first cut at author name normalization in place, the distribution of our authors with 1, 2, 3..n publications does not match Lotka's Law precisely, but neither is the distribution curve far off. For a definition of ACM's first set of publication statistics, see Bibliometrics
- Future Direction:
The initial release of the Author Edit Screen is open to anyone in the community with an ACM account, but it is limited to personal information. An author's photograph, a Home Page URL, and an email may be added, deleted or edited. Changes are reviewed before they are made available on the live site.
ACM will expand this edit facility to accommodate more types of data and facilitate ease of community participation with appropriate safeguards. In particular, authors or members of the community will be able to indicate works in their profile that do not belong there and merge others that do belong but are currently missing.
A direct search interface for Author Profiles will be built.
An institutional view of works emerging from their faculty and researchers will be provided along with a relevant set of metrics.
It is possible, too, that the Author Profile page may evolve to allow interested authors to upload unpublished professional materials to an area available for search and free educational use, but distinct from the ACM Digital Library proper. It is hard to predict what shape such an area for user-generated content may take, but it carries interesting potential for input from the community.
Bibliometrics
The ACM DL is a comprehensive repository of publications from the entire field of computing.
It is ACM's intention to make the derivation of any publication statistics it generates clear to the user.
- Average citations per article = The total Citation Count divided by the total Publication Count.
- Citation Count = cumulative total number of times all authored works by this author were cited by other works within ACM's bibliographic database. Almost all reference lists in articles published by ACM have been captured. References lists from other publishers are less well-represented in the database. Unresolved references are not included in the Citation Count. The Citation Count is citations TO any type of work, but the references counted are only FROM journal and proceedings articles. Reference lists from books, dissertations, and technical reports have not generally been captured in the database. (Citation Counts for individual works are displayed with the individual record listed on the Author Page.)
- Publication Count = all works of any genre within the universe of ACM's bibliographic database of computing literature of which this person was an author. Works where the person has role as editor, advisor, chair, etc. are listed on the page but are not part of the Publication Count.
- Publication Years = the span from the earliest year of publication on a work by this author to the most recent year of publication of a work by this author captured within the ACM bibliographic database of computing literature (The ACM Guide to Computing Literature, also known as "the Guide".
- Available for download = the total number of works by this author whose full texts may be downloaded from an ACM full-text article server. Downloads from external full-text sources linked to from within the ACM bibliographic space are not counted as 'available for download'.
- Average downloads per article = The total number of cumulative downloads divided by the number of articles (including multimedia objects) available for download from ACM's servers.
- Downloads (cumulative) = The cumulative number of times all works by this author have been downloaded from an ACM full-text article server since the downloads were first counted in May 2003. The counts displayed are updated monthly and are therefore 0-31 days behind the current date. Robotic activity is scrubbed from the download statistics.
- Downloads (12 months) = The cumulative number of times all works by this author have been downloaded from an ACM full-text article server over the last 12-month period for which statistics are available. The counts displayed are usually 1-2 weeks behind the current date. (12-month download counts for individual works are displayed with the individual record.)
- Downloads (6 weeks) = The cumulative number of times all works by this author have been downloaded from an ACM full-text article server over the last 6-week period for which statistics are available. The counts displayed are usually 1-2 weeks behind the current date. (6-week download counts for individual works are displayed with the individual record.)
ACM Author-Izer Service
Summary Description
ACM Author-Izer is a unique service that enables ACM authors to generate and post links on both their homepage and institutional repository for visitors to download the definitive version of their articles from the ACM Digital Library at no charge.
Downloads from these sites are captured in official ACM statistics, improving the accuracy of usage and impact measurements. Consistently linking to definitive version of ACM articles should reduce user confusion over article versioning.
ACM Author-Izer also extends ACM’s reputation as an innovative “Green Path” publisher, making ACM one of the first publishers of scholarly works to offer this model to its authors.
To access ACM Author-Izer, authors need to establish a free ACM web account. Should authors change institutions or sites, they can utilize the new ACM service to disable old links and re-authorize new links for free downloads from a different site.
How ACM Author-Izer Works
Authors may post ACM Author-Izer links in their own bibliographies maintained on their website and their own institution’s repository. The links take visitors to your page directly to the definitive version of individual articles inside the ACM Digital Library to download these articles for free.
The Service can be applied to all the articles you have ever published with ACM.
Depending on your previous activities within the ACM DL, you may need to take up to three steps to use ACM Author-Izer.
For authors who do not have a free ACM Web Account:
- Go to the ACM DL http://dl.acm.org/ and click SIGN UP. Once your account is established, proceed to next step.
For authors who have an ACM web account, but have not edited their ACM Author Profile page:
- Sign in to your ACM web account and go to your Author Profile page. Click "Add personal information" and add photograph, homepage address, etc. Click ADD AUTHOR INFORMATION to submit change. Once you receive email notification that your changes were accepted, you may utilize ACM Author-izer.
For authors who have an account and have already edited their Profile Page:
- Sign in to your ACM web account, go to your Author Profile page in the Digital Library, look for the ACM Author-izer link below each ACM published article, and begin the authorization process. If you have published many ACM articles, you may find a batch Authorization process useful. It is labeled: "Export as: ACM Author-Izer Service"
ACM Author-Izer also provides code snippets for authors to display download and citation statistics for each “authorized” article on their personal pages. Downloads from these pages are captured in official ACM statistics, improving the accuracy of usage and impact measurements. Consistently linking to the definitive version of ACM articles should reduce user confusion over article versioning.
Note: You still retain the right to post your author-prepared preprint versions on your home pages and in your institutional repositories with DOI pointers to the definitive version permanently maintained in the ACM Digital Library. But any download of your preprint versions will not be counted in ACM usage statistics. If you use these AUTHOR-IZER links instead, usage by visitors to your page will be recorded in the ACM Digital Library and displayed on your page.
FAQ
- Q. What is ACM Author-Izer?
A. ACM Author-Izer is a unique, link-based, self-archiving service that enables ACM authors to generate and post links on either their home page or institutional repository for visitors to download the definitive version of their articles for free.
- Q. What articles are eligible for ACM Author-Izer?
- A. ACM Author-Izer can be applied to all the articles authors have ever published with ACM. It is also available to authors who will have articles published in ACM publications in the future.
- Q. Are there any restrictions on authors to use this service?
- A. No. An author does not need to subscribe to the ACM Digital Library nor even be a member of ACM.
- Q. What are the requirements to use this service?
- A. To access ACM Author-Izer, authors need to have a free ACM web account, must have an ACM Author Profile page in the Digital Library, and must take ownership of their Author Profile page.
- Q. What is an ACM Author Profile Page?
- A. The Author Profile Page initially collects all the professional information known about authors from the publications record as known by the ACM Digital Library. The Author Profile Page supplies a quick snapshot of an author's contribution to the field and some rudimentary measures of influence upon it. Over time, the contents of the Author Profile page may expand at the direction of the community. Please visit the ACM Author Profile documentation page for more background information on these pages.
- Q. How do I find my Author Profile page and take ownership?
- A. You will need to take the following steps:
- Create a free ACM Web Account
- Sign-In to the ACM Digital Library
- Find your Author Profile Page by searching the ACM Digital Library for your name
- Find the result you authored (where your author name is a clickable link)
- Click on your name to go to the Author Profile Page
- Click the "Add Personal Information" link on the Author Profile Page
- Wait for ACM review and approval; generally less than 24 hours
- Q. Why does my photo not appear?
- A. Make sure that the image you submit is in .jpg or .gif format and that the file name does not contain special characters
- Q. What if I cannot find the Add Personal Information function on my author page?
- A. The ACM account linked to your profile page is different than the one you are logged into. Please logout and login to the account associated with your Author Profile Page.
- Q. What happens if an author changes the location of his bibliography or moves to a new institution?
- A. Should authors change institutions or sites, they can utilize ACM Author-Izer to disable old links and re-authorize new links for free downloads from a new location.
- Q. What happens if an author provides a URL that redirects to the author’s personal bibliography page?
- A. The service will not provide a free download from the ACM Digital Library. Instead the person who uses that link will simply go to the Citation Page for that article in the ACM Digital Library where the article may be accessed under the usual subscription rules.
However, if the author provides the target page URL, any link that redirects to that target page will enable a free download from the Service.
- Q. What happens if the author’s bibliography lives on a page with several aliases?
- A. Only one alias will work, whichever one is registered as the page containing the author’s bibliography. ACM has no technical solution to this problem at this time.
- Q. Why should authors use ACM Author-Izer?
- A. ACM Author-Izer lets visitors to authors’ personal home pages download articles for no charge from the ACM Digital Library. It allows authors to dynamically display real-time download and citation statistics for each “authorized” article on their personal site.
- Q. Does ACM Author-Izer provide benefits for authors?
- A. Downloads of definitive articles via Author-Izer links on the authors’ personal web page are captured in official ACM statistics to more accurately reflect usage and impact measurements.
Authors who do not use ACM Author-Izer links will not have downloads from their local, personal bibliographies counted. They do, however, retain the existing right to post author-prepared preprint versions on their home pages or institutional repositories with DOI pointers to the definitive version permanently maintained in the ACM Digital Library.
- Q. How does ACM Author-Izer benefit the computing community?
- A. ACM Author-Izer expands the visibility and dissemination of the definitive version of ACM articles. It is based on ACM’s strong belief that the computing community should have the widest possible access to the definitive versions of scholarly literature. By linking authors’ personal bibliography with the ACM Digital Library, user confusion over article versioning should be reduced over time.
In making ACM Author-Izer a free service to both authors and visitors to their websites, ACM is emphasizing its continuing commitment to the interests of its authors and to the computing community in ways that are consistent with its existing subscription-based access model.
- Q. Why can’t I find my most recent publication in my ACM Author Profile Page?
- A. There is a time delay between publication and the process which associates that publication with an Author Profile Page. Right now, that process usually takes 4-8 weeks.
- Q. How does ACM Author-Izer expand ACM’s “Green Path” Access Policies?
- A. ACM Author-Izer extends the rights and permissions that authors retain even after copyright transfer to ACM, which has been among the “greenest” publishers. ACM enables its author community to retain a wide range of rights related to copyright and reuse of materials. They include:
- Posting rights that ensure free access to their work outside the ACM Digital Library and print publications
- Rights to reuse any portion of their work in new works that they may create
- Copyright to artistic images in ACM’s graphics-oriented publications that authors may want to exploit in commercial contexts
- All patent rights, which remain with the original owner