Nima Shoghi
I’m an AI researcher with a background in computer science and machine learning. I earned both my B.S. and M.S. degrees in Computer Science from Georgia Tech. During my time there, I conducted research at the High Performance Computer Architecture Lab, where I focused on accelerating ML training and inference. I recently concluded my two-year AI residency at Meta AI’s FAIR Chemistry team, where I worked on developing large pre-trained models, trained on a large mixture of available chemical data across multiple different domains, for general-purpose chemical property prediction. I am particularly interested in the application of large-scale machine learning techniques to problems in science and engineering.
My CV is available here.
Recent Updates
- [Jan 2024] Our paper on large-scale diverse pre-training for chemical property prediction (also available on arXiv) has been accepted to ICLR 2024!
- [Dec 2023] I will be joining the High Performance Computer Architecture Lab at Georgia Tech as a temporary research staff member in December 2023!
Education
- M.S. with Highest Honors in Computer Science (Machine Learning Specialization), Georgia Institute of Technology, Jan 2020 - May 2021
- B.S. with High Honors in Computer Science (Machine Learning and Devices Threads), Georgia Institute of Technology, May 2015 - Dec 2019
- International Baccalaureate Diploma, Druid Hills High School, Aug 2011 - May 2015
Work experience
- Temporary Research Staff at the High Performance Computer Architecture Lab at Georgia Tech, Dec 2023 - Present
- Working on efficient inference strategies for pre-trained image diffusion models, with a focus on generating diverse, high-quality images.
- Advisors: Dr. Hyesoon Kim and Dr. Stefano Petrangeli
- AI Resident at Meta Fundamental AI Research (FAIR), Aug 2021 - Aug 2023
- Worked on the Open Catalyst Project on the FAIR Chemistry team, focusing on the development of large-scale pre-training methods for chemical property prediction.
- Advisors: Dr. Larry Zitnick and Dr. Abhishek Das
- Research Assistant at High Performance Computer Architecture Lab at Georgia Tech, May 2019 - May 2021
- Developed software-level and hardware-level techniques for accelerating deep learning training and inference.
- Advisors: Dr. Hyesoon Kim and Dr. Moinuddin Qureshi
- Graduate Teaching Assistant at Georgia Institute of Technology, Aug 2020 - May 2021
- CS 4510: Automata and Complexity, Spring 2021, Taught by Dr. Zvi Galil
- CS 4510: Automata and Complexity, Fall 2020, Taught by Dr. Merrick Furst
- Student Research Assistant at the Cyber Forensics Innovation (CyFI) Lab at Georgia Tech, Jan 2020 - Aug 2020
- Developed GNN-based ML models to analyze social media data for detecting incoming cyber attacks.
- Advisor: Dr. Maria Konte
- Software Engineering Intern at Ciena, May 2017 - May 2018
- Developed software to interface with network devices and maintained CI/CD pipelines for build processes.
Skills
- Data Science and Machine Learning:
- Proficient in Python data science libraries: NumPy, Pandas, Matplotlib, and Seaborn.
- Extensive experience with deep learning libraries: PyTorch and JAX.
- Experience with high-performance computing (HPC) and distributed (e.g., 128+ GPUs) training.
- Programming and Development:
- Proficient in Python, C, C++, Rust, C#, and JavaScript/TypeScript
- Experience with test-driven development, including unit tests, integration tests, and end-to-end tests.
- In-depth knowledge of virtualization, containers, and Docker/Podman/Singularity.
Publications
(* denotes equal contribution)
From Molecules to Materials: Pre-training Large Generalizable Models for Atomic Property Prediction
Nima Shoghi, Adeesh Kolluru, John Kitchin, Zachary Ulissi, C. Zitnick, Brandon Wood, arXiv preprint arXiv:2310.16802, 2023.
Context-Aware Task Handling in Resource-Constrained Robots with Virtualization
Ramyad Hadidi, Nima Shoghi, Bahar Asgari, Hyesoon Kim, IEEE International Conference on Edge Computing and Communications, 2023.
The Open Catalyst 2022 (OC22) dataset and challenges for oxide electrocatalysts
Richard Tran, Janice Lan, Muhammed Shuaibi, Brandon Wood, Siddharth Goyal, Abhishek Das, Javier Heras-Domingo, Adeesh Kolluru, Ammar Rizvi, Nima Shoghi, Anuroop Sriram, Félix Therrien, Jehad Abed, Oleksandr Voznyy, Edward Sargent, Zachary Ulissi, C. Zitnick, ACS Catalysis, 2023.
Open Challenges in Developing Generalizable Large-Scale Machine-Learning Models for Catalyst Discovery
Adeesh Kolluru, Muhammed Shuaibi, Aini Palizhati, Nima Shoghi, Abhishek Das, Brandon Wood, C Zitnick, John Kitchin, Zachary Ulissi, ACS Catalysis, 2022.
Transfer learning using attentions across atomic systems with graph neural networks (TAAG)
Adeesh Kolluru, Nima Shoghi, Muhammed Shuaibi, Siddharth Goyal, Abhishek Das, C Zitnick, Zachary Ulissi, The Journal of Chemical Physics, 2022.
SmaQ: Smart Quantization for DNN Training by Exploiting Value Clustering
Nima Shoghi*, Andrei Bersatti*, Moinuddin Qureshi, Hyesoon Kim, IEEE Computer Architecture Letters, 2021.
Quantifying the design-space tradeoffs in autonomous drones
Ramyad Hadidi, Bahar Asgari, Sam Jijina, Adriana Amyette, Nima Shoghi, Hyesoon Kim, Proceedings of the 26th ACM International Conference on Architectural Support for Programming Languages and Operating Systems, 2021.
Secure Location-Aware Authentication and Communication for Intelligent Transportation Systems
Nima Shoghi, Ramyad Hadidi, Lee Jaewon, Jun Chen, Arthur Siqueria, Rahul Rajan, Shaan Dhawan, Pooya Shoghi, Hyesoon Kim, arXiv preprint arXiv:2011.08936, 2020.
Understanding the Software and Hardware Stacks of a General-Purpose Cognitive Drone
Sam Jijina, Adriana Amyette, Nima Shoghi, Ramyad Hadidi, Hyesoon Kim, 2020 IEEE International Symposium on Performance Analysis of Systems and Software (ISPASS), 2020.
PISCES: Power-Aware Implementation of SLAM by Customizing Efficient Sparse Algebra
Bahar Asgari, Ramyad Hadidi, Nima Shoghi, Hyesoon Kim, 2020 57th ACM/IEEE Design Automation Conference (DAC), 2020.
Neural Network Weight Compression with NNW-BDI
Andrei Bersatti*, Nima Shoghi*, Hyesoon Kim, The International Symposium on Memory Systems, 2020.
SLAM Performance on Embedded Robots
Nima Shoghi, Ramyad Hadidi, Hyesoon Kim, Student Research Competition at Embedded System Week (SRC ESWEEK), 2019.
Talks
From Molecules to Materials: Pre-training Large Generalizable Models for Atomic Property Prediction
Talk at 2023 ACS Fall, San Francisco, California
SmaQ: Smart Quantization for DNN Training by Exploiting Value Clustering
Talk at CS 7290: Advanced Microarchitecture (Georgia Tech), Atlanta, Georgia
Legal Text Summarization Using Transformer Models
Talk at CS 8803-DLT: Deep Learning for Text (Georgia Tech), Atlanta, Georgia
Attention is All You Need
Talk at CS 8803-DLT: Deep Learning for Text (Georgia Tech), Atlanta, Georgia