
Date: Tuesday, February 18, 2025 | Time: 3:00 PM - 4:00 PM
The National Institute of Statistical Sciences (NISS) recently hosted an insightful discussion on large language models (LLMs) as part of its AI, Statistics & Data Science in Practice series. Dr. Frank Wei, Principal AI ML Scientist and Director of Data Scientists at General Motors, shared his expertise on the architecture, applications, and implications of LLMs, focusing on their transformative impact on artificial intelligence (AI) and industry operations. The event was moderated by Nancy McMillan (Battelle).
Exploring Transformer Architectures
Dr. Wei provided an in-depth look at transformer models, the backbone of modern LLMs. He explained key components such as multi-head self-attention and residual connections, which allow transformers to process vast amounts of text data effectively. He also discussed recent innovations, including Levy attention with linear biases and multi-head attention mechanisms, improving model performance for long sequences and complex language tasks.
Applications and Real-World Impact
Highlighting practical applications, Dr. Wei demonstrated how LLMs are being utilized in industries such as healthcare, finance, education, and e-commerce. He showcased an LLM-powered solution, emphasizing its potential to revolutionize AI-driven decision-making. The discussion extended to workforce transformation, addressing the social implications of AI, including bias, fairness, and automation in various sectors.
Challenges and Considerations
Despite their advantages, LLMs come with challenges. Dr. Wei discussed the high computational costs, issues of interpretability, and scalability concerns. He differentiated between GPT models optimized for text generation and BERT models tailored for classification tasks. Additionally, he highlighted recent advancements in sparse and multimodal models, which aim to improve efficiency and adaptability.
Security, Ethics, and Environmental Concerns
Security and ethical considerations were central topics in the discussion. Dr. Wei addressed concerns such as data leakage, bias amplification, and privacy risks. He emphasized the importance of transparency, algorithmic audits, and mitigation strategies to ensure fairness in AI systems. Additionally, he shed light on the environmental impact of training LLMs, underscoring the need for sustainable AI practices, including reinforcement learning with human feedback and multimodal AI innovations.
Introducing AI-Driven Career Enhancement
One of the highlights of the session was Dr. Wei’s introduction of a new AI-driven career enhancement tool. Designed to assist job seekers and hiring managers, this tool helps users generate tailored resumes and cover letters based on job descriptions. It also includes a mock interview feature, enhancing candidate preparation. Dr. Wei emphasized its efficiency and precision, making job applications more streamlined and effective.
The Future of AI-Generated Content
Concluding the discussion, Dr. Wei reflected on the growing reliance on AI-generated content and its implications for originality and misinformation detection. He acknowledged concerns about LLMs training on their own outputs, potentially reducing creativity and diversity in generated content. He also addressed the role of AI in detecting misinformation, emphasizing that effectiveness depends on training data and contextual factors.
Acknowledgments and Appreciation
NISS extends its sincere gratitude to Dr. Frank Wei for sharing his expertise and providing valuable insights into the world of large language models. His deep knowledge and engaging discussion helped attendees gain a better understanding of the transformative potential and challenges of AI-driven technologies. A special thank you also goes to Nancy McMillan for skillfully moderating the event and facilitating an engaging and thought-provoking conversation. Her contributions ensured a dynamic discussion that covered both technical and real-world implications of LLMs. We appreciate the participation of all attendees and look forward to continuing these important conversations in future NISS events.
Looking Ahead
The session provided valuable insights into the evolving landscape of AI and LLMs, sparking discussions on the balance between innovation, ethics, and sustainability. As AI continues to advance, industry leaders like Dr. Wei are paving the way for responsible development and deployment of transformative technologies.
About AI, Statistics and Data Science in Practice Webinars
The NISS AI, Statistics and Data Science in Practice is a monthly event series brings together leading experts from industry and academia to discuss the latest advances and practical applications in AI, data science, and statistics. Each session features a keynote presentation on cutting-edge topics, where attendees can engage with speakers on the challenges and opportunities in applying these technologies in real-world scenarios. This series is intended for professionals, researchers, and students interested in the intersection of AI, data science, and statistics, offering insights into how these fields are shaping various industries. The series is designed to provide participants with exposure to and understanding of how modern data analytic methods are being applied in real-world scenarios across various industries, offering both theoretical insights, practical examples, and discussion of issues.
During Spring 2026, from January through May 2026, the series will focus on large language models (LLMs) and the statistical and methodological foundations required to develop, evaluate, and deploy them responsibly and effectively. As LLMs become central to a wide range of scientific, industrial, and societal applications, careful attention to data generation, model training, evaluation, and inference is essential to ensure reliability, robustness, and transparency. As LLMs become increasingly central to scientific research, industry workflows, and societal decision-making, rigorous attention to how training data are constructed, curated, and sampled is critical for understanding model behavior and limitations. The series will highlight methodological considerations in model training and fine-tuning, including sources of bias, variability, and uncertainty, as well as principled approaches to benchmarking and evaluation that move beyond surface-level performance metrics. Emphasis will be placed on transparent and reproducible evaluation frameworks that support meaningful comparisons across models and use cases, and on statistical perspectives that help clarify what LLM outputs do and do not represent. By grounding discussions of LLM development and deployment in sound statistical reasoning, the series aims to promote more reliable, interpretable, and trustworthy language models in practice.
- Veridical Data Science - Speaker: Bin Yu, October 15,2024
- Random Forests: Why they Work and Why that’s a Problem - Speaker: Lucas Mentch, November 19, 2024
- Causal AI in Business Practices - Speakers: Victor Lo, and Victor Chen, January 24, 2025
- Large Language Models: Transforming AI Architectures and Operational Paradigms - Speaker: Frank Wei, February 18, 2025
- Machine Learning for Airborne Biological Hazard Detection - Speaker: Jared Schuetter, March 11, 2025
- Trustworthy AI in Weather, Climate, and Coastal Oceanography - Speaker: Dr. Amy McGovern, May 13, 2025
- Sequential Causal Inference in Experimental or Observational Settings - Speaker: Aaditya Ramdas, August 26, 2025
- Covariate Adjustment, Intro to Resampling, and Surprises - Speaker: Tim Hesterberg, October 3, 2025
- Bayesian Geospatial Approaches for Prediction of Opioid Overdose Deaths Utilizing the Real-Time Urine Drug Test - Speaker: Joanne Kim, November 18, 2025
- COVID-19 Focused Cost-benefit Analysis of Public Health Emergency Preparedness and Crisis Response Programs - Speaker: Nancy McMillan, December 11, 2025
- LabOS: The AI-XR Co-Scientist That Reasons, Sees and Works With Humans - Speaker: Mengdi Wang, January 20, 2026
- From LLMs to World Foundation Models & Robotics: The Next Frontier of Artificial Intelligence - Speaker: Robert Clark, February 24, 2026

