ACC Seminar: Applications of Transformers in Mathematics
Department of Mathematical Sciences
Location: North Building Room 316 and Zoom (Passcode: ACC)
Speaker: Paul Schwartz, Lecturer of Mathematics, Stevens Institute of Technology
ABSTRACT
Since the introduction of the transformer architecture in 2017 [Google DeepMind], large language models (LLMs) have a meteoric rise. We will discuss how the transformer works and state of the art methods for incentivizing reasoning by LLMs [Deepseek]. We will discuss some of the recent successes that LLMs have had in solving mathematics problems as well as the difficulties that still exist [DeepMind et al.]. Finally, we will discuss a result from March 2026 which proves that there are theoretical limitations to embedding-based retrieval [DeepMind] and pivot toward other machine learning methods that have contributed to recent mathematical discoveries.
BIOGRAPHY
Paul Schwartz received his PhD from the University of Florida in 2022 under the direction of Kevin Keating. He currently serves as a Lecturer of Mathematics at Stevens Institute of Technology.
Attendance: This is a technical talk open to all.
A campus map is available at https://tour.stevens.edu.
Additional information is available at https://web.stevens.edu/algebraic/.
At any time, photography or videography may be occurring on Stevens’ campus. Resulting footage may include the image or likeness of event attendees. Such footage is Stevens’ property and may be used for Stevens’ commercial and/or noncommercial purposes. By registering for and/or attending this event, you consent and waive any claim against Stevens related to such use in any media. See Stevens' Privacy Policy for more information.
