Jinwook Oh
Jinwook Oh is the Co-Founder and Chief Technology Officer of Rebellions, an AI chip company based in South Korea. After earning his Ph.D. from KAIST (Korea Advanced Institute of Science and Technology), he joined the IBM TJ Watson Research Center, where he contributed to several AI chip R&D projects as a Chip Architect, Logic Designer, and Logic Power Lead. At Rebellions, he has overseen the development and launch of two AI chips, with a third, REBEL, in progress. Jinwook's technical leadership has been crucial in establishing Rebellions as a notable player in AI technology within just three and a half years.
Rebellions
Website: https://rebellions.ai/
The founding team of Rebellions relocated to Korea from New York and elsewhere in 2020 to revolutionize AI chip industry.
At the heart of the Korean Silicon Eco-system, Rebellions has built a cutting-edge AI inference accelerator and full-stack software optimized for generative AI.
Within just three years of its inception, the company has introduced two groundbreaking chips: the finance market focused ION, released in 2021, and the datacenter-focused ATOM, taped-out in 2023. ATOM has demonstrated its superior performance in the MLPerf benchmarks and has been commercialized in a data center through a strategic partnership with KT(Korea Telecom), the biggest IDC company in South Korea.
Currently, Rebellions is developing its next-generation AI chip, REBEL, equipped with HBM3E in a collaboration with Samsung Electronics, paving the way for the advanced technology in the era of generative AI.
In this presentation, we will explore the advanced integration of Digital In-Memory Computing (D-IMC) and RISC-V technology by Axelera AI to accelerate AI inference workloads. Our approach uniquely combines the high energy efficiency and throughput of D-IMC with the versatility of RISC-V technology, creating a powerful and scalable platform. This platform is designed to handle a wide range of AI tasks, from advanced computer vision at the edge to emerging AI challenges.
We will demonstrate how our scalable architecture not only meets but exceeds the demands of modern AI applications. Our platform enhances performance while significantly reducing energy use and operational costs. By pushing the boundaries of Edge AI and venturing into new AI domains, Axelera AI is setting new benchmarks in AI processing efficiency and deployment capabilities.
Evangelos Eleftheriou
Evangelos Eleftheriou, an IEEE and IBM Fellow, is the Chief Technology Officer and co-founder of Axelera AI, a best-in-class performance company that develops a game-changing hardware and software platform for AI.
As a CTO, Evangelos oversees the development and dissemination of technology for external customers, vendors, and other clients to help improve and increase Axelera AI’s business.
Before his current role, Evangelos worked for IBM Research – Zurich, where he held various management positions for over 35 years. His outstanding achievements led him to become an IBM Fellow, which is IBM’s highest technical honour.
In 2002, Evangelos became a Fellow of the IEEE, and later in 2003, he was co-recipient of the IEEE ComS Leonard G. Abraham Prize Paper Award. He was also co-recipient of the 2005 Technology Award of the Eduard Rhein Foundation. In 2005, he was appointed an IBM Fellow and inducted into the IBM Academy of Technology. In 2009, he was co-recipient of the IEEE Control Systems Technology Award and the IEEE Transactions on Control Systems Technology Outstanding Paper Award. In 2016, Evangelos received an honoris causa professorship from the University of Patras, Greece. In 2018, he was inducted into the US National Academy of Engineering as Foreign Member. Evangelos has authored or coauthored over 250 publications and holds over 160 patents (granted and pending applications).
His primary interests lie in AI and machine learning, including emerging computing paradigms such as neuromorphic and in-memory computing.
Evangelos holds a PhD and a Master of Eng. degrees in Electrical Engineering from Carleton University, Canada, and a BSc in Electrical & Computer Engineering from the University of Patras, Greece.
Axelera
Website: https://www.axelera.ai/
Axelera AI is delivering the world’s most powerful and advanced solutions for AI at the Edge. Its industry-defining Metis™ AI platform – a complete hardware and software solution for AI inference at the edge – makes computer vision applications more accessible, powerful and user friendly than ever before. Based in the AI Innovation Center of the High Tech Campus in Eindhoven, The Netherlands, Axelera AI has R&D offices in Belgium, Switzerland, Italy and the UK, with over 170 employees in 18 countries. Its team of experts in AI software and hardware come from top AI firms and Fortune 500 companies.
For more information on Axelera AI, see: www.axelera.ai
Bijan Nowroozi
Bijan Nowroozi is Chief Technical Officer of The Open Compute Project Foundation and has more than 30 years of experience with hardware and software development, signal processing, networking, and research with technology companies. Prior to The OCP Bijan developed mission critical infrastructure and was on the leading-edge standards and technology development in multiple technology waves including edge computing, AI/ML, optical/photonics, quantum, RF, wireless, small cells, UAV’s, GIS, HPC, network security, energy and more.
Sadasivan Shankar
Sadasivan (Sadas) Shankar is Research Technology Manager at SLAC National Laboratory, adjunct Professor in Stanford Materials Science and Engineering, and Lecturer in the Stanford Graduate School of Business. He was an Associate in the Department of Physics at Harvard University, and was the first Margaret and Will Hearst Visiting Lecturer in Harvard and the first Distinguished Scientist in Residence at the Harvard Institute of Applied Computational Sciences. He has co-instructed classes related to design of materials, computing, sustainability in materials, and has received Excellence in Teaching award from Harvard University. He is co-instructing a class at Stanford University on Translation for
Innovations. He is a co-founder of and the Chief Scientist at Material Alchemy, a “last mile” translational and independent venture that has been recently founded to accelerate the path from materials discovery to adoption, with environmental sustainability as a key goal. In addition to research on fundamentals of Materials Design, his current research is on new architectures for specialized AI methods is exploring ways of bringing machine intelligence to system-level challenges in inorganic/biochemistry, materials, and physics and new frameworks for computing as information processing inspired by lessons from
nature.
Dr. Shankar’s current research and analysis on Sustainable Computing is helping provide directions for the US Department of Energy’s EES2 scaling initiatives (energy reduction in computing every generation for 1000X reduction in 2 decades) as part of the White House Plan to Revitalize American Manufacturing and Secure Critical Supply Chains in 2022 for investment in research, development, demonstration, and commercial application (RDD&CA) in conventional semiconductors.
In addition, his analysis is helping identify pathways for energy efficient computing. While in the industry, Dr. Shankar and his team have enabled several critical technology decisions in the semiconductor industrial applications of chemistry, materials, processing, packaging, manufacturing, and design rules for over nine generations of Moore’s law including first advanced
process control application in 300 mm wafer technology; introduction of flip chip packaging using electrodeposition, 100% Pb-elimination in microprocessors, design of new materials, devices including nano warp-around devices for the advanced semiconductor technology manufacturing, processing
methods, reactors, etc. Dr. Shankar managed his team members distributed across multiple sites in the US, with collaborations in Europe. The teams won several awards from the Executive Management and technology organizations.
He is a co-inventor in over twenty patent filings covering areas in new
chemical reactor designs, semiconductor processes, bulk and nano materials for the sub 10 nanometer generation of transistors, device structures, and algorithms. He is also a co-author in over hundred publications and presentations in measurements, multi-scale and multi-physics methods spanning from quantum scale to macroscopic scales, in the areas of chemical synthesis, plasma chemistry and processing, non-equilibrium electronic, ionic, and atomic transport, energy efficiency of information processing, and machine learning methods for bridging across scales, and estimating complex materials
properties and in advanced process control.
Dr. Shankar was an invited speaker at the Clean-IT Conference in Germany on Revolutionize Digital Systems and AI (2023), Telluride Quantum Inspired Neuromorphic Computing Workshop (2023) on Limiting Energy Estimates for Classical and Quantum Information Processing, Argonne National
Laboratory Director’s Special Colloquium on the Future of Computing (2022), panelist on Carnegie Science series on Brain and Computing (2020), lecturer in the Winter Course on Computational Brain Research in IIT-M-India (2020), invited participant in the Kavli Institute of Theoretical Physics program
on Cellular Energetics in UCSB (2019), invited speaker to the Camille and Henry Dreyfus Foundation meeting on Machine Learning for problems in Chemistry and Materials Science (2019), a Senior Fellow in UCLA Institute of Pure and Applied Mathematics during the program on Machine Learning and Manybody
Physics (2016), invited to the White House event for starting of the Materials Genome Initiative (2012), Invited speaker in Erwin Schrödinger International Institute for Mathematical Physics-Vienna (2007), Intel’s first Distinguished Lecturer in Caltech (1998) and MIT (1999). He has also given several
colloquia and lectures in universities all over the world and his research was also featured in the publications Science (2012), TED (2013), Nature Machine Intelligence (2022), Nature Physics (2022).
Ankita Singh
Dil Radhakrishnan
Dileeshvar Radhakrishnan is an engineer at MinIO where he focuses on AI/ML and Kubernetes. Prior to joining MinIO, Dil served as Chief Architect at ML pioneer Espressive. He previously worked in engineering roles at ServiceNow and Rewyndr. He began his career at Tata Consulting Services.
Dil has Bachelor of Engineering in Computer Science and Engineering from Anna University and a Masters in Computer Science from Carnegie Mellon.
Open Compute Project Foundation
Website: http://www.opencompute.org
The Open Compute Project (OCP) is a global collaborative Community of hyperscale data center operators, telecom, colocation providers and enterprise IT users, working with the product and solution vendor ecosystem to develop open innovations deployable from the cloud to the edge. The OCP Foundation is responsible for fostering and serving the OCP Community to meet the market and shape the future, taking hyperscale-led innovations to everyone. Meeting the market is accomplished through addressing challenging market obstacles with open specifications, designs and emerging market programs that showcase OCP-recognized IT equipment and data center facility best practices. Shaping the future includes investing in strategic initiatives and programs that prepare the IT ecosystem for major technology changes, such as AI & ML, optics, advanced cooling techniques, composable memory and silicon. OCP Community-developed open innovations strive to benefit all, optimized through the lens of impact, efficiency, scale and sustainability.
Hira Dangol
Industry experience in AI/ML, engineering, architecture and executive roles in leading technology companies, service providers and Silicon Valley leading organizations. Currently focusing on innovation, disruption, and cutting-edge technologies through startups and technology-driven corporation in solving the pressing problems of industry and world.
Puja Das
Dr. Puja Das, leads the Personalization team at Warner Brothers Discovery (WBD) which includes offerings on Max, HBO, Discovery+ and many more.
Prior to WBD, she led a team of Applied ML researchers at Apple, who focused on building large scale recommendation systems to serve personalized content on the App Store, Arcade and Apple Books. Her areas of expertise include user modeling, content modeling, recommendation systems, multi-task learning, sequential learning and online convex optimization. She also led the Ads prediction team at Twitter (now X), where she focused on relevance modeling to improve App Ads personalization and monetization across all of Twitter surfaces.
She obtained her Ph.D from University of Minnesota in Machine Learning, where the focus of her dissertation was online learning algorithms, which work on streaming data. Her dissertation was the recipient of the prestigious IBM Ph D. Fellowship Award.
She is active in the research community and part of the program committee at ML and recommendation system conferences. Shas mentored several undergrad and grad students and participated in various round table discussions through Grace Hopper Conference, Women in Machine Learning Program colocated with NeurIPS, AAAI and Computing Research Association- Women’s chapter.
Logan Grasby
Logan Grasby is a Senior Machine Learning Engineer at Cloudflare, based in Calgary, Alberta. As part of Cloudflare's Workers AI team he works on developing, deploying and scaling AI inference servers across Cloudflare's edge network. In recent work he has designed services for multi-tenant LLM LoRA inference and dynamic diffusion model pipeline servers. Prior to Cloudflare, Logan founded Azule, an LLM driven customer service and product recommendation platform for ecommerce. He also co-founded Conversion Pages and served as Director of Product at Appstle, a Shopify app development firm.
Daniel Valdivia
Daniel Valdivia is an engineer with MinIO where he focuses on Kubernetes, ML/AI and VMware. Prior to joining MinIO, Daniel was the Head of Machine Learning for Espressive. Daniel has held senior application development roles with ServiceNow, Oracle and Freescale. Daniel holds a Bachelor of Engineering from Tecnológico de Monterrey, Campus Guadalajara and Bachelor of Science in Computer Engineering from Instituto Tecnológico y de Estudios Superiores de Monterrey.
Baskar Sridharan
Baskar Sridharan is the Vice President for AI/ML and Data Services & Infrastructure at AWS, where he oversees the strategic direction and development of key services, including Amazon Bedrock, Amazon SageMaker, and essential data platforms like Amazon EMR, Amazon Athena, and AWS Glue.
Prior to his current role, Baskar spent nearly six years at Google, where he contributed to advancements in cloud computing infrastructure. Before that, he dedicated 16 years to Microsoft, playing a pivotal role in the development of Azure Data Lake and Cosmos, which have significantly influenced the landscape of cloud storage and data management.
Baskar earned a Ph.D. in Computer Science from Purdue University and has since spent over two decades at the forefront of the tech industry.
He has lived in Seattle for over 20 years, where he, his wife, and two children embrace the beauty of the Pacific Northwest and its many outdoor activities. In his free time, Baskar enjoys practicing music and playing cricket and baseball with his kids.
John Overton
John Overton is the CEO of Kove IO, Inc. In the late 1980s, while at the Open Software Foundation, Dr. Overton wrote software that went on to be used by approximately two thirds of the world’s workstation market. In the 1990s, he co-invented and patented technology utilizing distributed hash tables for locality management, now widely used in storage, database, and numerous other markets. In the 2000s, he led development of the first truly capable Software-Defined Memory offering, Kove:SDM™. Kove:SDM™ enables new Artificial Intelligence and Machine Learning capabilities, while also reducing power by up to 50%. Dr. Overton has more than 65 issued patents world-wide and has peer-reviewed publications across numerous academic disciplines. He holds post-graduate and doctoral degrees from Harvard and the University of Chicago.
Bill Wright
Arne Stoschek
Arne is the Vice President of AI, Autonomy & Digital Information and oversees the company’s development of autonomous flight and machine learning solutions to enable future, self-piloted aircraft. In his role, he also leads the advancement of large-scale data-driven processes to develop novel aircraft functions. He is passionate about robotics, autonomy and the impact these technologies will have on future mobility. After holding engineering leadership positions at global companies such as Volkswagen/Audi and Infineon, and at aspiring Silicon Valley startups, namely Lucid Motors/Atieva, Knightscope and Better Place, Arne dared to take his unique skill set to altitude above ground inside Airbus. Arne earned a Doctor of Philosophy in Electrical and Computer Engineering from the Technical University of Munich and held a computer vision and data analysis research position at Stanford University.
In this talk, Dr. Vinesh Sukumar will explain how Qualcomm has been successful in deploying large generative AI models on the edge for a variety of use cases in consumer and enterprise markets. He will examine key challenges that must be overcome before large models at the edge can reach their full commercial potential. He’ll also highlight how the industry is addressing these challenges and explore emerging large multimodal models.
Vinesh Sukumar
Vinesh Sukumar currently serves as Senior Director – Head of AI/ML product management at Qualcomm Technologies, Inc (QTI). In this role, he leads AI product definition, strategy and solution deployment across multiple business units.
•He has about 20 years of industry experience spread across research, engineering and application deployment. He currently holds a doctorate degree specializing in imaging and vision systems while also completing a business degree focused on strategy and marketing. He is a regular speaker in many AI industry forums and has authored several journal papers and two technical books.
Qualcomm
Website: https://www.qualcomm.com/
Qualcomm relentlessly innovates to deliver intelligent computing everywhere, helping the world tackle some of its most important challenges. Their leading-edge AI, high performance, low-power computing, and unrivaled connectivity deliver proven solutions that transform major industries. At Qualcomm, the team are engineering human progress.
Join Mark Russinovich, Azure CTO and Technical Fellow, for an in-depth exploration of Microsoft's AI architecture. Discover the technology behind our sustainable datacenter design, massive supercomputers used for foundational model training, efficient infrastructure for serving models, workload management and optimizations, AI safety, and advancements in confidential computing to safeguard data during processing.
Mark Russinovich
Mark Russinovich is Chief Technology Officer and Technical Fellow for Microsoft Azure, Microsoft’s global enterprise-grade cloud platform. A widely recognized expert in distributed systems, operating systems and cybersecurity, Mark earned a Ph.D. in computer engineering from Carnegie Mellon University. He later co-founded Winternals Software, joining Microsoft in 2006 when the company was acquired. Mark is a popular speaker at industry conferences such as Microsoft Ignite, Microsoft Build, and RSA Conference. He has authored several nonfiction and fiction books, including the Microsoft Press Windows Internals book series, Troubleshooting with the Sysinternals Tools, as well as fictional cyber security thrillers Zero Day, Trojan Horse and Rogue Code.