Loading…
Attending this event?
October 28-29, 2024 | Tokyo, Japan
View More Details & Registration
Note: The schedule is subject to change.

The Sched app allows you to build your schedule but is not a substitute for your event registration. You must be registered for Open Source Summit + AI_dev Japan 2024 to participate in the sessions. If you have not registered but would like to join us, please go to the event registration page to purchase a registration.

This schedule is automatically displayed in Japan Standard Time (UTC +9). To see the schedule in your preferred timezone, please select from the drop-down located at the bottom of the menu to the right.
AI_dev clear filter
Monday, October 28
 

11:15 JST

Democratizing Diffusion Models with Diffusers - Sayak Paul, Hugging Face
Monday October 28, 2024 11:15 - 11:55 JST
The talk “Democratizing Diffusion Models with Diffusers” will explore the diverse applications of the open-source Python library Diffusers in the image and video generation space. The talk will showcase how Diffusers, based on diffusion models, enables fast and high-quality image and video generation, making it accessible to a wide range of users. The presentation will cover various use cases, including image inpainting, image editing, and scene composition, demonstrating the capabilities of Diffusers in enabling users to create and edit photo-realistic images with minimum effort. The audience will gain insights into the potential of Diffusers in revolutionizing the way images and videos are generated and edited, making it a must-attend session for anyone interested in the latest advancements in this field.
Speakers
avatar for Sayak Paul

Sayak Paul

Research Engineer, Hugging Face
Sayak works on diffuson models at Hugging Face, focusing on training them, maintaining the diffusers library, and leading some applied research efforts. Off the work, he likes to binge-watch ICML tutorials and Suits.
Monday October 28, 2024 11:15 - 11:55 JST
Hall B (4)

12:05 JST

Data Prep Kit: A Comprehensive Cloud-Native Toolkit for Scalable Data Preparation in GenAI App - Daiki Tsuzuku & Takuya Goto, IBM
Monday October 28, 2024 12:05 - 12:45 JST
Every conversation on AI starts with models and ends with data. Data preparation is emerging as a very important phase of the GenAI journey, as high quantity and quality text and code corpora for GenAI model training have shown to play a crucial role in producing high performing Large Language Models (LLMs). The data preparation phase in the Generative AI lifecycle aims to clean, filter, and transform the datasets of text and code that are acquired from various sources into a tokenized form that is suitable for the training of LLMs, be it pre-training, or constructing LLM apps via fine-tuning or instruct tuning. The latter poses unique challenges, as each use case may necessitate tailored data preparation approaches. Given the enduring and evolving demand for data preparation techniques in LLM applications, we are introducing Data Prep Kit as an open-source software asset. This endeavour is geared towards fostering collaborative efforts within the community, enabling collective development and utilization, and ultimately reducing time to value. DPK has been instrumental in powering the IBM open-source Granite models.
Speakers
avatar for Takuya Goto

Takuya Goto

Software Engineer, IBM
Takuya is a software engineer at IBM where he works on software product development, and open-source development. Takuya specializes in NLP, ML, and text-based data processing. In his free time, Takuya likes running, and traveling with my wife and son.
avatar for Daiki Tsuzuku

Daiki Tsuzuku

Software Developer, IBM
I have been working in IBM as a software developer for about 7 years. I have been the backend developer, and sometimes frontend developer, of Watson Explorer, Watson Discovery, and watsonx Orchestrate. My field is to develop the application of processing a wide variety and large volume... Read More →
Monday October 28, 2024 12:05 - 12:45 JST
Hall B (4)

14:00 JST

Optimize Your AI Cloud Infrastructure: A Hardware Perspective - Liang Yan, CoreWeave
Monday October 28, 2024 14:00 - 14:40 JST
GPU Cloud has become a ubiquitous component of contemporary AI infrastructure, especially for distributed machine learning scenarios. While conversations around AI infrastructure optimization typically revolve around the application layer, such as machine learning tasks and distributed job schedulers, delving into the underhood of the GPU cloud is essential. Numerous factors, including POD Scheduler, Device Plugin, GPU/NUMA topology, ROCE/NCCL Stack, and more, can significantly impact performance.

This session will thoroughly explore the tuning of various machine models(CNN/RNN/Transformer) from MLPerf using an H100 Cluster as a reference. We will analyze the correlation between model performance and device operator configuration in nodes by presenting first-hand experimental results to unveil the hidden potential within a K8S GPU Cloud.
Speakers
avatar for Liang Yan

Liang Yan

Sr. Software Engineer, CoreWeave
Liang Yan is a senior software engineer at Coreweave, specializing in AI Infra, heterogeneous architecture acceleration and distributed machine learning systems from the cloud base. He collaborates closely with upstream communities and leading vendors like NVIDIA, AMD and ARM, delivering... Read More →
Monday October 28, 2024 14:00 - 14:40 JST
Hall B (4)
  AI_dev

14:50 JST

A Next-generation IoT Platform for Edge AI Apps Leveraging Sensors and Wasm - Munehiro Shimomura, Sony Semiconductor Solutions Corporation & Kenji Shimizu, Midokura
Monday October 28, 2024 14:50 - 15:30 JST
In this session, we will introduce the construction of a comprehensive platform that uses Edge AI and sensors to cover everything from devices to the cloud. The platform enables advanced cooperation between sensors and AI control, and emphasizes seamless and dynamic replacement of AI models by using WebAssembly (Wasm). Furthermore, through open sourcing, we aim to expand the ecosystem and form a technical community. Through technical details and real-world scenarios, we will provide insights that participants can apply to their own projects.
Speakers
avatar for Kenji Shimizu

Kenji Shimizu

Manager, Midokura
After spending more than 20 years in Japanese telecom & mobile company as an R&D engineering researcher and manager, he, inspired by Midokura's vision, has joined and started to play a role to expand the ecosystem for open source for an edge distributed AI sensing infrastructure which... Read More →
avatar for Munehiro Shimomura

Munehiro Shimomura

Open Source Program Manager, Sony Semiconductor Solutions Corporation
Munehiro is the division's OSPO Open Source Program Manager, where he leads open source strategy development and execution. He believes it is important to create a culture in which organizations can strategically and proactively utilize open source, and is working hard to achieve... Read More →
Monday October 28, 2024 14:50 - 15:30 JST
Meeting Room 1

14:50 JST

Unlocking Local LLMs with Quantization - Marc Sun, Hugging Face
Monday October 28, 2024 14:50 - 15:30 JST
This talk will share the story of quantization, its rise in popularity, and its current status in the open-source community. We'll begin by reviewing key quantization papers, such as QLoRA by Tim Dettmers and GPTQ by Elias Frantar. Next, we'll demonstrate how quantization can be applied at various stages of model development, including pre-training, fine-tuning, and inference. Specifically, we'll share our experience in pre-training a 1.58-bit model, show how fine-tuning is achievable using PEFT + QLoRA, and discuss optimizing inference performance with torch.compile or custom kernels. Finally, we'll highlight efforts within the community to make quantized models more accessible, including how transformers incorporate state-of-the-art quantization schemes and how to run GGUF models from llama.cpp.
Speakers
avatar for Marc Sun

Marc Sun

Machine Learning Engineer, HuggingFace
Marc is a ML Engineer working on the Open Source team at Hugging Face and he collaborates with researchers and developers to add new exciting features in the HF ecosystem and have contributed to various libraries in the HF ecosystem (transformers, accelerate, PEFT). Marc is also deeply... Read More →
Monday October 28, 2024 14:50 - 15:30 JST
Hall B (4)

15:40 JST

From Design to Launch: Implementing AI Governance at Scale - Nathália Kuromiya & Martin Winkler, Google
Monday October 28, 2024 15:40 - 16:20 JST
What does it take to design, implement, and roll out a comprehensive AI governance program? What challenges and opportunities do companies encounter when scaling AI governance across diverse products and AI applications, and how can AI governance programs be designed to stay agile in a dynamic technological and regulatory environment? Insights on scaling AI governance programs progressively across multiple jurisdictions while keeping them agile in a dynamic technological and regulatory landscape. Practical learnings, effective solutions and challenges to watch out for. Topics discussed: - What are the building blocks of your company’s AI Governance program? - What challenges did we face when building an AI Governance program? - The AI Governance landscape is evolving rapidly – both through technical innovation and regulatory advances. How to keep an agile approach to AI governance? - What kinds of risks are front and center when you’re building an AI Governance program? What kinds of opportunities?
Speakers
avatar for Martin Winkler

Martin Winkler

Software Engineer, Google
Martin Winkler is a Software Engineer at Google as part of the Privacy, Safety and Security team. He works as a lead for privacy and governance tooling and tackled cross-company privacy challenges to ensure the safety of users and their data. Additionally, he is establishing company... Read More →
avatar for Nathália Kuromiya

Nathália Kuromiya

Software Engineer, Google
Nathália Harumi Kuromiya is a Software Engineer at Google as part of the Privacy, Safety and Security team. She works as a lead for privacy and governance tooling and had a privacy reviewer role. For the past year, she's been working on AI safety space as one of the technical contributors... Read More →
Monday October 28, 2024 15:40 - 16:20 JST
Hall B (4)

16:40 JST

Leveraging Zephyr and ML to Bring Smart Devices to Market, Faster - Kate Stewart, The Linux Foundation
Monday October 28, 2024 16:40 - 17:20 JST
End point devices are resource constrained, either in terms of power, memory or communication capabilities - sometimes all three. However, being able to apply machine learning on these end point devices is possible and when applied strategically enables system wide efficiencies to be realized. This talk will explore the requirements and tradeoffs for such system to be considered when using the Zephyr RTOS and Tensorflow Lite for Embedded Microcontrollers projects.
Speakers
avatar for Kate Stewart

Kate Stewart

VP Dependable Embedded Systems, Linux Foundation
Kate Stewart works with the safety, security and license compliance communities to advance the adoption of best practices into embedded open source projects. Since joining The Linux Foundation, she has launched the ELISA and Zephyr Projects, as well as supporting other embedded projects... Read More →
Monday October 28, 2024 16:40 - 17:20 JST
Hall B (4)

17:30 JST

GPU Distributed Caching for PyTorch Leveraging NVMe, GDS, and RDMA - Hope Wang, Alluxio
Monday October 28, 2024 17:30 - 18:10 JST
As GPUs become increasingly powerful, the separation between compute and storage often results in underutilized GPUs waiting for data. Meanwhile, high-performance components on GPU machines, such as NVMe storage and fast networks leveraging InfiniBand or special NICs, remain idle. Effectively leveraging these hardware resources to address GPU underutilization is a critical challenge. In this talk, we introduce a Kubernetes-native distributed caching layer that leverages NVMe disks and fast networks to optimize PyTorch training data access. Utilizing stateless workers for scalability and ETCD for membership services, this caching layer efficiently manages and serves data. Cached data is rapidly and efficiently fed into GPU memory using NVIDIA's DALI data loader, GPUDirect Storage (GDS), and Remote Direct Memory Access (RDMA), significantly reducing data transfer bottlenecks and improving overall training performance.
Speakers
avatar for Hope Wang

Hope Wang

Developer Advocate, Alluxio
Hope Wang is a Presto Contributor and a Developer Advocate at Alluxio. She has a decade of experience in Data, AI, and Cloud. An open-source contributor to PrestoDB, Trino, and Alluxio, she currently works at Alluxio as a developer advocate and previously worked in venture capital... Read More →
Monday October 28, 2024 17:30 - 18:10 JST
Hall B (4)
 
Tuesday, October 29
 

11:10 JST

Monsters in the Deps: How to Protect Your AI/ML Systems from Supply Chain Attacks - Erin Glass & Patrick Smyth, Chainguard
Tuesday October 29, 2024 11:10 - 11:50 JST
Love developing AI/ML systems, but don’t want to become the next front-page cyberattack? We got you! In this fast-paced, meme-a-liscious, hands-on workshop, we’ll take a deep dive into the murky waters of the AI/ML supply chain, explore its many threats and terrors, and then – with our trusty box of supply chain security tools – build an island of safety for our AI/ML systems! Participants will come away with the skills and knowledge to significantly improve AI/ML supply chain security at their organization, as well as the unpleasant awareness about what might happen if the industry doesn’t do the same. Sorry! Hands-on activities will include vulnerability scanning, creating/consuming SBOMs/AIBOMs, digital signing using Sigstore tools, and provenance tracking. We will also provide a conceptual background on AI/ML supply chain components, attack categories, and global regulation and standards related to AI/ML security. Led by software supply chain and AI deployment experts at Chainguard, this workshop will enable participants to harden their AI/ML systems and evangelize others to do the same.
Speakers
avatar for Erin Glass

Erin Glass

Senior Product Manager, Chainguard
Dr. Erin Glass is a product manager at Chainguard, where she focuses on supply chain security education and meme R&D. She has published widely in developer education and other digital topics, including the courses Securing the AI/ML Supply Chain and Painless Vulnerability Management... Read More →
avatar for Patrick Smyth

Patrick Smyth

Staff Developer Relations Engineer, Chainguard
Dr. Patrick Smyth is Staff Developer Relations Engineer at Chainguard, where he shows developers how to deploy AI and other applications with 0 CVEs using Chainguard Images. Patrick has a PhD in the digital humanities and in a previous life led technical bootcamps for researchers... Read More →
Tuesday October 29, 2024 11:10 - 11:50 JST
Hall B (4)
  AI_dev

12:00 JST

Exploring Pillars of Trustworthy AI: Robustness and Fairness - Niharika Shrivastava, Workforce Optimizer
Tuesday October 29, 2024 12:00 - 12:40 JST
Due to the ever-increasing adoption of AI into the lives of daily users, trustworthy AI is of utmost priority. Even though advocates of AI globally have started talking about ethical considerations during ML model building, in reality, very few people know how to create robust, privacy-preserving, and fair AI models. In this talk, I'll explore 2 concrete technical concepts of trustworthy AI, namely ensuring robustness and fairness in ML models. Robustness: 1. Attendees will go through an in-depth understanding of critical vulnerabilities of common AI models and how to exploit them to adversarially attack the model (e.g., inference attacks, data poisoning). 2. This will be followed by simple defence strategies to increase robustness (e.g., gradient obfuscation, transformations). 3. This will be further followed by adaptive attacks on previous defence strategies thereby motivating the concept of certified robustness of AI models. Fairness: 1. Attendees will get to know how they can unconsciously encode bias (representational bias, model bias, etc) during training AI models. 2. This is followed by strategies to correct this bias using domain knowledge to create fair AI models.
Speakers
avatar for Niharika Shrivastava

Niharika Shrivastava

Data Scientist, Workforce Optimizer
Niharika's current interests lie in NLP and Applied Data Science. She holds a Master's in AI from the National University of Singapore. She was also an Outreachy fellow for The Fedora Project and has been the recipient of multiple awards such as the Red Hat Women in Open Source Award... Read More →
Tuesday October 29, 2024 12:00 - 12:40 JST
Hall B (4)

14:00 JST

Building a Thriving Open-Source AI Community with LF AI & Data Foundation - Vini Jaiswal, TikTok
Tuesday October 29, 2024 14:00 - 14:40 JST
Ever feel like contributing to groundbreaking AI projects but unsure where to start? Have a cool project that you are looking to find a neutral entity for? Through the session learn how the LF AI & Data foundation empowers open-source projects. Discover the framework to support for open development, governance models, and resources like legal assistance, marketing, and events. Gain exclusive updates from the LF AI and Data Technical Advisory Council (TAC). Learn about their vision, technical roadmap, success stories and how you can contribute. The session will cover practical tips and tricks to dive into project contributions, navigate the initiation process, and ultimately guide you towards project graduation. Also discover the diverse work streams within LF AI and Data foundation and how you can leverage your skills to make a real difference in the open-source AI community.
Speakers
avatar for Vini Jaiswal

Vini Jaiswal

Chair of Technical Advisory Council, Linux Foundation AI & Data
Vini Jaiswal is a renowned expert in AI and Data, acclaimed for her significant contributions to Apache Spark, MLflow, PrivacyGo and, notably, Delta Lake. Holding pivotal roles such as Chair of the Technical Advisory Committee (TAC) at Linux Foundation Data and AI, Governing Board... Read More →
Tuesday October 29, 2024 14:00 - 14:40 JST
Hall B (4)
  AI_dev

14:50 JST

Bringing AI on-Device: From Cloud to Edge - Catalin Vasile, Adobe
Tuesday October 29, 2024 14:50 - 15:30 JST
Step into the future of AI as we break free from the constraints of cloud-based processing and unlock the immense potential of edge computing. This cutting-edge talk explores the revolutionary shift of AI from centralized data centers to the devices in your pocket, on your wrist, and all around you. Discover how this paradigm shift is not just changing the game – it's rewriting the rules. We'll journey through the landscape of on-device AI, revealing how it's transforming user experiences, supercharging privacy, and pushing the boundaries of what's possible in real-time applications. From smart homes to autonomous vehicles, from augmented reality to personalized healthcare – learn how on-device AI is the key to unlocking a world of intelligent, responsive, and secure applications.
Speakers
avatar for Catalin Vasile

Catalin Vasile

Senior Computer Scientist, Adobe
Catalin is a Senior Computer Scientist at Adobe, surfing the clouds in the world of distributed systems and managing resilient high-scale solutions as part of the Cloud Platform team.
Tuesday October 29, 2024 14:50 - 15:30 JST
Hall B (4)

15:50 JST

Data Contracts Are Good for AI - Jean-Georges Perrin, Bitol / AbeaData
Tuesday October 29, 2024 15:50 - 16:30 JST
Bitol is one of the newest LF AI & Data projects focusing on open standards and open source tools for modern data engineering. Bitol's flagship standard is ODCS (Open Data Contract Standard). In this talk, as the chair of the Bitol TSC, I will first explain what a data contract is, the work of the Bitol project, and focus on the benefits of data contracts for AI & analytics. I will conclude with an end-to-end demo of creating and enforcing data contracts using open-source and free tools.
Speakers
avatar for Jean-Georges Perrin

Jean-Georges Perrin

Chair of the TSC / Chied Innovation Officer, Bitol / AbeaData
Jean-Georges “jgp” Perrin is CIO at AbeaData, focusing on building innovative and modern data tooling. He is also chair of the Bitol project at the Linux Foundation, a Lifetime IBM Champion, and author of multiple books, including Implementing Data Mesh (O’Reilly) and Spark... Read More →
Tuesday October 29, 2024 15:50 - 16:30 JST
Hall B (4)

16:40 JST

From Complexity to Clarity: Addressing Challenges in AI BOMs for Compliance - Gopi Krishnan Rajbahadur, Huawei Technologies Canada & Kate Stewart, The Linux Foundation
Tuesday October 29, 2024 16:40 - 17:20 JST
As global regulations on AI software tighten, developers face a complex set of new, ambiguous rules. The AI Software Bill of Materials (AI BOM), especially the new SPDX 3.0 with AI and dataset profiles, offers a promising solution for compliance, providing detailed, machine-readable documentation of AI systems. Despite its benefits, adoption has been slow, hindered by gaps in developer knowledge and the complex nature of AI systems. Many AI BOMs are incomplete or inaccurate, limiting their utility for compliance. Our talk will tackle these issues, drawing on our experience with SPDX 3.0 and AI BOM implementation. We'll share best practices and strategies to improve AI BOM accuracy and utility, equipping professionals with the insights to ensure their AI applications are compliant and prepared for future regulations.
Speakers
avatar for Gopi Krishnan Rajbahadur

Gopi Krishnan Rajbahadur

Senior Staff Researcher, Huawei Technologies Canada
Gopi Krishnan Rajbahadur is a Senior Staff Researcher at Huawei's Centre for Software Excellence in Canada. He is currently working on SE for Large Language Models and the governance of AI datasets. He is the co-lead for the AI and datasets profile in the ISO/IEC 5692 SPDX standard... Read More →
avatar for Kate Stewart

Kate Stewart

VP Dependable Embedded Systems, Linux Foundation
Kate Stewart works with the safety, security and license compliance communities to advance the adoption of best practices into embedded open source projects. Since joining The Linux Foundation, she has launched the ELISA and Zephyr Projects, as well as supporting other embedded projects... Read More →
Tuesday October 29, 2024 16:40 - 17:20 JST
Hall B (4)
 
  • Filter By Date
  • Filter By Venue
  • Filter By Type
  • Audience
  • Timezone

Share Modal

Share this link via

Or copy link

Filter sessions
Apply filters to sessions.