Skip to content

Commit 0997722

Browse files
authored
Updates (#310)
1 parent 8fa81c9 commit 0997722

File tree

2 files changed

+27
-2
lines changed

2 files changed

+27
-2
lines changed

source/_data/SymbioticLab.bib

Lines changed: 19 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -2009,7 +2009,7 @@ @Article{curie:arxiv25
20092009
20102010
@Article{cornstarch:arxiv25,
20112011
author = {Insu Jang and Runyu Lu and Nikhil Bansal and Ang Chen and Mosharaf Chowdhury},
2012-
title = {Cornstarch: Distributed Multimodal Training Must Be Multimodality-Aware },
2012+
title = {Cornstarch: Distributed Multimodal Training Must Be Multimodality-Aware},
20132013
year = {2025},
20142014
month = {March},
20152015
volume = {abs/2503.11367},
@@ -2024,4 +2024,21 @@ @Article{cornstarch:arxiv25
20242024
Multimodal large language models (MLLMs) extend the capabilities of large language models (LLMs) by combining heterogeneous model architectures to handle diverse modalities like images and audio. However, this inherent heterogeneity in MLLM model structure and data types makes makeshift extensions to existing LLM training frameworks unsuitable for efficient MLLM training.
20252025
In this paper, we present Cornstarch, the first general-purpose distributed MLLM training framework. Cornstarch facilitates modular MLLM construction, enables composable parallelization of constituent models, and introduces MLLM-specific optimizations to pipeline and context parallelism for efficient distributed MLLM training. Our evaluation shows that Cornstarch outperforms state-of-the-art solutions by up to 1.57x in terms of training throughput.
20262026
}
2027-
}
2027+
}
2028+
2029+
@Article{ai-eval-framework:arxiv25,
2030+
author = {Sarah Jabbour and Trenton Chang and Anindya Das Antar and Joseph Peper and Insu Jang and Jiachen Liu and Jae-Won Chung and Shiqi He and Michael Wellman and Bryan Goodman and Elizabeth Bondi-Kelly and Kevin Samy and Rada Mihalcea and Mosharaf Chowhury and David Jurgens and Lu Wang},
2031+
title = {Evaluation Framework for {AI} Systems in the Wild},
2032+
year = {2025},
2033+
month = {April},
2034+
volume = {abs/2504.16778},
2035+
archivePrefix = {arXiv},
2036+
eprint = {2504.16778},
2037+
url = {https://arxiv.org/abs/2504.16778},
2038+
publist_confkey = {arXiv:2504.16778},
2039+
publist_link = {paper || https://arxiv.org/abs/2504.16778},
2040+
publist_topic = {Systems + AI},
2041+
publist_abstract = {
2042+
Generative AI (GenAI) models have become vital across industries, yet current evaluation methods have not adapted to their widespread use. Traditional evaluations often rely on benchmarks and fixed datasets, frequently failing to reflect real-world performance, which creates a gap between lab-tested outcomes and practical applications. This white paper proposes a comprehensive framework for how we should evaluate real-world GenAI systems, emphasizing diverse, evolving inputs and holistic, dynamic, and ongoing assessment approaches. The paper offers guidance for practitioners on how to design evaluation methods that accurately reflect real-time capabilities, and provides policymakers with recommendations for crafting GenAI policies focused on societal impacts, rather than fixed performance numbers or parameter sizes. We advocate for holistic frameworks that integrate performance, fairness, and ethics and the use of continuous, outcome-oriented methods that combine human and automated assessments while also being transparent to foster trust among stakeholders. Implementing these strategies ensures GenAI models are not only technically proficient but also ethically responsible and impactful.
2043+
}
2044+
}
Lines changed: 8 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,8 @@
1+
---
2+
title: >-
3+
Mosharaf Delivered a Public Lecture and Q&A on Energy-Optimal AI at the Michigan Theater
4+
categories:
5+
- News
6+
date: 2025-04-22 23:05:10
7+
tags:
8+
---

0 commit comments

Comments
 (0)