Navigating Uncertainties: Understanding How GenAI Developers Document Their Models on Open-Source Platforms

📅 2025-03-30
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study identifies systemic documentation uncertainties faced by generative AI developers on open-source platforms (e.g., Hugging Face), including ambiguity in content selection, inconsistent presentation of critical model components, and unclear accountability attribution. Through semi-structured interviews with 13 developers and grounded-theory-driven thematic coding, we empirically derive and validate a novel “Three-Dimensional Model of Documentation Uncertainty”—the first such framework in the GenAI documentation literature. Our contribution includes a collaborative governance framework comprising three actionable recommendations: (1) cultivating community-endorsed documentation standards; (2) establishing shared model evaluation infrastructure; and (3) formalizing multi-stakeholder responsibility delineation mechanisms. The framework has been adopted by the Hugging Face Documentation Working Group. It advances both theoretical understanding and practical implementation of responsible, standardized documentation for generative AI models. (149 words)

Technology Category

Application Category

📝 Abstract
Model documentation plays a crucial role in promoting transparency and responsible development of AI systems. With the rise of Generative AI (GenAI), open-source platforms have increasingly become hubs for hosting and distributing these models, prompting platforms like Hugging Face to develop dedicated model documentation guidelines that align with responsible AI principles. Despite these growing efforts, there remains a lack of understanding of how developers document their GenAI models on open-source platforms. Through interviews with 13 GenAI developers active on open-source platforms, we provide empirical insights into their documentation practices and challenges. Our analysis reveals that despite existing resources, developers of GenAI models still face multiple layers of uncertainties in their model documentation: (1) uncertainties about what specific content should be included; (2) uncertainties about how to effectively report key components of their models; and (3) uncertainties in deciding who should take responsibilities for various aspects of model documentation. Based on our findings, we discuss the implications for policymakers, open-source platforms, and the research community to support meaningful, effective and actionable model documentation in the GenAI era, including cultivating better community norms, building robust evaluation infrastructures, and clarifying roles and responsibilities.
Problem

Research questions and friction points this paper is trying to address.

Understanding GenAI developers' model documentation practices on open-source platforms
Identifying uncertainties in what, how, and who documents GenAI models
Addressing gaps in effective and responsible GenAI model documentation
Innovation

Methods, ideas, or system contributions that make the work stand out.

Interviews with GenAI developers on documentation practices
Identifies uncertainties in model documentation content
Proposes norms and infrastructures for better documentation
🔎 Similar Papers
No similar papers found.