One Documentation Does Not Fit All: Case Study of TensorFlow Documentation

📅 2025-05-03
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
TensorFlow documentation exhibits low usability for machine learning (ML) practitioners without software engineering backgrounds. Method: We propose the first dual-dimensional “Nature & Trigger” taxonomy for ML software documentation—categorizing questions by *nature* (e.g., conceptual, debugging, usage) and *trigger* (e.g., example insufficiency, error occurrence)—and conduct an empirical study integrating qualitative content analysis, Stack Overflow data mining, and thematic coding. Results: 64.3% of documentation-related questions stem from inadequate or non-transferable code examples; 24.9% concern error diagnosis and debugging. Customized tutorials fail to improve effectiveness, exposing a fundamental tension between *generality* and *user adaptability*. The study systematically identifies structural deficiencies in current ML documentation—particularly in error support and generalizability—and establishes a transferable documentation assessment framework. This work provides both theoretical foundations and actionable design guidelines for improving ML tool documentation.

Technology Category

Application Category

📝 Abstract
Software documentation guides the proper use of tools or services. With the rapid growth of machine learning libraries, individuals from various fields are incorporating machine learning into their workflows through programming. However, many of these users lack software engineering experience, affecting the usability of the documentation. Traditionally, software developers have created documentation primarily for their peers, making it challenging for others to interpret and effectively use these resources. Moreover, no study has specifically focused on machine learning software documentation or analyzing the backgrounds of developers who rely on such documentation, highlighting a critical gap in understanding how to make these resources more accessible. This study examined customization trends in TensorFlow tutorials and compared these artifacts to analyze content and design differences. We also analyzed Stack Overflow questions related to TensorFlow documentation to understand the types of questions and the backgrounds of the developers asking them. Further, we developed two taxonomies based on the nature and triggers of the questions for machine learning software. Our findings showed no significant differences in the content or the nature of the questions across different tutorials. Our results show that 24.9% of the questions concern errors and exceptions, while 64.3% relate to inadequate and non-generalizable examples in the documentation. Despite efforts to create customized documentation, our analysis indicates that current TensorFlow documentation does not effectively support its target users.
Problem

Research questions and friction points this paper is trying to address.

Examining usability gaps in TensorFlow documentation for diverse users
Analyzing developer backgrounds and question trends on Stack Overflow
Assessing effectiveness of customized tutorials for machine learning tools
Innovation

Methods, ideas, or system contributions that make the work stand out.

Analyzed TensorFlow documentation customization trends
Developed taxonomies for machine learning software questions
Evaluated Stack Overflow questions on TensorFlow documentation
🔎 Similar Papers
No similar papers found.
S
S. Thirimanne
York University
E
Elim Yoseph Lemango
York University
G
Giulio Antoniol
Ecole Polytechnique de Montreal
Maleknaz Nayebi
Maleknaz Nayebi
Associate Professor at York University
Requirements EngineeringEmpirical software engineeringSoftware Analytics