I. Deconstructing the Black Box
Modern AI is not just code; it is a complex supply chain of data, compute, and human labor. As established in the seminal paper Model Cards for Model Reporting (Mitchell et al., 2019), transparency is a technical requirement, not just an ethical one.
We define the AI Supply Chain as the complete lifecycle of a model—from raw web-scrape data to the final inference API. Understanding this chain is critical for ensuring Algorithmic Accountability and identifying where bias or vulnerabilities enter the system.
Traditional software security focuses on the **Code**. AI security must also focus on the **Data** and the **Weights**. A vulnerability in the training data can manifest as a backdoor in the final model that remains invisible to standard static analysis.
Data Provenance
Tracking the origin and modifications of training datasets to prevent injection.
Primary Sources & Further Reading
- Mitchell et al. (2019). Model Cards for Model Reporting.
- Pushkarna et al. (2022). Data Cards: Purposeful Documentation for Data Sets.
- Gebru et al. (2021). Datasheets for Datasets.
- Bender et al. (2021). On the Dangers of Stochastic Parrots (Environmental & Ethics section).
- AI Now Institute. AI Supply Chain Research and Policy Reports.
- NIST (2024). AI Risk Management Framework (RMF 1.0).