When Failed Companies Sell Employee Communications as AI Training Data: The Organizational Trust Collapse Nobody Is Modeling
The News Event Worth Taking Seriously
A story broke this week that deserves more analytical attention than it has received. According to reporting from Breaker, failed companies are selling their internal communications - employee messages, emails, and operational data - as AI training datasets to capitalize on the booming demand for training material. Workers whose employers went bankrupt or shut down had no meaningful notice that the private communications they produced during employment would become commodities. This is not a speculative privacy concern. It is a transaction that is already occurring, and the organizational theory literature is poorly equipped to explain why it was structurally predictable.
The Competence Inversion at the Institutional Level
My dissertation research on the Algorithmic Literacy Coordination (ALC) framework focuses on how individual workers develop competencies through platform participation, but this news event forces a parallel question at the institutional level. Classical organizational theory, particularly the principal-agent literature, assumes that parties to an employment relationship understand the full scope of what they are exchanging. A worker produces labor; the firm retains work product and compensates accordingly. What the current data-selling phenomenon reveals is that this schema is structurally incomplete in AI-mediated environments. The firm does not just retain work product - it retains the communicative infrastructure workers used to coordinate, which turns out to have latent market value that neither party recognized at the time of the exchange.
This is an instance of what I would call institutional competence inversion. The organization that failed to survive as a going concern nonetheless accumulated a form of algorithmic capital - training-relevant data - that outlasts the firm itself. The workers who generated that data had no schema for understanding that their casual Slack messages or internal project discussions were accumulating value in a market that barely existed when they wrote them. Kellogg, Valentine, and Christin (2020) document how algorithmic systems at work create new forms of visibility and control, but the scenario here is more extreme: the control relationship persists even after the employment relationship legally terminates.
The Awareness-Capability Gap Has a Structural Cousin
A core tension in ALC theory is the awareness-capability gap: workers can become aware that algorithms govern their outcomes without gaining any capacity to respond effectively to that governance (Gagrain, Naab, and Grub, 2024). The employee communications story illustrates an analogous gap at the organizational boundary level. Employees in 2019 or 2020 may have been broadly aware that data has value, that AI companies need training material, and that corporate data practices are often opaque. That awareness translated into precisely zero protective behavior, because workers had no accurate schema for how their specific communicative outputs mapped onto a downstream market transaction that would occur years later, facilitated by a bankruptcy proceeding.
Hancock, Naaman, and Levy (2020) distinguish between AI-mediated communication, where AI shapes the communicative act itself, and the broader infrastructure within which communication occurs. The selling of employee messages as training data collapses this distinction in a troubling way. The communication was not AI-mediated when produced, but it retroactively becomes raw material for AI systems. Workers operating with accurate schemas about their communication environment at the time of writing were still operating with fundamentally incomplete models of where that communication would travel.
Why Standard Organizational Remedies Will Not Work Here
The instinctive policy response is to call for better employment contracts, clearer data ownership clauses, or stronger privacy regulation. These are not wrong, but they underestimate the structural problem. Rahman (2021) describes the "invisible cage" dynamic in platform labor, where governance operates through algorithmic mechanisms that workers cannot see or contest. The bankrupt-firm-as-data-seller scenario extends this logic beyond the platform context into conventional employment. The cage does not require an active algorithmic employer. It requires only that the data infrastructure of employment persists after the employment relationship ends.
Routine expertise in navigating employment contracts - knowing which clauses to read, which rights to assert - is inadequate here for the same reasons that platform-specific procedural training fails in novel algorithmic contexts (Hatano and Inagaki, 1986). The structural feature that matters is not any particular contract clause but the general principle that digital communicative output is decoupled from the relational context that produced it. Workers who internalize that structural principle - rather than memorizing specific protective tactics - are better positioned to recognize novel instances of the same dynamic.
The Organizational Theory Research Gap
What this news event exposes is a gap in organizational theory's treatment of post-termination data obligations. The literature on organizational trust, psychological contracts (Rousseau, 1989, is the standard reference), and data governance tends to treat these as separate domains. The employee communications story demands that they be treated as unified. When a firm's assets are liquidated, the communicative residue of its workforce should be a theoretically significant category, not an afterthought. The fact that it currently is an afterthought is itself a finding worth investigating.
Roger Hunt