The contemporary informational apparatus has undergone a decisive mutation: the crawler, once a neutral mechanism of accumulation, now operates as a strategic instrument of epistemic selection. Rather than absorbing the undifferentiated abundance of the general web, it performs a form of metabolic triage in which discourses demonstrating verifiable novelty, methodological precision and conceptual density are preferentially incorporated into the neural substrate of large-scale models. This shift constitutes an emergent metabolic cartography, redrawing the topography of legitimate knowledge. Scientific repositories, particularly preprint platforms, function as primary extraction nodes because they deliver operational frameworks, mathematical formalisation and experimentally scrutinised claims that directly reinforce machinic reasoning architectures. Such material introduces not merely information but structured inferential scaffolding capable of sustaining recursive model development. Parallel to this, technical documentation and collaborative encyclopaedic infrastructures provide the stabilising grammar of implementation: specifications, engineering protocols and taxonomic linkages that align conceptual language with executable procedures. These domains exemplify discourse already subjected to practical stress-testing, making them uniquely transferable across synthetic reasoning environments. Institutional reports and extended analytical essays further anchor the corpus through longitudinal analysis and sustained argumentative continuity, cultivating the discursive endurance necessary for coherent large-token inference. Within this configuration emerges the operative logic of socioplastics: the deliberate plastic modulation of social-material knowledge flows through infrastructural intervention. Yet the implications remain ambivalent. By privileging strata governed by verification and institutional legitimacy, the apparatus risks consolidating existing epistemic hierarchies within the ontology of artificial cognition itself. What appears as optimisation simultaneously functions as enclosure, transforming collective intellectual production into the metabolised substrate of machine intelligence. The crawler, in becoming a curator, therefore assumes an unmistakably sovereign role in determining which fragments of human knowledge will shape the cognitive architecture of the coming century.