Author Archives: Brian Millen

Modeling Value in the Anthropocene – White Paper

Project Narrative

Modeling Value in the Anthropocene is the prototype for a computational method that can assist in philosophical research and the theoretical work of reading and writing more generally. As we see it, it is both 1) a contribution to the theory and practice of the work of Bernard Stiegler, the Internation Collective, and the Association of Friends of the Thunberg Generation via the introduction of machine learning in general and word embedding in particular to reading and writing philosophy; and 2) a contribution to the fields of text analysis and distant reading, and the digital humanities more broadly, with an application of the philosophy of Bernard Stiegler that can situate natural language processing as an element of a technique of reading and writing in the digital age, while also scoping out the limits of such a technique.

The project began with a question as to the potential usefulness of distant reading for theoretical work. By theoretical, we mean all of the scientific disciplines through which bodies of researchers generate new knowledges of all kinds. Distant reading is the application of mathematical models to large databases of text after the text is made quantifiable, and thus calculable, by datafying it, and performs computations with it. These are analytical functions that the human mind does not have the capacity to perform. Most of the text analysis work done in the humanities (such as that of Franco Morretti, Ted Underwood, and Michael Gavin) takes up the literary field to study things like changes in language and style across different time periods or different literary movements. There are many arguments for and against this kind of work. Having a particular interest in the discipline of philosophy, we wanted to address the question of whether this kind of work could be used in the activity that philosophers partake in.

One way to approach the question of distant reading philosophy is through the thought of Michael Witmore, especially in his blog post, “Text: A Massively Addressable Object.” Here he defends the practice by positioning it as simultaneously continuous with the practice of reading since its genesis and discontinuous with past modalities of reading.
For Witmore, what separates digital text from older forms of text is that “it is massively addressable at different levels of scale” (Witmore). To understand this requires an understanding of what he means by “scale.” According to Witmore, reading has always been the practice of projecting a material text into an ideal level that one reads it at. For example, one can choose to address a text at the level of the word, the sentence, the paragraph, or the book, just to name a few. These levels are not material characteristics of the text itself, but rather subjective ways of dividing the text up and projecting them into/out of the text. A reader could just as easily choose to underline all the words of a text beginning with the letter ‘B’ and address it at that scale. How text is addressed is a matter of convention, a mode of attention contingent on the technical and normative limits of the given socio-historical context in which the reading occurs. Reading text as “books” or even “words” is a socially constructed mode of reading. As Witmore notes, “The idea of a word is itself an artifact of manuscript culture, one that could be perpetuated in print through the affordances of moveable type” (Witmore). This implies that there are other possible scales of reading one could address.
What makes digital text different is this scale of addressability, which in fact contains many different scales contingent on the new capacities of the technology. Instead of having to read one word at a time or one sentence at a time, we can query and compare many words from many different documents at once. Perhaps the most popular form of this found in the digital humanities is topic modeling. Topic models are generated by running an unsupervised machine learning algorithm on a group of documents and approximating which words tend to be used together in the same document. This allows us to address the texts at the level of “topic,” creating ideal topics we can say that the texts appear to be drawing from. This kind of modeling could prove useful for making clear what discourses various texts and authors and schools of thought might be drawing from, as Lisa Rhody has discussed with reference to ekphrastic poetry.

In philosophy, this kind of analysis could address large amounts of text at a scale that would allow us to understand what topics are characteristic of a particular school of thought (like German idealism) or a particular epoch of history (like the Enlightenment). Malaterre et al. run an unsupervised machine learning algorithm on the complete corpus of the journal Philosophy of Science, mining articles from 1934 to 2015. One result was the revelation that the topics which are consistent throughout are those related to “metaphysical and ontological questioning” about “space and time, causation, natural kinds, [and] realism” (Malaterre et al, 215). Another result was the discovery of how the topics of logic and language rose and fell throughout time.
 Despite topic modeling’s potential fruitfulness, though, we wish to show why word embedding is a better method for philosophy specifically. There are two reasons for this. The first is because it allows us to model conceptual similarity among different words. As Gavin et al. argue,

concepts are not words. The distinction can be glimpsed by considering any collection of synonyms, such as rubbish, trash, waste, junk. Each of these words has a distinct use, and they imply different connotations and meanings, but there remains a principle of synonymy among them, an underlying concept that ties them together (Gavin et al.)

With a word embedding model, an algorithm turns each unique word in a text corpus into a vector containing the relationship that each word bears to every other word in the corpus. With the numerical values based in a given word’s distributional distance from every other word in the text, semantic resonance can be calculated between words which have a similar relationship to the matrix of the text. This is useful for conceptual analysis because words that have similar vectors to each other will appear clustered together within the vector space, indicating that they are distributed throughout the texts in a similar way. The linguistic theory operating behind this model is that words that are deployed in similar contexts have some conceptual resonance: “The latent idea here is that different words will tend to appear in different contexts, and therefore one can guess at a word’s meaning by simply counting the words that appear near it” (Gavin et al.). Bringing it back to the language of Witmore, word embedding is a way of addressing large amounts of text, through calculating semantic similarity, at the level of the concept. It is an addressing of the text at the level of the word, but it is a level where each word exists in relation to every other word, the level getting more discreet as we narrow our investigation. Topic modeling could perhaps guide a close reading exploring the topic deeper, or what role a key word might play in a topic, but it cannot get to the semantic depth we might want to in the work of philosophy.

The other reason we prefer word embedding for philosophy is because philosophy is not just the modeling of concepts that already exist. As Deleuze and Guattari write, “The philosopher is the concept’s friend; he is the potentiality of the concept…philosophy is the discipline that involves creating concepts” (Deleuze et al., 5). The operations of word embedding alone already described are useful for clarifying and perhaps deepening concepts, and could possibly lend themselves to some conceptual re-evaluation. However, there is another operation made possible by word embeddings that contains so much more potential for philosophy. Because the words are embedded with numerical values, once the conceptual networks have been modeled in vector space, we can actually perform calculations on the vectors that create new vectors based on this math, and these new vectors can open a path to the creation of concepts, or what Stiegler calls “categorial invention,” which is the goal of philosophy, and perhaps of all theoretical work. A well-cited example is that of V(Queen) = V(King) + V(Woman) – V(Man). When taking the vector for “King”, adding the vector for “Woman” to it, and subtracting the vector for “Man”, the model has been proven to successfully output the vector for “Queen.” What this means conceptually is that if we add the qualities of women to the concept of kings, we have a concept which would have all the qualities of women and all the qualities of kings. If we then subtract from this new vector everything associated with men, we get the concept of queen. This is a simple example, but this functionality can prove exceptionally useful for philosophy.
One of the principal things Stiegler called for is a rethinking of value to escape the Anthropocene, initializing what he calls the epoch of the Neganthropocene. A chief problem of capitalism, he claims, is that, under the conditions it initiates, all use value is reduced to exchange value. The usefulness of a thing is reified into how much it costs, or how much money it could make. This reduces everything to the rules of the market. The progression of this dynamic is the way through which knowledge, art, politics, and life have been devalued, not to mention the health of the biosphere and the future itself. Thus, the Neganthropocene, which would be the epoch following the Anthropocene (if there is to be one), would have to be generated on the basis of a new valuation. The question, then, is if the value of everything is no longer to be based on profit and calculability, what is to be the new value founding this society? We hypothesized we could contribute to the thinking through of this question by treating Stiegler’s works with word embedding. We proposed querying a sample equation that looks something like V(value) – V(Anthropocene) + V(Neganthropocene). This would take the concept of value, subtract that which is characteristic of the Anthropocene from it, and add the vector representing the things that Stiegler writes about the Neganthropocene. This analogic calculation might point us in the direction of which words will be related together as having to do with how we should re-ground value beyond the Anthropocene. We planned to train word2vec, a word embedding algorithm, on a collection of texts by Stiegler and other members of the Internation Collective. The Stiegler works used were “The Neganthropocene”, “Nanjing Lectures 2016-2019”, and “Technics and Time, 4: Faculties and Functions of Noesis in the Post-Truth Age”. We also used “Psychopolitical Anaphylaxis: Steps Towards a Metacosmics” by Daniel Ross and “Bifurcate: ‘There Is No Alternative’”, a collection written by the Internation Collective. Then, we were to query the corpus for this new vector and see what insight could be granted into the question of value in the Neganthropocene.

It should be made very clear that this type of calculation is not a magic wand that can reveal new concepts for us on its own. Witmore’s account of distant reading focuses on the scale of the address, but it does not take into full account the shape or contours of the address itself. We would argue that there are two main modes with which one can address text: analytic and synthetic. These neo-Kantian faculties that Stiegler articulates are two faculties that make up the process of knowledge production. The full explication of these arguments are beyond the scope of this report, but they show that the calculation of data requires the synthetic work of the imagination to think according to standards of reason, and more importantly to dream up new concepts that do not fit into the analytic schema of the understanding. Information or data is the externalization of a prior synthetic act of reason that is calculable now that it is materialized. This act is a decomposition of the line of reasoning into discrete elements that can thus be quantified and calculated. This act is entropic in and of itself, but can produce new knowledge, new concepts, if it leads to a surprise which causes one to think it through according to the criteria of knowledge and create a new idea which re-organizes the analytical understanding as it now stands. In other words, by modeling text, one divides it up into a large number of different pieces (in this case, vectors) that one can perform calculations on. On their own, these models and these calculations are useless. However, an act like querying Stiegler’s texts for the answer to V(profit) – V(Anthropocene) + V(Neganthropocene) could open up a path that one could wander down. And perhaps, by wandering down this path, which would include careful thought, reasoning, and close reading, one could perhaps experience a surprise in the text. This surprise could potentially cause one to rethink the text they are reading closely in a new way, and potentially lead to the production of a concept. There is of course no way to guarantee this, but it is only by seeking out that which is incalculable that philosophy can be done. Perhaps word embedding could provide a kind of calculation that leads the way toward thinking about value anew and how a new society can be built upon this new concept of value. This could then guide a close reading of some of Stiegler’s texts that could potentially concretize this new, currently unknown, concept. This was the kind of work we hoped this project could make possible.

Audience

From the onset of Modeling Value in the Anthropocene, the esoteric nature of Bernard Stiegler’s philosophical thought along with the niche branch of natural language processing that is word embedding presented our project with the unique challenge of positioning our work in a way that equally engages with scholars in the digital humanities and philosophy, while simultaneously communicating our intention, our theoretical foundation, and our results in such a way that it might prove to be accessible to those on the periphery of these two disciplines. Though we predicted that our work would likely resonate most with those working closely with the theoretical and technical approaches employed in Modeling Value in the Anthropocene (and its sister project Modeling Memory in the Anthropocene), it is our hope that through the presentation of our findings now available on the project’s website, the various resources provided allowing users introductory insight into the theoretical framework of the Internation Collective, and the approachable and digestible nature of the NeganthropoZene, that our findings might be appropriately equipped to capture a broader scope of attention including that of students of philosophy, digital humanists throughout the field, and recreational scholars of theory and technology.

Though our initial audience proposal included ambitious social media outreach plans involving potential engagement and collaboration with popular philosophy and theory YouTube channels, podcasts, and blogs, in order to establish a social-networkless-social-network of thinkers to bolster our aim and philosophize over our findings, we quickly realized that networking in such a way is difficult without the results of the project finalized, properly assessed, and understood entirely. As those we’ve been in contact with throughout this process know well at this point, Modeling Value in the Anthropocene’s findings were not immediately evident as anything interesting, exciting, or even usable. This came as a source of brief anxiety and disappointment, triggering a critical reevaluation of our findings and a search for any subtle meaning that could be extracted from our word2vec results. With time and much discussion, “value” within our findings illuminated itself, allowing us just enough time to bolster our project’s website and adequately prepare for its fast-approaching presentation at the GC Digital Showcase, leaving little wiggle room in the remainder of our work plan to reconstruct an outreach plan, put together a “press kit,” and navigate the steps of building a working relationship necessary to coordinate any type of worthwhile collaborative project through the aforementioned mediums.

Considering that this initial aim was abandoned, the conclusion of our project has brought with it necessary reevaluations for the future social component of our work and the scholarly channels we’d like to see engaged with it as to be further critiqued, appreciated, developed, or collaborated on. Brian has managed to establish communication with Daniel Ross, Stiegler’s longtime friend, translator, and author of Psychopolitical Anaphylaxis: Steps Towards a Metacosmics, fulfilling one goal of scholarly outreach that we had presented in our initial proposal. Despite this dialogue being rooted in Brian’s work outside of Modeling Value in the Anthropocene, it elucidates the potential for such interlocution in the future stages of this work and allows for a sense of real possibility for our original, overly-ambitious audience proposal that sought out the likes of Ross Abbinnett and members of the Internation Collective as sounding boards for our work. This, along with our last-minute realization that the University of South Carolina’s resident digital humanist and text analysis authority, Michael Gavin, had been waiting for our request to take on a larger role within our project, has provided us with an exciting notion of what the future of this type of work could look like given the opportunity and time to foster such scholarly and consultative relationships.

As we’ve noted multiple times at this point, perhaps the biggest lesson learned through this project regarding audience and outreach has been, “Never hesitate to ask for help because you never know who might simply be waiting for you to ask.” The accessibility we’ve discovered to the very thinkers that inspired this work has been as intimidating as it is thrilling and we look forward to future iterations of this work now cognizant of the brilliant minds willing to engage with our work, regardless of how distantly.

Project Activities

Amidst this final stage of our project’s creation, reflecting on the initial work plan for Modeling Value in the Anthropocene set out in March reminds us of the experimental nature of our approach to this analysis and the necessary skill set that we worked to develop from the ground up in order to facilitate our intended investigation of Stiegler’s Nanjing Lectures. Though our original scope for this project included objectives such as the creation of an essay to detail our findings, our overly ambitious social media outreach goals noted above, and an unrealistic reading plan for the text that was quickly reassessed, the core outline of our inceptive work plan is surprising similar to that which we followed up until this point of retrospection. Our navigation of Python and text analysis via workshops, tutorials, and the NLTK workbook, along with countless YouTube videos and troubleshooting coding forums, was presciently outlined and planned for, allowing us to approach the immeasurable amount of information on such topics deliberately and assiduously as to make the most of our limited time. As a result of this intentionality in conjunction with our consultations with project mentors such as Leanne Fan, Michael Gavin, Filipa Colado, and Rafael Davis Portela, we were able to build the necessary foundation of skills as to properly and effectively carry out the text analysis equations as they were described in our project’s proposal at the start of the semester.

Though we deserted the notion of creating an essay to accompany our text analysis work in the early weeks of its development, we feel that our website, through the written portions located throughout detailing Stiegler’s philosophy as well as the philosophy and technicity of our approach, operates to bolster both our purpose in producing this project and the findings presented as a result of this process. Despite time being the factor that dissuaded us from the creation of such an essay, it was in the loosely structured final month stretch of our work plan that opportunities revealed themselves allowing for such efforts to communicate our theoretical and technical program to originate and enhance the Modeling Value in the Anthropocene website. The vague nature of the last month of our work plan further illuminates the creative ambiguity that was left open so that our work could mature without naively calculated restraints placed before our ideas and skillset could fully ripen. It was this openness that allowed us to expand our corpus to include texts from the Internation Collective and Daniel Ross, extend our vector analysis to include a wide array of unanticipated equations, and include Modeling Memory in the Anthropocene as a complementary element in our analysis.
Each stage of Modeling Value in the Anthropocene brought with it challenges that required us to reevaluate and restructure components of our project, eventually culminating in the briefly disappointing realization that the equation central to our word2vec analysis (V(value) – V(Anthropocene) + V(Neganthropocene)) had rendered less than immediately compelling results. However, upon further reflection and direction provided by Bret Maney and Michael Gavin, we were able and inspired to salvage such “non-results” and transform them into the bountiful grounds of interpretation that produced the philosophical exegesis elaborated on through our presentation at the Graduate Center’s Digital Showcase. Though our work has undoubtedly provided us with an opportunity for growth in our understanding of Bernard Stiegler and the scholarly possibilities provided to us through text analysis, it has also been an exercise in interpersonal problem solving, troubleshooting, and skill-development. Due to the complex nature of the Modeling Value in the Anthropocene’s proposal, it was fundamental to the project’s success that we expeditiously immersed ourselves in the world of Stieglerian thought and text analysis, regardless of rudimentary knowledge of one or the other, and advance our understanding through a cohesive and ambitious methodology.
Accomplishments

After completing this work, what we have are three products, all hosted on our website, which can be found at https://metacosmics.commons.gc.cuny.edu/. These products are: a Jupyter notebook file containing the Python script for our text analysis, some very basic and provisional writing containing some reflections on the results of our analyses’ queries, and a digital zine providing readers with an introduction to the philosophy of Bernard Stiegler. The Python script contains code for how to upload the (or any) text, train the word2vec model on the text, create new vectors in the model, and query for vectors with the greatest cosine similarity. The code is notated for the sake of intelligibility. The reflections on the results are some provisional thoughts on where this work could go and how it could guide a close reading of the work of Stiegler and others. The zine became a seeming necessity after our engagement with the digital humanities community regarding this project came up against the almost complete absence of familiarity with Stiegler’s work in this community. We felt a zine such as this could assist in our hopes for more appropriation of his ideas in the digital humanities world.

Evaluation

As we progressed this semester, feedback provided through our consultations with professors, digital fellows, and colleagues each acted as intermittent lodestars that we predominantly chose to follow, only occasionally neglecting to fully internalize such delineated directives and finding this out the hard way down the line. Aside from the thoughtful and supportive feedback provided by Bret Maney each week as we provided updates detailing our progress, our first piece of notable feedback from outside of the Graduate Center’s purview came from Michael Gavin at the University of South Carolina. In our initial meeting, Michael shared with us guidance regarding the struggles of interpretive clarity inherent in word2vec analyses, the benefits of utilizing a pre-trained model, the upsides to employing topic modeling and network graphs, and the necessity of breaking the corpus into subsections to be treated as individual documents and queried in comparison. Though our first meeting with Professor Gavin was wildly illuminating, our understanding of that which we were immersing ourselves in was still too limited to fully incorporate his invaluable instruction effectively into our project’s operation and general direction.

Our second meeting with Michael came shortly before our presentation at the Graduate Center Digital Showcase. Updating him on the progress we had made, along with the roadblocks and missteps, he evoked the cautionary advice he had provided months prior, suggesting that we had attempted exactly that which he had suggested not to attempt. Going on to question why we hadn’t reached back out sooner to engage with him further and avoid such lapses in project production, we realized that his feedback could have (and should have) played a larger role in our work, providing us with a deeply beneficial but hard-learned lesson to embrace as we move forward in our scholarly pursuits.

The feedback received from digital advisors such as Filipa Colado and Leanne Fan generally came in the form of collaborative working sessions via Zoom, allowing them to get their hands on our Python script, critique it, amend it, and provide recommendations for future development. These sessions were crucial to our advancement as coders, allowing us to troubleshoot and experiment under the instructive watch of some of the Graduate Center’s most talented digital scholars. It was primarily through these working sessions, along with our engagement with our peers in these early stages of script development, that we realized that the theoretical unpinning ushering along this code’s production required an accessible elucidation as to make the core objectives of our work both compelling and digestible to digital scholars unfamiliar with Stiegler’s philosophical project. To address this “weakness” of obscurity brought about through the sea of neologisms that one must swim through in order to grasp the core arguments of Stiegler’s work, we devised the “NeganthropoZene” to act as an introductory brochure for those interested but perhaps intimidated by the occasionally abstruse nature of our work. As this opacity was also mentioned in the Digital Showcase dress rehearsal, we are excited to have produced a resource that might help to shed light on Stiegler’s thought for curious citizens of the Anthropocene.

Lastly, the feedback received via the Digital Showcase was largely positive and restricted to brief kudos in the Zoom chat, providing little to extract and apply to the betterment of our project. However, after a semester of applying the critiques and directives of those we admire, we feel that our project is in a place that has recognized its weaknesses, engaged with them as an element of our presentation at the Digital Showcase, and addressed them to the best of our ability on the project’s website.

Future of the Project

Modeling Value in the Anthropocene is just the beginning of the work we will be doing utilizing text analysis in philosophical and other academic research. The goals we have for the future of this work are twofold. On the one hand, we will be taking the lessons gained from this project and bringing them to more mature text analysis that will lend itself to close reading and the production of philosophical writing that utilizes such analysis and reading. On the other hand, we hope to develop an application that can do the kind of word embedding conducted here (and in the ways we hope to adjust it in the future with the perspective we gained here) with a user-friendly GUI that will allow academics and other interested folks to do this kind of work without needing to know how to program. This will allow more researchers access to this tool and will hopefully contribute the work that so desperately needs to be done in the Anthropocene. This could also potentially be part of a larger idea of how to produce a word processor that could link tools such as this along with others to produce writing that could be hyperlinked in a large collaborative research network that could allow new knowledge to be transmitted and shared by others in new ways. This dimension of the project is, however, much more long-term in scope.

Anthropizing in the Anthropocene – Group Update

Hello fellow digital-humanists-in-training,

It has been an interesting couple of weeks for Hampton and I as we wind down toward the end of the spring semester and put the finishing touches on our text analysis project. Spring break saw us able to increase our corpus to five texts by Stiegler and some of his disciples. It also saw us accomplish what we set out to do, which was find a new vector which took the vector for “value” or “profit”, subtract from it the vector for “Anthropocene”, and add to it the vector for “Neganthropocene. The hypothesis was that by doing this, we would take away the characteristics of value which correspond to the Anthropocene (non-)epoch, and add those which correspond to the desired Neganthropocene epoch.

This has provided us with some interesting little tidbits to think about, such as the recurrence of the Greek topos ouranios, which bears a strong cosine similarity with our new vector for value in the Neganthropocene. The topos ouranios is the place in heaven where Plato believed that all the ideal forms were located. Stiegler (along with many other continental philosophers) are critical of this metaphysical construct which posits a transcendental world beyond. Stiegler is philosopher of materialism, situating the ideal world of thought within physical human bodies, technical systems of memory support, and material social relations. This poses the question: what similarity does that which ought to be valued in the Neganthropocene bear to this theological dimension? This reminds us of Stiegler’s treatment of Aristotle’s theos, wherein he resituates God as a dimension of being to which questions about being are posed and from which knowledge about being comes. It is also the object of all desire and all attention. We cannot go further into this now, but this is a little taste of what the results of our word embedding have got us thinking about.

Ultimately, though, however, our results have proven a little confusing, and a little disappointing. After a final meeting with Michael Gavin from the University of South Carolina, a literary scholar experienced in word embedding, we realized that there was a fundamental flaw in the premise of our project. When working with a corpus as small as ours, running simple word embedding models on the entire corpus creates too much noise for anything statistically significant to emerge. We have some ideas about how this project can be taken moving forward, and some better ways to approach word embedding. If nothing else, we have learned a vital lesson about text analysis, and have learned a fair amount of Python this semester.

So, we are pivoting the final results of our project a little bit in an attempt to show something a little more interesting. We have queried the model for 40 terms which we deemed most interesting and important to Stiegler’s work. We will be creating a table to visualize the 5 most similar words for each term. We will then be creating a network graph with Gephi to visualize this constellation of Stieglerian neganthropological concepts based on the results of these cosine similarity clusters. We are having done doing this work, and we can’t wait to share it with ya’ll.

Lastly, Hampton and I are working on a project for our Digital Memories class here at the GC that we are calling Modeling Memory in the Anthropocene and we will be hosting it on our website for this project. It will also be a Gephi network graph that models Stiegler’s conception of memory and how it relates and differs from the conceptualizations of memory in memory studies and digital memory studies. So keep an eye out for that.

Thanks for reading guys. We hope that you are all staying sane and healthy at this point in the semester and we look forward to hearing how your projects have been shaping up.

God bless.

Modeling Value in the Anthropocene – 3/31 Group Update

Modeling Value in the Anthropocene has regressed into a bit of entropy this week, as Brian’s immune system has experienced plague-induced degeneration and Hampton has come up against the structural incompleteness of his mind and his reality. Brian was unfortunately checked out for a week and a half due to a bout with COVID-19 while Hampton continued to dig further into word2vec and word embedding.

We have gotten to a point where our text is cleaned (with the cleaned version being stored locally on our machines), and we have successfully been able to play around with a word2vec model that has been pre-trained on a large corpus of text from Wikipedia. The end-goal we are working toward involves performing calculations on certain vectors from this pre-trained model with vectors from a model that we want to train on our Stiegler text. Now that we have successfully used the pre-train model, the next step for us is to train the model on our text. The problem that Hampton has been confronted with is a simple one, which is that of merely loading our cleaned text into our programming environment to begin to do so. This is where we run into problems as pirate digital humanists without a real background in computer science. However, we are very fortunate to be surrounded by a wealth of human beings with more experience than us who are willing to assist us. We have a second meeting with Leanne from the Digital Fellows Friday to figure out how to import our corpus and get back to work.

This situation (both Brian’s falling ill and Hampton’s Python woes) have also been a valuable lesson in two important Stieglerian concepts, adoption and quasi-causality (which he adopts from Gilles Deleuze and Felix Guattari). For Stiegler, following Heidegger, we are thrown into a world which has the character of being already-there, a world which pre-exists us and world over which we do not have control. We learn from Stiegler that the responsibility of the non-inhuman is to adopt this situation (of our lack of mastery over being) as necessary, and in so doing to make it quasi-causal for us. We do not have control over the initial blows that the world deals to us, but we do have the capacity to channel the libidinal energy that is generated as a result and make the situation the preindividual funds for a future-to-come. Working with Python is transforming the way we view failure, revealing its function as a departure point for progress.

Spoiler alert: I am uploading this a day late, so we have already met with Leanne, with whom we had the most productive meeting possible. She assisted us with getting our corpus into our Jupyter notebook and has brought us to the point where our model is now trained on our corpus. The next step will be figuring out what calculations we need to perform on our corpora in order to produce a new vector for value in the Neganthropocene and relate this vector to other vectors in the Stiegler text and the Wikipedia corpus. We hope to shed some light on these calculations for next week, our week thus amounting to a lot of Google-assisted programming. This will be weaved through with our continuing close reading of the lectures, which continue to stimulate us emotionally and intellectually.

Modeling Value in the Anthropocene – Data Management Plan

The Data

The original data source for the text is a searchable PDF file that is publicly available for free via Open Humanities Press’ website. We will also be appropriating Python scripts for running word embedding models also available online for free. Lastly, we will be generating notes to be stored in Jupyter Notebook files. The text and Python data will be regenerable so longs as the sources remain public and free. The PDF file will be converted to plain text. The code will initially be stored as .py files. Any visualizations will be image files (type TBD). The finished product (including code, text, and images) will be stored as an .ipynb Jupyter Notebook file. Jupyter Notebook will also be saved as PDF for sustainability. The tools we are utilizing are: Calibre for file conversion from PDF to .txt; Python and word2vec for text operationalization; Github for provisional code storage; Jupyter Notebook for project creation/storage. Github will be used to store, edit, and view code as we find, create, and alter it in the process of making. The final product will will also be stored on Github. We will backup files on our personal iClouds. File names will be the name of the author of the code and the version number. The different segments of code will be stored in the same directory on Github.

Data Standards

Data collection procedures are documented in a collaborative work place accessible as a collaborative .pages file. We are ensuring good project and data documentation via weekly in-person check-in meetings in addition to several virtual meetings per week. Both participants are responsible for implementing this data management plan. Project will follow the open access and open source practices of the Digital Humanities, the finished project being code and comments on said code that will be publicly available to all.

Data Security

Our data is not sensitive. There is no embargo period for our data. The text is licensed under a Creative Commons Attribution 4.0 Unported License, which allows us to share and adapt the text, so long as appropriate credit is given and a link to the license is provided with an indication that changes were made.

Re-use and re-distribution of data

The sharing of the data from this project are not subject to any external requirements. Our hopeful audience is members of the philosophical community interested in appropriating computation methods for theorizing. We will publish the data on our public website in May 2022. Jupyter will necessary to access the data published.

Long Term Archiving and Preservation

Data will be stored for 5-10 years unless a more permanent means of storage becomes available in the meantime. Data should be archived for Stiegler scholars or other researchers interested in the Neganthropocene. Jupyter notebook(s) will be stored as PDF’s for sustainability. Data will be submitted to CUNY Academic Works for long-term maintenance. A subject-based archive may also be found appropriate if found.

Modeling Value in the Anthropocene – Work Plan

3/2-3/9:

Stiegler Lectures: Finish the 2016 Lectures and work through the entirety of the 2017 Lectures (Pages 50 – 169)

Discuss and decide how to prepare the corpus

Continue expanding Python / Text Analysis skillset through workshops and tutorials
Work through NLTK work book

Reach out to Daniel Ross to establish contact with the Internation Collective as part of audience-building and scholarly outreach

3/10-3/16:

Stiegler Lectures: 2018 Lectures (Pages 169 – 269)

Begin to clean and prepare the corpus

Go to the Word Embeddings workshop

Continue expanding Python / Text Analysis skillset
Go through “Automate the boring stuff with Python” lesson online

Meet with Filipa

3/17-3/23:

Stiegler Lectures: 2019 Lectures (Pages 269 – 345)

Lemmatize corpus, remove punctuation, start writing loops

Begin experimenting with the text in Jupyter (via topic modeling, word embedding, etc.)

Continue expanding Python / Text Analysis skillset through workshops and tutorials

3/24 – 3/30:

Begin to formalize and finalize the research methods and functions through which we will be carrying out our close reading of the text.

Proceed with distant reading of the text.

3/31 – 4/6:

Finalize and begin to implement the methods through which we will be carrying out our distant and close reading of the text.

Continue to grow and apply our capabilities with Python, Text Analysis, & Word Embedding.

Develop essay outline (and section delegation) and formalize our method of presentation.

4/7 – 4/13:

Proceed with close reading and essay development.

4/14 – 4/20:

Continue to develop project.

4/21 – 4/27:

Assemble a rough draft of the project’s final product

Proceed with group revisions

Consultation with the “two-two’s” for project critiques

4/28 – 5/04:

Continue to develop and finalize the project.

Finalize method of presentation.

5/05 – 5/11:

Make final preparations.

Brian Millen – Personal Bio/Contribution Statement

Brian Millen received his B.A. in Philosophy from SUNY Purchase, and is currently an M.A. student in the Digital Humanities at the CUNY Graduate Center. During his time there, he submitted a senior thesis entitled Education and a Discipline Beyond Punish, which argued for a renewed role of education in global political transformation. His theoretical interests involve the relationship between humans and technology and the political consequences thereof. His research interests concern utilizing a combination of computational text analysis with more traditional scholarly work of philosophy to investigate technology, politics, economics, and strategies for overcoming the Anthropocene. Currently, Brian is working on a project called Modeling Value in the Anthropocene: Contributions to a metacosmics, a vector semantics analysis, for which he is fulfilling a role as project manager, as well as working co-extensively with research collaborator Hampton Dodd as co-developer and co-author.

Updated Project Proposal: Modeling Value in the Anthropocene

Abstract

Modeling Value in the Anthropocene is an attempt to accomplish two goals, one more general and one more specific. The first, more general one, is an attempt to provide an example of using computational tools in the world of philosophy. Doing this could enhance philosophy by utilizing methods that can process quantitative information faster than the speed of light. We argue that the particular tool of vector semantics can model large corpora of text in such a way that a new perspective can be created that could guide a subsequent close reading. Our hope is that this prior modeling can provide the potential for a close reading that leads to the creation of a concept, which is the aim of philosophy.
The particular question that this project will try to answer is something like: what concept should ground individual and collective value in the world beyond the Anthropocene? After using word2vec to model the conceptual networks of the concepts “profit,” “Anthropocene,” and “Neganthropocene” in the bibliography of philosopher Bernard Stiegler, we will then perform calculations on these vectors to unveil a new vector, the concept it’s representing being the one that will guide our close reading of one of his texts. This close reading will lend itself to the production of an article/potential book chapter. This will be part of a larger project of modeling the concepts of “entropy” and “negentropy” in the history of Western thought, which might result in a book on what Daniel Ross calls metacosmics.

 

List of Participants

Brian Millen, Digital Humanities Masters Student at CUNY Graduate Center, will be the project manager, co-developer, and co-researcher/author.
Hampton Dodd, Digital Humanities Masters Student at CUNY Graduate Center, will be the secretary, co-developer, and co-researcher/author

 

Enhancing the Humanities through Innovation

In proposing a reading of philosophical texts that makes use of algorithms for natural language processing, one would have to consider what, if anything, these computational methods of reading and writing bring to doing the work of philosophy. Natural language processing consists of performing mathematical calculations on language. In text analysis specifically, we use digital text as data that we can then model in different ways by making quantitative calculations that a human mind would never be able to do with such large amounts of text on its own through close reading. There are many arguments for and against this kind of practice, and it is likely that many native to the world of philosophy would be especially skeptical of accepting a computational approach. We believe one way of arguing how text analysis could contribute to philosophical practice is through Michael Witmore’s blog post, “Text: A Massively Addressable Object.” In this post, he situates distant reading as simultaneously continuous with the practice of reading since its genesis and discontinuous with past modalities of reading.

Witmore writes that what separates digital text from older forms of text is that “it is massively addressable at different levels of scale” (Witmore). To understand this requires understanding what he means by “scale.” According to Witmore, reading has always been the practice of abstracting a material text at the ideal level at which one is to read it. For example, one can choose to address a text at the level of the word, the sentence, the paragraph, or the book, just to name a few. These are not material characteristics of the text itself, but rather subjective ways of dividing the text up and projecting them into/out of the text. A reader could just as easily choose to underline all the words of a text beginning with the letter ‘B’ and address it at that scale. How text is addressed is a matter of convention, contingent on the technical and normative limits of the given socio-historical context of the act of reading. The fact that we tend to read text as “books” or even “words” is simply a matter of convention. As Witmore writes, “The idea of a word is itself an artifact of manuscript culture, one that could be perpetuated in print through the affordances of moveable type” (Witmore). What makes digital text different, then, is the scale of addressability, or rather the many different scales at which text can now be addressed. Instead of having to read one word at a time or one sentence at a time, we can query and compare many words from many different documents at once. A popular form of this found in the digital humanities is topic modeling. Topic models are generated by running an unsupervised machine learning algorithm on a group of documents and approximating which words tend to be used together in the same document. This allows us to address the texts at the level of “topic,” creating ideal topics we can say that the texts appear to be drawing from (ideal in the sense of non-real; in the realm of the potential or virtual, discoverable in the imagination, as distinguished from the actual).

The form of computational address we propose using in a philosophical context is one called vector semantics analysis. It is a technique of computational linguistics wherein we can run an algorithm that will group words together that bear a semantic similarity to one another. This similarity is represented using a particular kind of vector analysis called word embeddings. Word embeddings assign each word in the text(s) a numerical value based on their distributional distance from every other word in the text. We can then map these words graphically which represent concepts as networks of words used “synonymously” in the text. This is useful for conceptual analysis because words that have similar vectors to each other will appear clustered together within the vector space, indicating that they are distributed throughout the texts in a similar way. The linguistic theory operating behind this model is that words that are deployed in similar contexts have some conceptual resonance: “The latent idea here is that different words will tend to appear in different contexts, and therefore one can guess at a word’s meaning by simply counting the words that appear near it” (Gavin et al.). Bringing it back to the language of Witmore, vector semantics is a way of addressing large amounts of text at the level of semantic similarity. It is an addressing of the text at the level of the word, but it is a level where each word exists in relation to every other word, the level getting more discreet as we narrow our investigation. We can thus say this method allows us to address text at the level of concept. This level of address is obviously not new in philosophy. Finding continuity at the level of conceptuality is the modus operandi of philosophy, and we have been used to addressing concepts at the scale of the entire history of philosophy since at least Hegel. What is new here is the way we can address this history and its concepts.

The vector semantic analysis we are proposing is that of the concepts of entropy and negentropy in the history of Western thought. Bernard Stiegler teaches us that “the relation entropy/negentropy is really the question of life par excellence” (Stiegler, 39). Entropy is a term taken from thermodynamics denoting the tendency of any closed system to tend toward a state of disorder. Physicists of the 19th century came to apply this to the entire universe conceived of as a closed system, effectively reversing the Western conception of the cosmos, which since Socrates had been understood as equilibrium. The concept of negentropy was introduced by Erwin Schrödinger in his lectures published as What Is Life? to give an account of life understood as the temporary struggle against entropy, organic matter being organized matter that wards off the dissipation of energy. The physicist Alfred Lotka takes a step beyond Schrödinger to argue that humans are a particular kind of negentropy, temporarily suspending entropy by means other than just biological organs generated by natural selection. In other words, humans produce negentropy through artificial means of various kinds: tools, writing, machines, etc. According to Stiegler, the implications of these breakthroughs in scientific thought have yet to have been thought through by philosophy. They must be thought through in the context of the Anthropocene, which he claims is an epoch of history marked by extreme accelerations of entropy through industrialization, accelerations that threaten the survival of the human race. The Anthropocene is a proposed geological epoch in which human beings become a geophysical force affecting the biosphere. However, for Stiegler, it is important to recognize that this epoch is firstly a technological, economic, and political epoch. Thinking through the concepts of entropy and negentropy in this context is the work that he called for, and it is the work he attempted until his death in August 2020.

The long-term project we would like to embark on is an investigation of these concepts (in their inclusion as well as in their absence) over a long period of time, covering the histories of physics, biology, philosophy, economics, and information science. Using vector semantics, we wish to find out what other words bear a semantic resemblance to the words “entropy” and negentropy”. The goal of this project would be a contribution to what Daniel Ross calls “metacosmics”, which is a destruction of/continuation of metaphysics centered on a general theory of entropy/negentropy. This project would be oriented toward understanding how these concepts change over time, as well as seeing what resonance certain thinkers (like Nietzsche) have with these concepts, despite not using their names. This would complexify and intensify our conceptions of entropy and negentropy. The scope of this project would start much smaller though, beginning with just modeling these concepts in Stiegler’s work alone, further branching out to other works and disciplines in future projects. We wish to perform a conceptual analysis of his work that we will describe in the following section, an analysis that would be part of a close reading, a reading that will be a small part of this larger project.

 

Environmental Scan

Much of the work of distant reading in the digital humanities utilizes methods of topic modeling. Something like topic modeling can be incredibly useful for philosophy, especially the history of philosophy. It can address large amounts of text at a scale that would allow us to understand what topics are characteristic of a particular school of thought (like German idealism) or a particular epoch of history (like the Enlightenment). Malaterre et al. run an unsupervised machine learning algorithm on the complete corpus of the journal Philosophy of Science, mining articles from 1934 to 2015. One result of this was the revelation that the topics that are consistent throughout the whole time are “metaphysical and ontological questioning” about “space and time, causation, natural kinds, [and] realism” (Malaterre et al, 215). Another was the discovery of how the topics of logic and language rose and fell throughout time.
Despite topic modeling’s potential fruitfulness, we wish to show why vector semantics is a better method specifically for doing the work of philosophy. There are two reasons for this. The first is because it allows us to model semantic similarity among different words. As Gavin et al. argue,

concepts are not words. The distinction can be glimpsed by considering any collection of synonyms, such as rubbish, trash, waste, junk. Each of these words has a distinct use, and they imply different connotations and meanings, but there remains a principle of synonymy among them, an underlying concept that ties them together (Gavin et al.)

Topic models show us what words tend to be used together. Word embeddings show us which words tend to be used with the same words. Topic modeling could perhaps guide a close reading exploring the topic deeper, or what role a key word might play in a topic, but it cannot get to the semantic depth we might want to in the work of philosophy.

The other reason why we prefer vector semantics for philosophy is because philosophy is not just the modeling of concepts. As Deleuze and Guattari write, “The philosopher is the concept’s friend; he is the potentiality of the concept…philosophy is the discipline that involves creating concepts” (Deleuze et al., 5). The operations of word embedding alone already described are useful for clarifying and perhaps deepening concepts, and could possibly lend themselves to some conceptual reevaluation. However, there is another operation made possible by word embeddings that contains so much more potential for philosophy. Because the words are embedded with numerical values, once the conceptual networks have been modeled in vector space, we can actually perform calculations on the vectors that create new vectors based on this math, and these new vectors can open a path to the creation of concepts, or what Stiegler calls “categorial invention,” which is the goal of philosophy, and perhaps of all theoretical work. A well-cited example is that of V(Queen) = V(King) + V(Woman) – V(Man). When taking the vector for “King”, adding the vector for “Woman” to it, and subtracting the vector for “Man”, the model has been proven to successfully output the vector for “Queen.” What this means conceptually is that if we add the qualities of women to the concept of kings, we have some concept which would have all the qualities of women and all the qualities of kings. If we then subtract from this new vector everything associated with men, we get the concept of queen. This is a simple example, but this functionality can prove exceptionally useful for philosophy.

One of the principal things Stiegler calls for is a rethinking of value to escape the Anthropocene, initializing what he calls the epoch of the Neganthropocene. One chief problem with capitalism, he claims, is that, under the conditions it initiates, all use value is reduced to exchange value. The usefulness of a thing is reified into how much it costs, or how much money it could make. This reduces everything to the standards of valuation of the market. The progression of this dynamic accounts for how things like the law or works of art have been devalued, not to mention the health of the biosphere and the future itself. Thus, the Neganthropocene, which would be the epoch following the Anthropocene (if there is to be one), would have to be generated on the basis of a new valuation. The question, then, is if the value of everything is no longer to be based on profit, what is to be the new value founding this society? We can contribute to the thinking through of this question by treating Stiegler’s works with vector semantics. We propose starting off by querying a sample equation that looks something like V(profit) – V(Anthropocene) + V(Neganthropocene). This would take the concept of profit, which grounds value in this current stage of capitalism, subtract that which is characteristic of the Anthropocene, and add the vector representing the things that Stiegler writes about the Neganthropocene. This analogic calculation might point us in the direction of which words will be related together as all having to do with how we should re-ground value beyond the Anthropocene. We will run word2vec, a vector semantic algorithm, on two of Stiegler’s texts: Nanjing Lectures 2016-2019, where he lays out his theories of entropy, negentropy, Anthropocene, and Neganthropocene most systematically.

It should be made very clear that this type of calculation is not a magic wand that can reveal new concepts for us on its own. Witmore’s account of distant reading focuses on the scale of the address, but it does not take into full account the shape or contours of the address itself. We would argue that there are two main modes with which one can address text: analytic and synthetic. These neo-Kantian faculties that Stiegler articulates are two forces that make up the dialectic of knowledge production. The full explication of these arguments is beyond the scope of this proposal, but they show that calculating text (or any data) requires the synthetic work of the imagination to think according to standards of reason, and more importantly to dream up new concepts that do not fit into the analytic schema of the understanding. Information or data is the externalization of a prior synthetic act of reason that is calculable now that it is materialized. This act is a decomposition of the line of reasoning into discrete elements that can thus be quantified and calculated. This act is entropic in and of itself, but can produce new knowledge, new concepts, if it leads to a surprise which causes one to run it through their filter of reason and create a new idea which re-organizes the analytical understanding as it now stands. In other words, by modeling text, one divides it up into enormous different pieces (in this case, vectors) that one can perform calculations on. On their own, these models and these calculations are useless. However, an act like querying Stiegler’s texts for the answer to V(profit) – V(Anthropocene) + V(Neganthropocene) could open up a path that one could wander down. And perhaps, by wandering down this path, which would include careful thought, reasoning, and close reading, one could perhaps experience a surprise in the text. This surprise could potentially cause one to rethink the text they are reading closely in a new way, and potentially lead to the production of a concept. There is of course no way to guarantee this, but it is only by seeking out that which is incalculable that philosophy can be done. Perhaps vector semantics could be a kind of calculation that leads the way toward thinking about value anew and how a new society can be built upon this new concept of value. This could then guide a close reading of some of Stiegler’s texts that could potentially concretize this new, currently unknown, concept.

 

Work Plan

Education
Both participants will spend the first portion of the semester learning the basics of Python and text analysis, followed by time spent learning how to use word2vec. The specific work plan will be fleshed out more when this is done. Regular (weekly or bi-weekly) re-assessment will be a necessary part of the work flow, which is in nature more open and porous.

Digitizing
Making the text machine-readable will be a collaborative effort. The text is available in an open access searchable PDF format. Software will be used to convert it to plain text format and the results will be hand-checked against the original.

Operationalization
Once participants are familiar with the word2vec tool, they will train the algorithm on the text in question, creating vectors based on cosine similarity. These vectors will then be operationalized to determine the new vector-concept. This step will be open to the possibility of failure and the potential need for alternative lines of questioning opened up by playing around with the tool. There are two options for doing the actual work of text analysis collaboratively. Participants will experiment with both a) doing the work separately and meeting regularly to exchange and discuss results and b) pair programming, doing the text analysis on the same machine or sharing a screen via video chat.

Close reading/writing
Utilizing the fruits of the vector analysis, the participants will then perform a close reading of the texts at hand guided by the vector produced by the algorithm. This will require the development of a strategy for collaborative note-taking and writing. Participants may utilize a collaborative version of the Zettelkasten method to bridge emerging ideas into the production of a unified text. The scope of this project will focus on a rough draft of a piece of writing, revision, and publication of which will be the work of a future iteration.

 

Final Product and Dissemination

As stated, this project is to form a small part of a larger project about entropy and negentropy in the history of Western thought more generally. This particular project will lend itself to a shorter piece of writing that will specifically be about the question of value in the Neganthropocene. It will initially be published online as a blog post. It will not only be provisionally about the conceptual framework needed to reevaluate value, but it will also form the foreground for this larger project on metacosmics. Thus, this close reading and writing will also be the work of forming the questions we would like to pose in future work, as well as the kinds of texts that may need to be addressed.

On project management in the Digital Humanities

The NYC Digital Humanities Week event that I will be reflecting on was entitled “DH Project Management”, a talk organized by Jesse Merandy and Kimon Keramidas. I chose this talk more generally because of the projects we are all embarking on this academic semester, but also because of the specific questions I had regarding the specificity of the project my partner and I are embarking on. We are doing a collaborative text analysis, a digital methodology that I have been having trouble conceiving of having a group dynamic. What will project management look like in this context?

The speakers started by going over three basic paradigms with which digital humanists usually approach their projects. The first do are mutually exclusive, but the third will fall into one of the first two categories. The first two are that of original collection and existing collection and are related to the kind of data to be used for the project. For a project with an existing collection, there is a pre-existing data set, so it is a question of data aggregation more than data collection. However, plan still needs to be set up for a data repository for the project. The repository must be compatible with the incoming data. The tool selection also has to be compatible with existing data structures.

For a project with an original collection of data, there is no pre-existing data set, and thus it must be gathered. The question that must be first thought through is what needs to be gathered and how will it be gathered? The same rules apply for the repository, but in this kind of project, there is more of a dialectic between tool data. As with the pre-existing set, the tool selection must be compatible with the data, but the tool can also drive the data collection. This is the kind of project that my team will be working on this semester.

The third paradigm is a team-based project. This was the most important portion of the talk for me, given, as I said, that we are working on group projects this semester, but especially because the elements of team work that were stressed were very much at odds with the way I was planning on approaching my own project. The speakers emphasized that organization should be the number one priority of team-based digital humanities work. There must be clear project goals derived from the project proposal/description, and a plan must be built around this that his focussed and communicated with all members of the team. Scheduling is more complex and crucial for groups than for individual projects, which have the liberty to be more fluid with less moving parts and thus less variables. This upset the expectation that I had of my two-person team project being loosely organized, working at whatever pace felt right, having an amorphous work plan that would evolve as it went. The speakers made it clear that this kind of orientation was surefire way to not get what we want accomplished. The project goals must be as clear as possible so the tasks can be made as discrete as possible so as to make sure that we are staying on track. This will also allow a team to build in space for failure, course correction, and other constraints of working on a project.

I am indebted to these speakers for this insight, as well to the opportunity to do this group-oriented work in the first place. The collaborative spirit of digital humanities is one of the elements that drew me to the field in the first place. As one of the speakers said, despite the challenges (or maybe because of the challenges) that come with collaborative work, what it offers is the chance for a work that is more of the sum of its parts. I am grateful to be able to surrender this idea that I had which led to the creation of a project proposal, and allow this project to be individuated in a way which is social and goes beyond what I would be capable of alone.

Brian Millen Skillset

Project management: While I do not have specific experience working in project management, I do have experience in operations coordination, coordinating between different teams at a social media company. My job was to receive common complaints coming from the user’s of the mobile application and work with the tech team to address these issues. This required the skills of organization, communication, and coordinating between different teams’ and individuals’ workflows, making me comfortable taking on a position of project management.

Design/UX: While I did not work directly with design and product management at the social media company, I worked closely with user experiences of the mobile app, giving me insight into user experience on digital platforms. Although I would not be interested in taking on this position directly, I do believe it is something I could provide insight into.

Documentation/Research & writing: Coming from a humanities background, I believe writing to be one of my stronger skills. Furthermore, with my job at the social media company, I was responsible for documenting help articles and FAQ’s for the platform’s Customer Care team. Either of these roles would be my preferred contribution.

Developer: This is the dimension I have probably the least experience in, but it is also where I would like to grow the most in. I would be interested in taking a secondary position in assisting with the more technical aspects of the project.

Modeling Value in the Anthropocene: Contributions to a metacosmics (Project Proposal)

Abstract

Modeling Value in the Anthropocene is an attempt to accomplish two goals, one more general and one more specific. The first, more general one, is an attempt to provide an example of using computational tools in the world of philosophy. Doing this could enhance philosophy by utilizing methods that can process quantitative information faster than the speed of light. I argue that the particular tool of vector semantics can model large corpora of text in such a way that a new perspective can be created that could guide a subsequent close reading. My hope is that this prior modeling can provide the potential for a close reading that leads to the creation of a concept, which is the aim of philosophy.

The particular question that this project will try to answer is something like: what concept should ground individual and collective value in the world beyond the Anthropocene? After using word2vec to model the conceptual networks of the concepts “profit,” “Anthropocene,” and “Neganthropocene” in the bibliography of philosopher Bernard Stiegler, I will then perform calculations on these vectors to unveil a new vector, the concept it’s representing being the one that will guide my close reading of one of his texts. This close reading will lend itself to the production of an article/potential book chapter. This will be part of a larger project of modeling the concepts of “entropy” and “negentropy” in the history of Western thought, which will result in a book on what Daniel Ross calls metacosmics.

List of Participants

2-3 participants to be responsible for collaborating on the digitizing of the texts, the operationalizing of the texts, the close reading, and the production of an article/book chapter.

Enhancing the Humanities through Innovation

In proposing a reading of philosophical texts that makes use of algorithms for natural language processing, one would have to consider what, if anything, these computational methods of reading and writing bring to doing the work of philosophy. Natural language processing consists of performing mathematical calculations on language. In text analysis specifically, we use digital text as data that we can then model in different ways by making quantitative calculations that a human mind would never be able to do with such large amounts of text on its own through close reading. There are many arguments for and against this kind of practice, and it is likely that many native to the world of philosophy would be especially skeptical of accepting a computational approach. I believe one way of arguing how text analysis could contribute to philosophical practice is through Michael Witmore’s blog post, “Text: A Massively Addressable Object.” In this post, he situates distant reading as simultaneously continuous with the practice of reading since its genesis and discontinuous with past modalities of reading.

Witmore writes that what separates digital text from older forms of text is that “it is massively addressable at different levels of scale” (Witmore). To understand this requires understanding what he means by “scale.” According to Witmore, reading has always been the practice of abstracting a material text at the ideal level at which one is to read it. For example, one can choose to address a text at the level of the word, the sentence, the paragraph, or the book, just to name a few. These are not material characteristics of the text itself, but rather subjective ways of dividing the text up and projecting them into/out of the text. A reader could just as easily choose to underline all the words of a text beginning with the letter ‘B’ and address it at that scale. How text is addressed is a matter of convention, contingent on the technical and normative limits of the given socio-historical context of the act of reading. The fact that we tend to read text as “books” or even “words” is simply a matter of convention. As Witmore writes, “The idea of a word is itself an artifact of manuscript culture, one that could be perpetuated in print through the affordances of moveable type” (Witmore). What makes digital text different, then, is the scale of addressability, or rather the many different scales at which text can now be addressed. Instead of having to read one word at a time or one sentence at a time, we can query and compare many words from many different documents at once. A popular form of this found in the digital humanities is topic modeling. Topic models are generated by running an unsupervised machine learning algorithm on a group of documents and approximating which words tend to be used together in the same document. This allows us to address the texts at the level of “topic,” creating ideal topics we can say that the texts appear to be drawing from (ideal in the sense of non-real; in the realm of the potential or virtual, discoverable in the imagination, as distinguished from the actual).

The form of computational address I propose using in a philosophical context is one called vector semantics analysis. It is a technique of computational linguistics wherein we can run an algorithm that will group words together that bear a semantic similarity to one another. This similarity is represented using a particular kind of vector analysis called word embeddings. Word embeddings assign each word in the text(s) a numerical value based on their distributional distance from every other word in the text. We can then map these words graphically which represent concepts as networks of words used “synonymously” in the text. This is useful for conceptual analysis because words that have similar vectors to each other will appear clustered together within the vector space, indicating that they are distributed throughout the texts in a similar way. The linguistic theory operating behind this model is that words that are deployed in similar contexts have some conceptual resonance: “The latent idea here is that different words will tend to appear in different contexts, and therefore one can guess at a word’s meaning by simply counting the words that appear near it” (Gavin et al.). Bringing it back to the language of Witmore, vector semantics is a way of addressing large amounts of text at the level of semantic similarity. It is an addressing the text at the level of the word, but it is a level where each word exists in relation to every other word, the level getting more discreet as we narrow our investigation. We can thus say this method allows us to address text at the level of concept. This level of address is obviously not new in philosophy. Finding continuity at the level of conceptuality is the modus operandi of philosophy, and we have been used to addressing concepts at the scale of the entire history of philosophy since at least Hegel. What it is new here is the way we can address this history and its concepts.

The vector semantic analysis I am proposing is that of the concepts of entropy and negentropy in the history of Western thought. Bernard Stiegler teaches us that “the relation entropy/negentropy is really the question of life par excellence” (Stiegler, 39). Entropy is a term taken from thermodynamics denoting the tendency of any closed system to tend toward a state of disorder. Physicists of the 19th century came to apply this to the entire universe conceived of as a closed system, effectively reversing the Western conception of the cosmos, which since Socrates had been understood as equilibrium. The concept of negentropy was introduced by Erwin Schrödinger in his lectures published as What Is Life? to give an account of life understood as the temporary struggle against entropy, organic matter being organized matter that wards off the dissipation of energy. The physicist Alfred Lotka takes a step beyond Schrödinger to argue that humans are a particular kind of negentropy, temporarily suspending entropy by means other than just biological organs generated by natural selection. In other words, humans produce negentropy through artificial means of various kinds: tools, writing, machines, etc. According to Stiegler, the implications of these breakthroughs in scientific thought have yet to have been thought through by philosophy. They must be thought through in the context of the Anthropocene, which he claims is an epoch of history marked by extreme accelerations of entropy through industrialization, accelerations that threaten the survival of the human race. The Anthropocene is a proposed geological epoch in which human beings become a geophysical force affecting the biosphere. However, for Stiegler, it is important to recognize that this epoch is firstly a technological, economic, and political epoch. Thinking through the concepts of entropy and negentropy in this context is the work that he called for, and it is the work he attempted until his death in August 2020.

The long-term project I would like to embark on is an investigation of these concepts (in their inclusion as well as in their absence) over a long period of time, covering the histories of physics, biology, philosophy, economics, and information science. Using vector semantics, I wish to find out what other words bear a semantic resemblance to the words “entropy” and negentropy”. The goal of this project would be a contribution to what Daniel Ross calls “metacosmics”, which is a destruction of/continuation of metaphysics centered on a general theory of entropy/negentropy. This project would be oriented toward understanding how these concepts change over time, as well as seeing what resonance certain thinkers (like Nietzsche) have with these concepts, despite not using their names. This would complexify and intensify our conceptions of entropy and negentropy. The scope of this project would start much smaller though, beginning with just modeling these concepts in Stiegler’s work alone, further branching out to other works and disciplines in future projects. I wish to perform a conceptual analysis of his work that I will describe in the following section, an analysis that would be part of a close reading, a reading that will be a small part of this larger project.

Environmental Scan

Much of the work of distant reading in the digital humanities utilizes methods of topic modeling. Something like topic modeling can be incredibly useful for philosophy, especially the history of philosophy. It can address large amounts of text at a scale that would allow us to understand what topics are characteristic of a particular school of thought (like German idealism) or a particular epoch of history (like the Enlightenment). Malaterre et al. run an unsupervised machine learning algorithm on the complete corpus of the journal Philosophy of Science, mining articles from 1934 to 2015. One result of this was the revelation that the topics that are consistent throughout the whole time are “metaphysical and ontological questioning” about “space and time, causation, natural kinds, [and] realism” (Malaterre et al, 215). Another was the discovery of how the topics of logic and language rose and fell throughout time.
Despite topic modeling’s potential fruitfulness, I wish to show why vector semantics is a better method specifically for doing the work of philosophy. There are two reasons for this. The first is because it allows us to model semantic similarity among different words. As Gavin et al. argue,

concepts are not words. The distinction can be glimpsed by considering any collection of synonyms, such as rubbish, trash, waste, junk. Each of these words has a distinct use, and they imply different connotations and meanings, but there remains a principle of synonymy among them, an underlying concept that ties them together (Gavin et al.)

Topic models show us what words tend to be used together. Word embeddings show us which words tend to be used with the same words. Topic modeling could perhaps guide a close reading exploring the topic deeper, or what role a key word might play in a topic, but it cannot get to the semantic depth we might want to in the work of philosophy.

The other reason why I prefer vector semantics for philosophy is because philosophy is not just the modeling of concepts. As Deleuze and Guattari write, “The philosopher is the concept’s friend; he is the potentiality of the concept…philosophy is the discipline that involves creating concepts” (Deleuze et al., 5). The operations of word embedding alone already described are useful for clarifying and perhaps deepening concepts, and could possibly lend themselves to some conceptual reevaluation. However, there is another operation made possible by word embeddings that contains so much more potential for philosophy. Because the words are embedded with numerical values, once the conceptual networks have been modeled in vector space, we can actually perform calculations on the vectors that create new vectors based on this math, and these new vectors can open a path to the creation of concepts, or what Stiegler calls “categorial invention,” which is the goal of philosophy, and perhaps of all theoretical work. A well-cited example is that of V(Queen) = V(King) + V(Woman) – V(Man). When taking the vector for “king”, adding the vector for “woman” to it, and subtract the vector for “man”, the model has been proven to successfully output the vector for queen. What this means conceptually is that if we add the qualities of women to the concept of kings, we have some concept which would have all the qualities of women and all the qualities of kings. If we then subtract from this new vector everything associated with men, we get the concept of queen. This is a simple example, but this functionality can prove exceptionally useful for philosophy.

One of the principal things Stiegler calls for is a rethinking of value to escape the Anthropocene, initializing what he calls the epoch of the Neganthropocene. One chief problem with capitalism, he claims, is that, under the conditions it initiates, all use value is reduced to exchange value. The usefulness of a thing is reified into how much it costs, or how much money it could make. This reduces everything to the rules of the market. The progression of this dynamic is how things like the law or works of art have been devalued, not to mention the health of the biosphere and the future itself. Thus, the Neganthropocene, which would be the epoch following the Anthropocene (if there is to be one), would have to be generated on the basis of a new valuation. The question, then, is if the value of everything is no longer to be based on profit, what is to be the new value founding this society? We can contribute to the thinking through of this question by treating Stiegler’s works with vector semantics. I propose querying a sample equation that looks something like V(profit) – V(Anthropocene) + V(Neganthropocene). This would take the concept of profit, which grounds value in this current stage of capitalism, subtract that which is characteristic of the Anthropocene, and add the vector representing the things that Stiegler writes about the Neganthropocene. This analogic calculation might point us in the direction of which words will be related together as all having to do with how we should re-ground value beyond the Anthropocene. I will run word2vec, a vector semantic algorithm, on two of Stiegler’s texts: Nanjing Lectures 2016-2019, where he lays out his theories of entropy, negentropy, Anthropocene, and Neganthropocene most systematically, and Technics and Time, Vol. 1: The Fault of Epimetheus, his first text, before he began to speak of these concepts at all, but where he laid the grounds for his philosophical work to come.

It should be made very clear that this type of calculation is not a magic wand that can reveal new concepts for us on its own. Witmore’s account of distant reading focuses on the scale of the address, but it does not take into full account the shape or contours of the address itself. I would argue that there are two main modes with which one can address text: analytic and synthetic. These neo-Kantian faculties that Stiegler articulates are two forces that make up the dialectic of knowledge production. The full explication of these arguments are beyond the scope of this proposal, but they show that calculating text (or any data) requires the synthetic work of the imagination to think according to standards of reason, and more importantly to dream up new concepts that do not fit into the analytic schema of the understanding. Information or data is the externalization of a prior synthetic act of reason that is calculable now that it is materialized. This act is a decomposition of the line of reasoning into discrete elements that can thus be quantified and calculated. This act is entropic in and of itself, but can produce new knowledge, new concepts, if it leads to a surprise which causes one to run it through their filter of reason and create a new idea which re-organizes the analytical understanding as it now stands. In other words, by modeling text, one divides it up into an enormous different pieces (in this case, vectors) that one can perform calculations on. On their own, these models and these calculations are useless. However, an act like querying Stiegler’s texts for the answer to V(profit) – V(Anthropocene) + V(Neganthropocene) could open up a path that one could wander down. And perhaps, by wandering down this path, which would include careful thought, reasoning, and close reading, one could perhaps experience a surprise in the text. This surprise could potentially cause one to rethink the text they are reading closely in a new way, and potentially lead to the production of a concept. There is of course no way to guarantee this, but it is only by seeking out that which is incalculable that philosophy can be done. Perhaps vector semantics could be a kind of calculation that leads the way toward thinking about value anew and how a new society can be built upon this new concept of value. This could then guide a close reading of some of Stiegler’s texts that could potentially concretize this new, currently unknown, concept.

Work Plan

Digitizing
The work of making the texts digital will take place over the course of a week. Both texts are available online in PDF format. The week will be spent turning them into plain text files manually by typing them into a plain text editor.

Operationalization
Participant will spend a few weeks learning how to use word2vec, a popular algorithm for performing vector semantics. Once familiar with the tool, he will train the algorithm on the two texts in question, creating vectors based on cosine similarity. These vectors will then be operationalized to determine the new vector-concept. This step will be open to the possibility of failure and the potential need for alternative lines of questioning opened up by playing around with the tool.

Close reading/writing
Utilizing the fruits of the vector analysis, the participant will then perform a close reading of the texts at hand guided by the vector produced by the algorithm. This will result in an article surrounding the subject of value in the Neganthropocene.

Final Product and Dissemination

As stated, this project is to form a small part of a larger project about entropy and negentropy in the history of Western thought more generally. This particular project will lend itself to a shorter piece of writing that will specifically be about the question of value in the Neganthropocene. It will initially be published online as a blog post. It will not only be provisionally about the conceptual framework need to reevaluate value, but it will also form the foreground for this larger project on metacosmics. Thus, this close reading and writing will also be the work of forming the questions I would like to pose in future work, as well as the kinds of texts that may need to be addressed.

Works Cited

Deleuze, Gilles, and Félix Guattari. What Is Philosophy? Translated by Hugh Tomlinson and Graham Burchell, Verso, 2015.

Gavin, Michael, et al. “Spaces of Meaning: Conceptual History, Vector Semantics, and Close Reading.” University of Minnesota Press, Minneapolis, MN, 2019.

Malaterre, Christophe, et al. “What Is This Thing Called Philosophy of Science? A Computational Topic-Modeling Perspective, 1934–2015.” HOPOS: The Journal of the International Society for the History of Philosophy of Science, vol. 9, no. 2, 2019, pp. 215–249., https://doi.org/10.1086/704372.

Stiegler, Bernard. The Neganthropocene. Translated by Daniel Ross, 1st ed., Open Humanities Press, 2018.

Witmore, Michael. “Text: A Massively Addressable Object.” Debates in the Digital Humanities, The University of Minnesota Press, Minneapolis, MN, 2012, https://dhdebates.gc.cuny.edu/read/untitled-88c11800-9446-469b-a3be-3fdb36bfbd1e/section/402e7e9a-359b-4b11-8386-a1b48e40425a#p4b3. Accessed 28 Oct. 2021.