Tooling Up for Digital Humanities

  • Development Blog
  • Documentation
  • Plugins
  • Suggest Ideas
  • Support Forum
  • Themes
  • WordPress Planet
  • Home
  • Workshop Series
  • About
  • Virtual You
    • 1: Virtual You
    • 2: Keeping a Finger on the Pulse
    • 3: Building Community
    • 4: Further Reading
    • 5: Discussion
  • Digitization
    • 1: Making Documents Digital
    • 2: Metadata and Text Markup
    • 3: Further Reading
    • 4: Discussion
  • Text Analysis
    • 1: The Text Deluge
    • 2: A Brief History
    • 3: Stylometry
    • 4: Content-Based Analysis
    • 5: Metadata Analysis
    • 6: Conclusion
    • 7: Further Reading
    • 8: Discussion
  • Spatial Analysis
    • 1: The Spatial Turn
    • 2: Spatial History Lab
    • 3: Geographic Information Systems
    • 4: Further Reading
    • 5: Discussion
  • Databases
    • 1: The Basics
    • 2: Managing Your Bibliography
    • 3: Cloud Computing
    • 4: Organizing Images
    • 5: Further Reading
    • 6: Discussion
  • Pedagogy
    • 1: In the Classroom
    • 2: Student Collaboration
    • 3: Debating Pedagogical Efficacy
    • 4: Further Reading
    • 5: Discussion
  • Data Visualization
    • 1: Introduction
    • 2: Getting Started
    • 3: For Analysis and Understanding
    • 4: For Communication and Storytelling
    • 5: Visualizations and Accountability
    • 6: Recommended Reading/Viewing
    • 7: Discussion
  • Discussion

2: A Brief History

Early forms of text analysis relied on paper technology and were incredibly labor intensive. While there were examples of Parisian monks creating concordance lists for the Vulgate bible in the thirteenth century, a more modern forerunner was Father Roberto Busa’s Index Thomisticus, which started in the 1940s as an attempt to analyze the complete works of Thomas Aquinas. Busa linguistically annotated and indexed all of the words by hand on index cards, which were later converted to analogue form and published in the 1970s.

The quantitative turn in history of the 1960s and 1970s also opened the field to more macro-level analysis . Historians began counting and measuring, from accumulating probate records of New England towns, for instance, to doing bigger-picture analysis of census data. But with the new wave of quantitative analysis emerged a strong backlash, as many humanities scholars critiqued the social scientific methods for what they saw as a failure to draw humanistic meaning out of data (see, for example, the controversial arguments concerning the profitability of slavery in Stanley Engerman and Robert Fogel’s book Time on the Cross).

In the realm of literature, forms of text mining have recently been characterized more broadly as a form of “distant reading”, a term coined by Franco Moretti to describe a macro-level analysis of literature. Distant reading allows analysis of a larger sample size, which can complement a more traditional analysis relying on a close reading of a small number of texts. While traditional methods often involve extrapolation based on a small sample size, mining a broader swath of data offers an opportunity to find wider larger, more representative patterns. Moretti argues that a broader scope allows scholars to analyze the “great unread” of non-canonical literature so often ignored by literary critics.

1: The Text Deluge 3: Stylometry

Navigation

  • Welcome
  • Workshop Series
  • About
  • Virtual You
  • Digitization
  • Text Analysis
    • 1: The Text Deluge
    • 2: A Brief History
    • 3: Stylometry
    • 4: Content-Based Analysis
    • 5: Metadata Analysis
    • 6: Conclusion
    • 7: Further Reading
    • 8: Discussion
  • Spatial Analysis
  • Databases
  • Pedagogy
  • Data Visualization
  • Discussion
Powered by WordPress | “Blend” from Spectacu.la WP Themes Club