SyntaxHighlighter

Showing posts with label collaboration. Show all posts
Showing posts with label collaboration. Show all posts

Saturday, April 9, 2016

Stable Orbits or Clear Air Turbulence: Capacity, Scale, and Use Cases in Geospatial Antiquity


I delivered the following talk on 8 April 2016 at the Mapping the Past: GIS Approaches to Ancient History conference at the University of North Carolina at Chapel Hill. Update (19 April 2016): video is now available on YouTube, courtesy of the Ancient World Mapping Center.

How many of you are familiar with Jo Guldi's on-line essay on the "Spatial Turn" in western scholarship? I highly recommend it. It was published in 2011 as a framing narrative for the Spatial Humanities website, a publication of the Scholar's Lab at the University of Virginia. The website was intended partly to serve as a record of the NEH-funded Institute for Enabling Geospatial Scholarship. That Institute, organized in a series of three thematic sessions, was hosted by the Scholars Lab in 2009 and 2010. The essay begins as follows:
“Landscape turns” and “spatial turns” are referred to throughout the academic disciplines, often with reference to GIS and the neogeography revolution ... By “turning” we propose a backwards glance at the reasons why travelers from so many disciplines came to be here, fixated upon landscape, together. For the broader questions of landscape – worldview, palimpsest, the commons and community, panopticism and territoriality — are older than GIS, their stories rooted in the foundations of the modern disciplines. These terms have their origin in a historic conversation about land use and agency.
Professor Guldi's essay takes us on a tour through the halls of the Academy, making stops in a variety of departments, including Anthropology, Literature, Sociology, and History. She traces the intellectual innovations and responses -- prompted in no small part by the study and critique of the modern nation state -- that iteratively gave rise to many of the research questions and methods that concern us at this conference. I don't think it would be a stretch to say that not only this conference but its direct antecedents and siblings -- the Ancient World Mapping Center and its projects, the Barrington Atlas and its inheritors -- are all symptoms of the spatial turn.

So what's the point of my talk this evening? Frankly, I want to ask: to what degree do we know what we're doing? I mean, for example, is spatial practice a subfield? Is it a methodology?  It clearly spans chairs in the Academy. But does it answer -- better or uniquely? -- a particular kind of research question? Is spatial inquiry a standard competency in the humanities, or should it remain the domain of specialists? Does it inform or demand a specialized pedagogy? Within ancient studies in particular, have we placed spatially informed scholarship into a stable orbit that we can describe and maintain, or are we still bumping and bouncing around in an unruly atmosphere, trying to decide whether and where to land?

Some will recognize in this framework questions -- or should we say anxieties -- that are also very much alive for the digital humanities. The two domains are not disjoint. Spatial analysis and visualization are core DH activities. The fact that the Scholar's Lab proposed and the NEH Office of Digital Humanities funded the Geospatial Institute I mentioned earlier underscore this point.

So, when it comes to spatial analysis and visualization, what are our primary objects of interest? "Location" has to be listed as number one, right? Location, and relative location, are important because they are variables in almost every equation we could care about. Humans are physical beings, and almost all of our technology and interaction -- even in the digital age -- are both enabled and constrained by physical factors that vary not only in time, but also in three-dimensional space. If we can locate people, places, and things in space -- absolutely or relatively -- then we can open our spatial toolkit. Our opportunities to explore become even richer when we can access the way ancient people located themselves, each other, places, and things in space: the rhetoric and language they used to describe and depict those locations.

The connections between places and between places and other things are also important. The related things can be of any imaginable type: objects, dates, events, people, themes. We can express and investigate these relationships with a variety of spatial and non-spatial information structures: directed graphs and networks for example. There are digital tools and methods at our disposal for working with these mental constructs too, and we'll touch on a couple of examples in a minute. But I'd like the research questions, rather than the methods, to lead the discussion.

When looking at both built and exploited natural landscapes, we are often interested in the functions humans impart to space and place. These observations apply not only to physical environments, but also to their descriptions in literature and their depictions in art and cartography. And so spatial function is also about spatial rhetoric, performance, audience, and reception.

Allow me a brief example: the sanctuary of Artemis Limnatis at Volimnos in the Tayegetos mountains (cf. Koursoumis 2014; Elliott 2004, 74-79 no. 10). Its location is demonstrated today only by scattered architectural, artistic, and epigraphic remains, but epigraphic and literary testimony make it clear that it was just one of several such sanctuaries that operated at various periods and places in the Peloponnese.  Was this ancient place of worship located in a beautiful spot, evocative of the divine? Surely it was! But it -- and its homonymous siblings -- also existed to claim, mark, guard, consecrate, and celebrate political and economic assertions about the land it overlooked. Consequently, the sanctuary was a locus of civic pride for the Messenians and the Spartans, such that -- from legendary times down to at least the reign of Vespasian -- it occasioned both bloodshed and elite competition for the favor of imperial powers. Given the goddess's epithet (she is Artemis Of The Borders), the sanctuary's location, and its history of contentiousness, I don't think we're surprised that a writer like Tacitus should take notice of delegations from both sides arriving in Rome to argue for and against the most recent outcome in the struggle for control of the sanctuary. I can't help but imagine him smirking as he drops it into the text of his Annals (4.43), entirely in indirect discourse and deliberately ambiguous of course about whether the delegations appeared before the emperor or the Senate. It must have given him a grim sort of satisfaction to be able to record a notable interaction between Greece and Rome during the reign of Tiberius that also served as a metaphor for the estrangement of emperor and senate, of new power and old prerogatives.

Epigraphic and literary analysis can give us insight into issues of spatial function, and so can computational methods. The two approaches are complementary, sometimes informing, supporting, and extending each other, other times filling in gaps the other method leaves open. Let's spend some time looking more closely at the computational aspects of spatial scholarship.

A couple of weeks ago, I got to spend some time talking to Lisa Mignone at Brown about her innovative work on the visibility of temples at Rome with respect to the valley of the Tiber and the approaches to the city. Can anyone doubt that, among the factors at play in the ancient siting and subsequent experience of such major structures, there's a visual expression of power and control at work? Mutatis mutandis, you can feel something like it today if you get the chance to walk the Tiber at length. Or, even if you just go out and contemplate the sight lines to the monuments and buildings of McCorkle Place here on the UNC campus. To be sure, in any such analysis there is a major role for the mind of the researcher ... in interpretation, evaluation, narration, and argument, and that researcher will need to be informed as much as possible by the history, archaeology, and literature of the place. But, depending on scale and the alterations that a landscape has undergone over time, there is also the essential place of viewshed analysis. Viewsheds are determined by assessing the visibility of every point in an area from a particular point of interest. Can I see the University arboretum from the north-facing windows of the Ancient World Mapping Center on the 5th floor of Davis Library? Yes, the arboretum is in the Center's viewshed. Well, certain parts of it anyway. Can I see the Pit from there? No. Mercifully, the Pit is not in the Center's viewshed.

In one methodological respect, Professor Mignone's work is not new. Viewshed analysis has been widely used for years in archaeological and historical study, at levels ranging from the house to the public square to the civic territory and beyond. I doubt anyone could enumerate all the published studies without a massive amount of bibliographical work. Perhaps the most well known -- if you'll permit an excursion outside the domain of ancient studies -- is Anne Kelly Knowles' work (with multiple collaborators) on the Battle of Gettysburg. What could the commanders see and when could they see it? There's a fascinating, interactive treatment of the data and its implications published on the website of Smithsonian Magazine.

Off the top of my head, I can point to a couple of other examples in ancient studies. Though their mention will only scratch the surface of the full body of work, I think they are both useful examples. There's Andrew Sherrat's 2004 treatment of Myceneae, which explores the site's visual and topographical advantages in an accessible, online form. It makes use of cartographic illustration and accessible text to make its points about strategically and economically interesting features of the site.

I also recall a poster by James Newhard and several collaborators that was presented at the 2012 meeting of the Archaeological Institute of America. It reported on the use of viewshed analysis and other methods as part of an integrated approach to identifying Byzantine defensive systems in North Central Anatolia. The idea here was that the presence of a certain kind of viewshed -- one offering an advantage for surveillance of strategically useful landscape elements like passes and valleys -- might lend credance to the identification of ambiguous archaeological remains as fortifications. Viewshed analysis is not just revelatory, but can also be used for predictive and taxonomic tasks.

In our very own conference, we'll hear from Morgan Di Rodi and Maria Kopsachelli about their use of viewshed analysis and other techniques to refine understanding of multiple archaeological sites in northwest Greece. So we'll get to see viewsheds in action!

Like most forms of computational spatial analysis, viewshed work is most rigorously and uniformly accomplished with GIS software, supplied with appropriately scaled location and elevation data. To do it reliably by hand for most interesting cases would be impossible. These dependencies on software and data, and the know-how to use them effectively, should draw our attention to some important facts. First of all, assembling the prerequisites of non-trivial spatial analysis is challenging and time consuming. More than once, I've heard Prof. Knowles say that something like ninety percent of the time and effort in a historical GIS project goes into data collection and preparation.  Just as we depend on the labor of librarians, editors, philologists, Renaissance humanists, medieval copyists, and their allies for our ability to leverage the ancient literary tradition for scholarly work, so too we depend on the labor of mathematicians, geographers, programmers, surveyors, and their allies for the data and computational artifice we need to conduct viewshed analysis. This inescapable debt -- or, if you prefer, this vast interdisciplinary investment in our work -- is a topic to which I'd like to return at the end of the talk.

Before we turn our gaze to other methods, I'd like to talk briefly about other kinds of sheds. Watershed analysis -- the business of calculating the entire area drained and supplied by a particular water system -- is a well established method of physical geography and the inspiration for the name viewshed. It has value for cultural, economic, and historical study too, and so should remain on our spatial RADAR. In fact, Melissa Huber's talk on the Roman water infrastructure under Claudius will showcase this very method.

Among Sarah Bond's current research ideas is a "smells" map of ancient Rome. Where in the streets of ancient Rome would you have encountered the odors of a bakery or a latrine a fullonica. And -- God help you -- what would it have smelled like? Will it be possible at some point to integrate airflow and prevailing wind models with urban topography and location data to calculate "smellsheds" or "nosescapes" for particular installations and industries? I sure hope so! Sound sheds ought to be another interesting possibility; we ought to look for leadership to the work of people like Jeff Veitch who is investigating acoustics and architecture at Ostia, and the Virtual Paul's Cross project at North Carolina State.

Every bit as interesting as what the ancients could see, and from where they could see it, is the question of how they saw things in space and how they described them. Our curiosity about ancient geographic mental models and worldview drives us to ask questions like ones Richard Talbert has been asking: did the people living in a Roman province think of themselves as "of the province" in the way modern Americans think of themselves as being North Carolinians or Michiganders? Were the Roman provinces strictly administrative in nature, or did they contribute to personal or corporate identity in some way? Though not a field that has to be plowed only with computers, questions of ancient worldview do sometimes yield to computational approaches.

Consider, for example, the work of Elton Barker and colleagues under the rubric of the Hestia project. Here's how they describe it:
Using a digital text of Herodotus’s Histories, Hestia uses web-mapping technologies such as GIS, Google Earth and the Narrative TimeMap to investigate the cultural geography of the ancient world through the eyes of one of its first witnesses. 
In Hestia, word collocation -- a mainstay of computational text analysis -- is brought together and integrated with location-based measures to interrogate not only the spatial proximity of places mentioned by Herodotus, but also the textual proximities of those place references. With these keys, the Hestia team opens the door to Herodotus' geomind and that of the culture he lived in: what combinations of actual location, historical events, cultural assumptions, and literary agenda shape the mention of places in his narrative?

Hestia is not alone in exploring this particular frontier. Tomorrow we'll hear from Ryan Horne about his collaborative work on the Big Ancient Mediterranean project. Among its pioneering aspects is the incorporation of data about more than the collocation of placenames in primary sources and the relationships of the referenced places with each other. BAM also scrutinizes personal names and the historical persons to whom they refer. Who is mentioned with whom where? What can we learn from exploring the networks of connection that radiate from such intersections?

The introduction of a temporal axis into geospatial calcuation and visualization is also usually necessary and instructive in spatial ancient studies, even if it still proves to be more challenging in standard GIS software than one might like. Amanda Coles has taken on that challenge, and will be telling us more about what it's helped her learn about the interplay between warfare, colonial foundations, road building, and the Roman elites during the Republic.

Viewsheds, worldviews, and temporality, oh my!

How about spatial economies? How close were sources of production to their markets? How close in terms of distance? How close in terms of travel time? How close in terms of cost to move goods?

Maybe we are interested in urban logistics. How quickly could you empty the colosseum? How much bread could you distribute to how many people in a particular amount of time at a particular place? What were the constraints and capacities for transport of the raw materials? What do the answers to such questions reveal about the practicality, ubiquity, purpose, social reach, and costs of communal activities in the public space? How do these conclusions compare with the experiences and critiques voiced in ancient sources?

How long would it take a legion to move from one place to another in a particular landscape? What happens when we compare the effects of landscape on travel time with the built architecture of the limes or the information we can glean about unit deployment patterns from military documents like the Vindolanda tablets or the ostraca from Bu Njem?

The computational methods involved in these sorts of investigations have wonderful names, and like the others we've discussed, require spatial algorithms realized in specialized software. Consider cost surfaces: for a particular unit of area on the ground, what is the cost in time or effort to pass through it? Consider network cost models: for specific paths between nodal points, what is the cost of transit? Consider least cost path analysis: given a cost surface or network model, what is the cheapest path available between two points?

Many classicists will have used Orbis: The Stanford Geospatial Network Model of the Roman World. The Orbis team, assembled by Walter Scheidel, has produced an online environment in which one can query a network model of travel costs between key nodal points in the Roman world, varying such parameters as time of year and mode of transport. This model, and its digital modes of access, bring us to another vantage point. How close were two places in the Roman world, not as the crow flies, not in terms of miles along the road, but as the boat sailed or the feet walked.

Barbora Weissova is going to talk to us tomorrow about her work in and around Nicaea. Among her results, she will discuss another application of Least Cost Path Analysis: predicting the most likely route for a lost ancient roadway.

It's not just about travel, transport, and cost. Distribution patterns are of interest too, often combined with ceramic analysis, or various forms of isotopic or metallurgical testing, to assess the origin, dissemination, and implications of ancient objects found in the landscape. Inscriptions, coins, portable antiquities, architectural and artistic styles, pottery, all have been used in such studies. Corey Ellithorpe is going to give us a taste of this approach in numismatics by unpacking the relationship between Roman imperial ideology and regional distribution patterns of coins.

I'd like to pause here for just a moment and express my hope that you'll agree with the following assessment. I think we are in for an intellectual feast tomorrow. I think we should congratulate the organizers of the conference for such an interesting, and representative, array of papers and presentations. That there is on offer such a tempting smorgasbord is also, of course, all to the credit of the presenters and their collaborators. And surely it must be a factor as we consider the ubiquity and disciplinarity of spatial applications in ancient studies.

Assiduous students of the conference program will notice that I have neglected yet to mention a couple of the papers. Fear not, for they feature in the next section of my talk, which is -- to borrow a phrase from Meghan Trainor and Kevin Kadish -- all about that data.

So, conference presenters, would you agree with the dictum I've attributed to Anne Knowles? Does data collection and preparation take up a huge chunk of your time?

Spatial data, particularly spatial data for ancient studies, doesn't normally grow on trees, come in a jar, or sit on a shelf. The ingredients have to be gathered and cleaned, combined and cooked. And then you have to take care of it, transport it, keep track of it, and refashion it to fit your software and your questions. Sometimes you have to start over, hunt down additional ingredients, or try a new recipe. This sort of iterative work -- the cyclic remaking of the experimental apparatus and materials -- is absolutely fundamental to spatially informed research in ancient studies.

If you were hoping I'd grind an axe somewhere in this talk, you're in luck. It's axe grinding time.

There is absolutely no question in my mind that the collection and curation of data is part and parcel of research. It is a research activity. It has research outcomes. You can't answer questions without it. If you aren't surfacing your work on data curation in your CV, or if you're discounting someone else's work on data curation in decisions about hiring, tenure, and promotion, then I've got an old Bob Dylan song I'd like to play for you.

  • Archive and publish your datasets. 
  • Treat them as publications in your CV. 
  • Write a review of someone else's published dataset and try to get it published. 
  • Document your data curation process in articles and conference presentations.

Right. Axes down.

So, where does our data come from? Sometimes we can get some of it in prepared form, even if subsequent selection and reformatting is required. For some areas and scales, modern topography and elevation can be had in various raster and vector formats. Some specialized datasets exist that can be used as a springboard for some tasks. It's here that the Pleiades project, which I direct, seeks to contribute. By digitizing not the maps from the Barrington Atlas, but the places and placenames referenced on those maps and in the map-by-map directory, we created a digital dataset with potential for wide reuse. By wrapping it in a software framework that facilitates display, basic cartographic visualization, and collaborative updates, we broke out of the constraints of scale and cartographic economy imposed by the paper atlas format. Pleiades now knows many more places than the Barrington did, most of these outside the cultures with which the Atlas was concerned. More precise coordinates are coming in too, as are more placename variants and bibliography. All of this data is built for reuse. You can collect it piece by piece from the web interface or download it in a number of formats. You can even write programs to talk directly to Pleiades for you, requesting and receiving data in a computationally actionable form. The AWMC has data for reuse too, including historical coastlines and rivers and map base materials. It's all downloadable in GIS-friendly formats.

But Pleiades and the AWMC only help for some things. It's telling that only a couple of the projects represented at this conference made use of Pleiades data. That's not because Pleiades is bad or because the authors didn't know about Pleiades or the Center. It's because the questions they're asking require data that Pleiades is not designed to provide.

It's proof of the point I'd like to underline: usually -- because your research question is unique in some way, otherwise you wouldn't be pursuing it -- you're going to have to get your hands dirty with data collection.

But before we get dirty, I'm obliged to point out that, although Pleiades has received significant, periodic support from the National Endowment for the Humanities since 2006, the views, findings, conclusions, or recommendations expressed in this lecture do not necessarily reflect those of the National Endowment for the Humanities.

We've already touched on the presence of spatial language in literature. For some studies, the placenames, placeful descriptions, and narratives of space found in both primary and secondary sources constitute raw data we'd like to use. Identifying and extracting such data is usually a non-trivial task, and may involve a combination of manual and computational techniques, the latter depending on the size and tractability of the textual corpus in question and drawing on established methods in natural language processing and named entity recognition. It's here we may encounter "geoparsing" as a term of art. Many digital textual projects and collections are doing geoparsing: individual epigraphic and papyrological publications using the Text Encoding Initiative and EpiDoc Guidelines; the Perseus Digital Library; the Pelagios Commons by way of its Recogito platform. The China Historical GIS is built up entirely from textual sources, tracking each placename and each assertion of administrative hierarchy back to its testimony.

For your project, you may be able to find geoparsed digital texts that serve your needs, or you may need to do the work yourself. Either way, some transformation on the results of geoparsing is likely to be necessary to make them useful in the context of your research question and associated apparatus.

Relevant here is Micah Myers's conference paper. He is going to bring together for us the analysis and visualization of travel as narrated in literature. I gather from his abstract that he'll show us not only a case study of the process, but discuss the inner workings of the on-line publication that has been developed to disseminate the work.

Geophysical and archaeological survey may supply your needs. Perhaps you'll have to do fieldwork yourself, or perhaps you can collect the information you need from prior publications or get access to archival records and excavation databases. Maybe you'll get lucky and find a dataset that's been published into OpenContext, the British Archaeology Data Service, or tDAR: the Digital Archaeological Record. But using this data requires constant vigilance, especially when it was collected for some purpose other than you own. What were the sampling criteria? What sorts of material were intentionally ignored? What circumstances attended collection and post-processing?

Sometimes the location data we need comes not from a single survey or excavation, but from a large number of possibly heterogeneous sources. This will be the case for many spatial studies that involve small finds, inscriptions, coins, and the like. Fortunately, many of the major documentary, numismatic, and archaeological databases are working toward the inclusion of uniform geographic information in their database records. This development, which exploits the unique identifying numbers that Pleiades bestows on each ancient place, was first championed by Leif Isaksen, Elton Barker, and Rainer Simon of the Pelagios Commons project. If you get data from a project like the Heidelberg Epigraphic Databank, papyri.info, the Arachne database of the German Archaeological Institute, the Online Coins of the Roman Empire, or the Perseus Digital Library, you can count on being able to join it easily with Pleiades data and that of other Pelagios partners. Hopefully this will save some of us some time in days to come.

Sometimes what's important from a prior survey will come to us primarily through maps and plans. Historical maps may also carry information we'd like to extract and interpret. There's a whole raft of techniques associated with the scanning, georegistration, and georectification (or warping) of maps so that they can be layered and subjected to feature tracing (or extraction) in GIS software. Some historic cartofacts -- one thinks of the Peutinger map and medieval mappae mundi as examples -- are so out of step with our expectations of cartesian uniformity that these techniques don't work. Recourse in such cases may be had to first digitizing features of interest in the cartesian plane of the image itself, assigning spatial locations to features later on the basis of other data. Digitizing and vectorizing plans and maps resulting from multiple excavations in order to build up a comprehensive archaeological map of a region or site also necessitates not only the use of GIS software but the application of careful data management practices for handling and preserving a collection of digital files that can quickly grow huge.

We'll get insight into just such an effort tomorrow when Tim Shea reports on Duke's "Digital Athens Project".

Let's not forget remote sensing! In RS we use sensors -- devices that gather information in various sections of the electro-magnetic spectrum or that detect change in local physical phenomena. We mount these sensors on platforms that let us take whatever point of view is necessary to achieve the resolution, scale, and scope of interest: satellites, airplanes, drones, balloons, wagons, sleds, boats, human hands. The sensors capture emitted and reflected light in the visible, infrared, and ultraviolet wavelengths or magnetic or electrical fields. They emit and detect the return of laser light, radio frequency energy, microwaves, millimeter waves, and, especially underwater, sound waves. Specialized software is used to analyze and convert such data for various purposes, often into rasterized intensity or distance values that can be visualized by assigning brightness and color scales to the values in the raster grid. Multiple images are often mosaicked together to form continuous images of a landscape or 360 degree seamless panoramas.

Remotely sensed data facilitate the detection and interpretation of landforms, vegetation patterns, and physical change over time, revealing or enhancing understanding of built structures and exploited landscapes, as well as their conservation. This is the sort of work that Sarah Parcak has been popularizing, but it too has decades of practice behind it. In 1990, Tom Sever's dissertation reported on a remote-sensing analysis of the Anasazi road system, revealing a component of the built landscape that was not only invisible on the ground, but that demonstrates that the Anasazi were far more willing than even the Romans to create arrow-straight roads in defiance of topographical impediments. More recently, Prof. Sever and his NASA colleague Daniel Irwin have been using RS data for parts of Guatemala, Honduras, and Mexico, to distinguish between vegetation that thrives in alkaline soils and vegetation that doesn't. Because of the Mayan penchant for coating monumental structures with significant quantities of lime plaster, this data has proved remarkably effective in the locating of previously unknown structures beneath forest canopy. The results seem likely to overturn prevailing estimates of the extent of Mayan urbanism, illustrating a landscape far more cleared and built upon than heretofore proposed (cf. Sever 2003).

Given the passion with which I've already spoken about the care and feeding of data, you'll be unsurprised to learn that I'm really looking forward to Nevio Danelon's presentation tomorrow on the capture and curation of remotely sensed data in a digital workflow management system designed to support visualization processes.

I think it's worth noting that both Professor Parcak's recently collaborative work on a possible Viking settlement in Newfoundland, as well as Prof. Sever's dissertation, represent a certain standard in the application of remote sensing to archaeology. RS analysis is tried or adopted for most archaeological survey and excavation undertaken today. The choice of sensors, platforms, and analytical methods will of course vary in response to landscape conditions, expected archaeological remains, and the realities of budget, time, and know-how.

Similarly common, I think, in archaeological projects is the consideration of geophysical, alluvial, and climatic features and changes in the study area. The data supporting such considerations will come from the kinds of sources we've already discussed, and will have to be managed in appropriate ways. But it's in this area -- ancient climate and landscape change -- that I think ancient studies has a major deficit in both procedure and data. Computational, predictive modeling of ancient climate, landscape, and ground cover has made no more than tentative and patchy inroads on the way we think about and map the ancient historical landscape. That's a deficit that needs addressing in an interdisciplinary and more comprehensive way.

I'd be remiss if, before moving on to conclusions, I kept the focus so narrowly on research questions and methods that we miss the opportunity to talk about pedagogy, public engagement, outreach, and cultural heritage preservation. Spatial practice in the humanities is increasingly deeply involved in such areas. The Ancient World Mapping Center's Antiquity a-la Carte website enables users to create and refine custom maps from Pleiades and other data that can then be cited, downloaded, and reused. It facilitates the creation of map tests, student projects, and maps to accompany conference presentations and paper submissions.

Meanwhile, governments, NGOs, and academics alike are brining the full spectrum of spatial methods to bear as they try to prevent damage to cultural heritage sites through assessment, awareness, and intervention. The American Schools of Oriental Research conducts damage assessments and site monitoring with funding in part from the US State Department. The U.S. Committee of the Blue Shield works with academics to prepare geospatial datasets that are offered to the Department of Defense to enhance compliance with the 1954 Hague Convention for the Protection of Cultural Property in the Event of Armed Conflict.

These are critical undertakings as well, and should be considered an integral part of our spatial antiquity practice.

So, how should we gather up the threads of this discussion so we can move on to the more substantive parts of the conference?

I'd like to conclude as I began, by recommending an essay. In this case, I'm thinking of Bethany Nowiskie's recent essay on "capacity and care" in the digital humanities. Bethany is the former director of UVA's Scholars Lab. She now serves as Director of the Digital Library Federation at the Council on Library and Information Resources. I had the great good fortune to hear Bethany deliver a version of this essay as a keynote talk at a project director's meeting hosted by the NEH Office of Digital Humanities in Washington in September of last year. You can find the essay version on her personal website.

Bethany thinks the Humanities must expand its capacity in order not only to survive the 21st century, but to contribute usefully to its grand challenges. To cope with increasing amounts and needs for data of every kind. To move gracefully in analysis and regard from large scales to small ones and to connect analysis at both levels. To address audiences and serve students in an expanding array of modes. To collaborate across disciplines and heal the structurally weakening divisions that exist between faculty and "alternate academics", even as the entire edifice of faculty promotion and tenure threatens to shatter around us.

What is Bethany's prescription? An ethic of care. She defines an ethic of care as "a set of practices", borrowing the following quotation from the political scientist Joan Tronto:
a species of [collective] activity that includes everything we do to maintain, contain, and repair our world, so that we can live in it as well as possible.
I think our practice of spatial humanities in ancient studies is just such a collective activity. We don't have to turn around much to know that we are cradled in the arms and buoyed up on the shoulders of a vast cohort, stretching back through time and out across the globe. Creating data and handing it on. Debugging and optimizing algorithms. Critiquing ideas and sharpening analytical tools.

The vast majority of projects on the conference schedule, or that I could think of to mention in my talk, are explicitly and immediately collaborative.

And we can look around this room and see like-minded colleagues galore. Mentors. Helpers. Friends. Comforters. Makers. Guardians.

And we have been building the community infrastructure we need to carry on caring about each other and about the work we do to explain the human past to the human present and to preserve that understanding for the human future. We have centers and conferences and special interest groups and training sessions. We involve undergraduates in research and work with interested people from outside the academy. We have increasingly useful datasets and increasingly interconnected information systems. Will all these things persist? No, but we get to play a big role in deciding what and when and why.

So if there's a stable orbit to be found, I think it's in continuing to work together and to do so mindfully, acknowledging our debts to each other and repaying them in kind.

I'm reminded of a conversation I had with Scott Madry, back in the early aughts when we were just getting the Mapping Center rolling and Pleiades was just an idea. As many of you know, Scott together with Carole Crumley and numerous other collaborators here at UNC and beyond, have been running a multidimensional research project in Burgundy since the 1970s. At one time or another the Burgundy Historical Landscapes project has conducted most of the kinds of studies I've mentioned tonight, all the while husbanding a vast and growing store of spatial and other data across a daunting array of systems and formats.

I think that the conversation I'm remembering with Scott took place after he'd spent a couple of hours teaching undergraduate students in my seminar on Roman roads and land travel how to do orthophoto analysis the old fashioned way: with stereo prints and stereoscopes. He was having them do the Sarah Parcak thing: looking for crop marks and other indications of potentially buried physical culture. After the students had gone, Scott and I were commiserating about the challenges of maintaining and funding long-running research projects. I was sympathetic, but know now that I really didn't understand those challenges then. Scott did, and I remember what he said. He said: "We were standing on that hill in Burgundy twenty years ago, and as we looked around I said to Carol: 'somehow, we are going to figure out what happened here, no matter how long it takes.'"

That's what I'm talking about.

Thursday, April 10, 2014

Batch XML validation at the command line

Updated: 8 August, 2017 to reflect changes in the installation pattern for jing.

Against a RelaxNG schema. I had help figuring this out from Hugh and Ryan at DC3:

$ find {searchpath} -name "*.xml" -print | parallel --tag jing {relaxngpath}
The find command hunts down all files ending with ".xml" in the directory tree under searchpath. The parallel command takes that list of files and fires off (in parallel) a jing validation run for each of them. The --tag option passed to jing ensures we get the name of the file passed through with each error message. This turns out (in general terms as seen by me) to be much faster than running each jing call in sequence, e.g. with the --exec primary in find.

As I'm running on a Mac, I had to install GNU Parallel and the Jing RelaxNG Validator. That's what Homebrew is for:
$ brew install jing
$ brew install jing-trang
$ brew install parallel
NB: you may have to install a down version of Java before you can get the jing-trang formula to work in homebrew (e.g., brew install java6).

What's the context, you ask? I have lots of reasons to want to be able to do this. The proximal cause was batch-validating all the EpiDoc XML files for the inscriptions that are included in the Corpus of Campā Inscriptions before regenerating the site for an update today. I wanted to see quickly if there were any encoding errors in the XML that might blow up the XSL transforms we use to generate the site. So, what I actually ran was:
$ curl -O http://www.stoa.org/epidoc/schema/latest/tei-epidoc.rng
$ find ./texts/xml -name '*.xml' -print | parallel --tag jing tei-epidoc.rng
 Thanks to everybody who built all these tools!


Thursday, June 20, 2013

It is happening

A couple of hours ago, I was sitting out on the back deck with my wife and pets, enjoying perfect temperatures, morning birdsong, lavender-scented country air, and a cup of freshly brewed Costa Rican coffee (roasted by the good folks at the Kaffeeklatsch in Huntsville). Idyllic.

I was flipping through the latest news stories, blog posts, and such, brought into my phone by my feed reader (currently Feedly). I was trying to ignore the omnipresent bad news of the world, when this popped up:

screen capture of a feed summary in Feedly on my Android phone
Forma[m] Lollianus fecit?!? I'm predisposed by my dissertation experience to trigger on certain Latin and Greek words because of their sometime significance for the study of Roman geography. Forma is of course one of those words, and it does (probably more often than justified) get translated as "map" or "plan." Could this be — admittedly against the odds —an inscription on a map or plan drafted or surveyed by some guy named Lollianus?

If you're me, the possibility warrants a click-through to a corresponding record in the Heidelberg Epigraphic Databank (EDH).

My mappish hopes were quickly dashed, but just as quickly were replaced by interest in a group of inscribed objects I hadn't run across before: mirrors from Roman Dacia bearing makers' inscriptions. "Forma" can mean "mirror"? A quick check of Lewis & Short at Perseus doesn't support that idea, but builds confidence in a better interpretation: "mold, stamp, form". Was this mirror, or some part of it, somehow cast or stamped out? The EDH entry tells me there are 9 identical mirrors extant and that the inscription goes around the "Fassung" (frame?). Yup.

Cool. I learned something today before breakfast. And it's knowledge I can use when I come back to doing more with the geographical/mapping/surveying vocabulary.

And then it hits me: that's not information I went looking for, not a search I initiated. New information of interest was pushed to me because I had previously used a software tool to express interest in a number of information sources including, but not limited to, ancient inscriptions. The software kept an eye on new output from those sources and made it available to me for review and engagement in a mode and at a time and place of my choosing. And because the source data was online, open, and linked in a standard format, I was able to drink coffee and pet my dog on the back deck in Moontown, Alabama while making use of the scholarly work done yesterday(!) by Brigitte Gräf in Heidelberg, Germany.

Isn't this one of the things we've been working toward?

How did that happen?


Sometime earlier this year, Frank Grieshaber in Heidelberg rolled out web page listings and corresponding Atom feeds of recently changed content in the EDH (e.g., latest updates to the inscriptions database). I added them, along with similar data-oriented feeds, to a feed aggregator I dubbed Planet Potamos (with "Potamos" trying lamely to evoke a rushing river of data; the "Planet" acknowledges the feed aggregation software I use). I put the same feed subscriptions into my personal feed reader (I could also have put the Potamos aggregator's feed, but it only updates periodically and I'm an immediacy junkie). I installed and configured my feed reader on every device I use.

The rest is magic. Magic made the old-fashioned way by lots of people in many different places and times developing standards, building software, creating data, doing research, and sharing.

What next?


Well, I hope that Frank and his colleagues in Heidelberg will eventually add thumbnail images (where they have them) to the EDH feeds. I hope that the other epigraphic databases (and indeed all kinds of ancient studies web applications) will set up similar feeds. I hope that we can all start using more linked-data approaches in and alongside such feeds in order to communicate seminal interpretive/discovery facets (like geography, personography, temporality and genre) in machine-actionable ways. I hope the spirit and practice of openness that lubricates and accelerates this sort of synergy continues to grow and flower.

As for me, I'm thinking about how I might set up some kind of filtering mechanism that would highlight or prioritize content in my feed reader that's potentially relevant to my (e.g.) geo/map/survey vocabulary interests. Hmmmmm....


Thursday, April 18, 2013

Citing Sources in Digital Annotations

I'm collaborating with other folks both in and outside ISAW on a variety of digital scholarly projects in which Linked Open Data is playing a big role. We're using the Resource Description Framework (RDF) to provide descriptive information for, and make cross-project assertions about, a variety of entities of interest and the data associated with them (places, people, themes/subjects, creative works, bibliographic items, and manuscripts and other text-bearing objects). So, for example, I can produce the following assertions in RDF (using the Terse RDF Triple Language, or TuRTLe):

<http://syriaca.org/place/45> a <http://geovocab.org/spatial#Feature> ;
  rdfs:label "Serugh" ;
  rdfs:comment "An ancient city where Jacob of Serugh was bishop."@en ;
  foaf:primaryTopicOf <http://en.wikipedia.org/wiki/Suruç> ;
  owl:sameAs <http://pleiades.stoa.org/places/658405#this> .

This means: 'There's a resource identified with the Uniform Resource Identifier (URI) "http://syriaca.org/place/45" about which the following is asserted:
(Folks familiar with what Sean Gillies has done for the Pleiades RDF will recognize my debt to him in the what proceeds.)

But there are plenty of cases in which just issuing a couple of triples to encode an assertion about something isn't sufficient; we need to be able to assign responsibility/origin for those assertions and to link them to supporting argument and evidence (i.e., standard scholarly citation practice). For this purpose, we're very pleased by the Open Annotation Collaboration, whose Open Annotation Data Model was recently updated and expanded in the form of a W3C Community Draft (8 February 2013) (the participants in Pelagios use basic OAC annotations to assert geographic relationships between their data and Pleiades places).


A basic OADM annotation uses a series of RDF triples to link together a "target" (the thing you want to make an assertion about) and a "body" (the content of your assertion). You can think of them as footnotes. The "target" is the range of text after which you put your footnote number (only in OADM you can add a footnote to any real, conceptual, or digital thing you can identify) and the "body" is the content of the footnote itself. The OADM draft formally explains this structure in section 2.1. This lets me add an annotation to the resource from our example above (the ancient city of Serugh) by using the URI "http://syriaca.org/place/45" as the target of an annotation) thus:
<http://syriaca.org/place/45/anno/desc6> a oa:Annotation ;
  oa:hasBody <http://syriaca.org/place/45/anno/desc6/body> ;
  oa:hasTarget <http://syriaca.org/place/45> ;
  oa:motivatedBy oa:describing ;
  oa:annotatedBy <http://syriaca.org/editors.xml#tcarlson> ;
  oa:annotatedAt "2013-04-03T00:00:01Z" ;
  oa:serializedBy <https://github.com/paregorios/srpdemo1/blob/master/xsl/place2ttl.xsl> ;
  oa:serializedAt "2013-04-17T13:35:05.771-05:00" .

<http://syriaca.org/place/45/anno/desc6/body> a cnt:ContentAsText, dctypes:Text ;
  cnt:chars "an ancient town, formerly located near Sarug."@en ;
  dc:format "text/plain" ;

I hope you'll forgive me for not spelling that all out in plain text, as all the syntax and terms are explained in the OADM. What I'm concerned about in this blog post is really what the OADM doesn't explicitly tell me how to do, namely: show that the annotation body is actually a quotation from a published book. The verb oa:annotatedBy lets me indicate that the annotation itself was made (i.e., the footnote was written) by a resource identified by the URI "http://syriaca.org/editors.xml#tcarlson". If I'd given you a few more triples, you could have figured out that that resource is a real person named Thomas Carlson, who is one of the editors working on the Syriac Reference Portal project. But how do I indicate (as he very much wants to do because he's a responsible scholar and has no interest in plagiarizing anyone) that he's deliberately quoting a book called The Scattered Pearls: A History of Syriac Literature and Sciences? Here's what I came up with (using terms from Citation Typing Ontology and the DCMI Metadata Terms):
<http://syriaca.org/place/45/anno/desc7/body> a cnt:ContentAsText, dctypes:Text ;
  cnt:chars "a small town in the Mudar territory, between Ḥarran and Jarabulus. [Modern name, Suruç (tr.)]"@en ;
  dc:format "text/plain" ;
  cito:citesAsSourceDocument <http://www.worldcat.org/oclc/255043315> ;
  dcterms:biblographicCitation  "The Scattered Pearls: A History of Syriac Literature and Sciences, p. 558"@en .

The addition of the triple containing cito:citesAsSourceDocument lets me make a machine-actionable link to the additional structured bibliographic data about the book that's available at Worldcat (but it doesn't say anything about page numbers!). The addition of the triple containing dcterms:bibliographicCitation lets me provide a human-readable citation.

I'd love to have feedback on this approach from folks in the OAC, CITO, DCTERMS, and general linked data communities. Could I do better? Should I do something differently?


The SRP team is currently evaluating a sample batch of such annotations, which you're also welcome to view. The RDF can be found here. These files are generated from the TEI XML here using the XSLT here.

Friday, June 1, 2012

Ancient Studies Needs Open Bibliographic Data and Associated URIs

Update 1:  links throughout, minor formatting changes, proper Creative Commons Public Domain tools, parenthetical about import path from Endnote and such, fixing a few typos.

The NEH-funded Linked Ancient World Data Institute, still in progress at ISAW, has got me thinking about a number of things. One of them is bibliography and linked data. Here's a brain dump, intended to spark conversation and collaboration.

What We Need

  • As much bibliographic data as possible, for both primary and secondary sources (print and digital), publicly released to third parties under either a public domain declaration or an unrestrictive open license.
  • Stable HTTP URIs for every work and author included in those datasets.

Why

Bibliographic and citation collection and management are integral to every research and publication in project in ancient studies. We could save each other a lot of time, and get more substantive work done in the field, if it was simpler and easier to do. We could more easily and effectively tie together disparate work published on the web (and appearing on the web through retrospective digitization) if we had a common infrastructure and shared point of reference. There's already a lot of digital data in various hands that could support such an effort, but a good chunk of it is not out where anybody with good will and talent can get at it to improve it, build tools around it, etc.

What I Want You (and Me) To Do If You Have Bibliographic Data
  1. Release it to the world through a third party. No matter what format it's in, give a copy to someone else whose function is hosting free data on the web. Dump it into a public repository at github.com or sourceforge.net. Put it into a shared library at Zotero, Bibsonomy, Mendeley, or another bibliographic content website (most have easy upload/import paths from Endnote, and other citation management applications). Hosting a copy yourself is fine, but giving it to a third party demonstrates your bona fides, gets it out of your nifty but restrictive search engine or database, and increments your bus number.
  2. Release it under a Creative Commons Public Domain Mark or Public Domain Dedication (CC0).  Or if you can't do that, find as open a Creative Commons or similar license as you can. Don't try to control it. If there's some aspect of the data that you can't (because of rights encumberance) or don't want to (why?) give away to make the world a better place, find a quick way to extract, filter, or excerpt that aspect and get the rest out.
  3. Alert the world to your philanthropy. Blog or tweet about it. Post a link to the data on your institutional website. Above all, alert Chuck Jones and Phoebe Acheson so it gets announced via Ancient World Online and/or Ancient World Open Bibliographies.
  4. Do the same if you have other useful data, like identifiers for modern or ancient works or authors.
  5. Get in touch with me and/or anyone else to talk about the next step: setting up stable HTTP URIs corresponding to this stuff.
Who I'm Talking To

First of all, I'm talking to myself, my collaborators, and my team-mates at ISAW. I intend to eat my own dogfood.

Here are other institutions and entities I know about who have potentially useful data.
  • The Open Library : data about books is already out there and available, and there are ways to add more
  • Perseus Project : a huge, FRBR-ized collection of MODS records for Greek and Latin authors, works, and modern editions thereof.
  • Center for Hellenic Studies: identifiers for Greek and Latin authors and works
  • L'Année Philologique and its institutional partners like the American Philological Association: the big collection of analytic secondary bibliography for classics (journal articles)
  • TOCS-IN: a collaboratively collected batch of analytic secondary bibliography for classics
  • Papyri.info and its contributing project partners: TEI bibliographic records for  much of the bibliography produced for or cited by Greek and Latin papyrologists (plus other ancient language/script traditions in papyrology)
  • Gnomon Bibliographische Datenbank: masses of bibliographic data for books and articles for classics
  • Any and every university library system that has a dedicated or easily extracted set of associated catalog records. Especially any with unique collections (e.g., Cincinnati) or those with databases of analytical bibliography down to the level of articles in journals and collections.
  • Ditto any and every ancient studies digital project that has bibliographic data in a database.
Comments, Reactions, Suggestions

Welcome, encouraged, and essential. By comment here or otherwise (but not private email please!).

Friday, November 4, 2011

It's all coming together at PELAGIOS

For years (over a decade in fact) we've been dreaming and talking about linking up ancient world resources on the web along the thematic axis of geography. Pleiades was launched in no small part in pursuit of that vision. And today comes more proof -- to which many can relate -- that hard work, collaboration, and openness bears really tasty fruit.
The Perseus geospatial data now includes annotations of ancient places with Pleiades URIs. Beginning next week, the Places widget in the Perseus interface will include links to download the Pleiades annotations in OAC compliant RDF format. These links will appear for any text with place entity markup which also has places from this dataset. We are also providing a link to search on the top five most frequently mentioned of these places in the Pelagios graph explorer.
(Check out the rest of the story, which provides a screenshot of the interface changes and a step-by-step description of how the work was done).

How did this come to be possible? Here's a very much abridged history:

  • Perseus built a path-breaking, web-based digital library of resources for the study of the ancient world; released a bunch of their code and content under open licenses; and managed the geographic aspects of the content as data
  • Pleiades built on and marshaled the efforts of the Classical Atlas Project, the Digital Atlas of Roman and Medieval Civilization, and other collaborators to publish an ever-improving geographic dataset on the web under a permissive open license
  • Leif Isaksen, on behalf of the Google Ancient Places project, took that dataset, mashed it up with another open geographical dataset (GeoNames) and published the results (Pleiades+) under a public domain declaration (more openness).
  • The PELAGIOS team took Pleiades+ and started matching it with their data. Perseus is just the latest member of that team to do so, and there are more on the way.
The resulting interface enhancements Perseus is announcing today are just the latest visible example of how the web of people benefits from the creation and exploitation of the web of data, and it's all super-charged by openness.

I'm grateful to the hard-working folks, and the array of funding agencies and host institutions, whose commitment and support are making these dreams come true.

Friday, October 28, 2011

Pleiades Hack Day

I'm thinking it's time. Time we designated a particular day to give over to ganging up on Pleiades content and making it better. Looking at the low-hanging-and-annoying-fruit list, the first thing that comes to mind is improving our titles and descriptions to facilitate discovery and disambiguation. I've laid out what I see as the landscape in a document on the Pleiades site: Improving Titles and Descriptions for Prominent Places.

So what remains is to see who would be willing to devote at least a couple of hours (if not a whole day) to this enterprise and to fix a day for it. You don't need to be an expert to help with this job. Anyone interested at all in ancient geography who has basic web skills and can get on the internet at the scheduled time should be able to make a substantive contribution, whether they are a student, a scholar or an interested "layperson".

If you're interested, leave a comment on my blog and I'll send you a link to a doodle poll to do the scheduling.

Tuesday, October 25, 2011

Meetup: The Future of the Ancient World Image Bank

So, my brain has mashed up its happiness at the success of last week's Pleiades community meetup with the hanging conversations that started in response to my blog post on Pleiades machine tags in Flickr. Nate Nagy (AWIB Managing Editor) and I had a good conversation a couple of weeks ago about the future of AWIB, one that would be more open and collaborative than the current model.

I'd like to light a fire under (and refine) those ideas. I'm starting a Doodle poll to find a time for a "future of AWIB" meetup, to be held via Google+. I'll send invites to all of our contacts on Flickr and others who have expressed interest. Please post a comment here if you'd like to be part of the event.

Saturday, September 10, 2011

Feeds of Flickr Photos Depicting Pleiades Places

Some months ago, ISAW started adding Pleiades machine tags to the Ancient World Image Bank (AWIB) photos we've been uploading to Flickr. This post will explain what that means, how it might be useful to you and how you can add Pleiades machine tags to your own photos so we can find out about them.

Updated: 8:45pm EDT, 10 September 2011 (changes highlighted in orange).
Updated: 10:43am EST, 20 December 2011 (some of what's here is now superseded by recent developments; see further this new post: Pleiades, Flickr, and the Ancient World Image Bank)

Pleiades Machine Tags

Pleiades is a collaborative, open-access digital gazetteer for the ancient world. AWIB is an open-access publication that uses the Flickr photo-sharing site to publish free, reusable photos of ancient sites and artifacts. Machine tags are an extension to Flickr's basic tag-this-photo functionality that "use a special syntax to define extra information about a tag" (Aaron Straup Cope, "Ladies and Gentlemen: Machine Tags," 24 January 2007).

A Pleiades machine tag looks like this:
pleiades:place=795868
where 795868 is the stable identifier portion of a Pleiades Uniform Resource Identifier (URI).  In this example, the URI corresponding to the machine tag above is:
http://pleiades.stoa.org/places/795868
Note what's in common between the machine tag and the URI (highlighted in yellow).

What Pleiades Machine Tags Are Good For

The Flickr API makes it possible to request lists of machine-tagged photos in the RSS webfeed format. So, to get a list of all photos in Flickr that are tagged with the example machine tag above, pop this into your feed reader:
http://api.flickr.com/services/feeds/photos_public.gne?tags=pleiades%3Aplace=795868&lang=en-us&format=rss_200
The same results can be viewed in HTML in a browser by resolving the following:
http://www.flickr.com/photos/tags/pleiades:place=795868
To get a list of all photos in Flickr that are tagged with any Pleiades machine tag, try this (the API syntax supports wildcards!):
http://api.flickr.com/services/feeds/photos_public.gne?tags=pleiades%3Aplace%3D&lang=en-us&format=rss_200
The same results, viewed in HTML on the Flickr site:
http://www.flickr.com/photos/tags/pleiades:place
Feeds like these aren't just for feed readers anymore. You can add user-interface widgets to your blog or website to summarize the latest content for your readers (check out the right-hand column in this blog). You can hook up services like Networked Blogs or Twitterfeed to pass on the latest changes to your Facebook friends or Twitter followers. If you've got a web-facing numismatic database that you've already linked up with Pleiades for the mint locations, you could write custom code to pull a corresponding picture of the ancient site into your web interface (say, alongside the map you've already got).

Add Pleiades Machine Tags to Your Own Photos on Flickr

Many of you have been taking amazing photos of ancient sites and artifacts for years. Many of you have posted some of them to Flickr and shared them with great groups like Chiron, Visibile Parlare - Greek Inscriptions and Visibile Parlare - Latin Inscriptions. If you'd like these photos to appear in queries and feeds (like those described above), right alongside the photos that we're publishing via AWIB, all you have to do is add the appropriate Pleiades machine tags in Flickr. Just look up your site on Pleiades, copy the numeric ID from the URI in your browser's location bar, append it to the string "pleiades:place=" and tag your Flickr photos with it. In this way, you can help us improve findability of good photos of ancient sites and the artifacts found there for everyone on the web. Who knows ... maybe enough people will join us in this effort that we can someday get the Flickr development team to give Pleiades machine tags some extra love.

Kudos to:

Wednesday, December 1, 2010

Flavia Faustina, version 3: chi-rho, dolium, multiple editors, rationale

Ryan Baumann and Georgia Tsouvala have joined the mob!

Ryan forked my Mob Epigraphy repository on github and added markup to the EpiDoc XML file to represent the Chi-Rho and dolium(?) that appear below the inscribed text. Then he sent me a pull request. I merged his changes and pushed them back to github, and then I pushed a few more modifications to show his contribution in the EpiDoc/TEI header and to modify the stylesheets to handle whitespace and multiple editors better (and to write out an HTML doctype). Here's the result:
Ryan's change -- which parallels the treatment in ICVR II as reported via EDB -- raises some questions in my mind:
  1. Is the second illustration really a dolium? It doesn't look that much like what's illustrated at http://en.wikipedia.org/wiki/Dolium. Why would a dolium appear on a Christian sepulchral inscription? Maybe someone like Sebastian Heath or Charlotte Tupman will have an idea about that.
  2. Are those two items really glyphs that should be "read" as part of the inscription and therefore marked up using the TEI "g" element (as Ryan has done), or should they be treated as figures or illustrations and therefore marked up a different way? If they are "glyphs", then what would be the corresponding glyph definition markup (if any) and where should it go in an EpiDoc file? Maybe someone like Gabriel Bodard or Marion Lamé will have an opinion about that.
Meanwhile, Georgia wrote to me as follows:
I like version 2. For one, I could see it and read it without any problems; something I could not do with version 1. I like the idea of being able to see pictures, texts, and translations of inscriptions on a single page. My question is: what are you trying to do here? What's the purpose, goal, etc. of Mob Epigraphy? And how can others help, contribute, etc.?
My goal with Mob Epigraphy is two-fold. First, I want to create more on-line, open examples of real inscriptions marked up in EpiDoc. Secondly, I want to see how far we can push an openly collaborative model in the practice of digital epigraphy, welcoming all interested parties in editing the text and pushing the boundaries on what we can and can't do with standard encoding and web publication.

How to contribute? There are many ways. This post highlights two examples. Ryan saw something missing and, exploiting the digital collaboration infrastructure provided by github, pitched in to fill the gap. Georgia had comments and questions and, after having some trouble with Blogger's comment functionality, sent me an email. Both are great ways to contribute, and I bet readers of this post can come up with more -- like suggesting answers to my questions above, or proposing more robust or interesting documentation of the inscription or elaboration of the encoding or HTML representation.

Previous post.





Flavia Faustina, version 2: style

This is a follow-on to my initial posting about the Flavia Faustina inscription from St. Paul's Outside the Walls in Rome. Another contribution to the "Mob Epigraphy" thread. Still a mob of one, alas ... if you see something you think could be done better -- epigraphically or technically -- please chime in! There are deliberate (and no doubt accidental) omissions and mistakes.

Not much substantive change, just style and inline image:

Tuesday, November 23, 2010

Mob Epigraphy: Sepulchral Inscription of Flavia Faustina

First installment in an irregular series (entitled "Mob Epigraphy") exploring the collaborative encoding, enrichment and publication of epigraphic texts on the web.

Here's the deal: what follows is surely incomplete, or even wrong, from any number of perspectives (textual, historical, technical?). So, if you have ideas or expertise with respect to the text, translation, descriptive information, EpiDoc/TEI encoding of the XML, HTML encoding, etc.), then please weigh in via comment or another blog post (just make sure I discover it somehow!).

What do you think would make this a better digital publication?


Thursday, July 29, 2010

EpiDoc Tools Released "as is"

If you visit http://sourceforge.net/projects/epidoc/files/ you'll now find readily downloadable releases of the following EpiDoc tools:
  • Guidelines
  • P5 Conversion Tools
  • Transcoder
  • Example P5 XSLTs
  • Example P4 XSLTs (deprecated; last/final release)
  • DTD (deprecated; last/final release)
  • Schema
  • CHETC JavaScript
These releases reflect the current state of code or documentation as it is to be found in our SVN repository. All of the tools have had README.txt files added in order to help the person downloading them figure out what they are and how to start using them. They also all have LICENSE.txt files that spell out the terms under which they are distributed. If you want to see our agenda, feel free to visit: http://epidocroadmap.pbworks.com/Release-Sprint-July-2010

Some of these packages are out-of-date or not feature-complete (e.g., especially the guidelines). We'll want to marshal volunteers in coming weeks and months to work on these discrepancies. There is in fact, already a group working hard on the guidelines. If you're not part of that group and would like to be, please shout out about it on the markup list.

My hearty thanks to Gabriel Bodard, Hugh Cayless and Charlotte Tupman, who assisted in today's sprint, and to Marion Lame, who also volunteered but could not be available during the time that I had scheduled.

Our next big step is to update http://epidoc.sourceforge.net/resources.shtml so that it properly reports on the state of each tool and links directly to the appropriate release. I'll be issuing a call for volunteers for that follow-up sprint shortly.

Friday, July 23, 2010

Linking to Google Books Content in an Ancient Geographic Way

I'm very interested in finding ways through Pleiades and other ISAW digital projects to support the efforts of Leif, Elton and Eric on the "Google Ancient Places (GAP): Discovering historic geographical entities in the Google Books corpus" project. In particular, I'd hope we can integrate this into the web interfaces for our projects:

ECS will work on a Web Service and Web Widget [that] will make it possible for Webmasters to add links to the ancient texts [in Google Books] within their websites, enabling the public and researchers to search for them easily.

Tuesday, April 6, 2010

Josh Greenberg on the Mellon/UVA "Shape of Things to Come" conference

The general perception of the academic humanities as far removed from the daily lives of the general public that is only heightened by isolationist jargon and publishing mechanisms that create rather than break down silos represents a massive failure to make the case for the value of that work to society ...
Epistemographer | Notes from “The Shape of Things to Come”

Wednesday, March 31, 2010

NEH Awards Grant for Pleiades Project

I'm happy to report that the National Endowment for the Humanities, through the Humanities Collections and Reference Resources program of the Division of Preservation and Access, has granted New York University $298,457 in outright grant funds to support an additional three years of funding for the development of Pleiades. Watch this space, and Sean's blog, for further details in coming weeks. Here's the official NEH announcement (we're listed in the "Nebraska to Wyoming" PDF, page 7).

Our sincere thanks to NEH, the anonymous reviewers of our application, and to all those in our user community who have helped us reach this important milestone!

Any views, findings, conclusions, or recommendations expressed in this blog do not necessarily reflect those of the National Endowment for the Humanities.

Thursday, December 17, 2009

Interoperation with Pleiades

I've had a few questions lately about how other web-based publications could be designed to support interoperation with Pleiades. Here's my working advice:

Any project that wants to lay the groundwork for geographic interoperability on the basis of Pleiades should:

1. Capture and manage Pleiades identifiers (stable URLs like http://pleiades.stoa.org/places/638753/) for each place one might want to cite.

2. Request membership in the Pleiades community and add/modify content therein as necessary in order to create new resources (and new URLs) for places that Pleiades doesn't yet document, but which are provably historical and relevant to content controlled by the external project.

3. Capture and manage stable URLs from Wikipedia or GeoNames that correspond to modern geographic entities that are relevant to the content controlled by the external project. Don't conflate modern and ancient locations, as this will eventually lead to heartbreak.

4. Emit paged web feeds in the Atom Syndication Format (RFC 4287) that also conform to the guidance documented (with in-the-wild, third-party examples) at:

http://www.atlantides.org/trac/concordia/wiki/ConcordiaAtomFeeds

and make use of the terms defined at

http://www.atlantides.org/trac/concordia/wiki/ConcordiaThesaurus

to indicate publicly relationships such as "findspot" and "original location" between the content controlled by the external project, Pleiades resources, Wikipedia resources, GeoNames resources and resources published by other third parties.

5. Alert us so we can include the entry-point URL for the feeds in the seeded search horizon list for the web crawler and search index service we are developing.

You can see how the Epigraphic Databank Heidelberg team has been thinking about how to accomplish this at:

http://www.atlantides.org/trac/concordia/wiki/PleiadesMoI

and

http://www.atlantides.org/trac/concordia/wiki/EDHgeographyTable

Tuesday, September 30, 2008

Happy Birthday, Chiron!

Χείρων·(Chiron), the "collaborative space for teachers of classics" is two years' old. Congratulations!

If you're not familiar with Chiron, check out the "About" page (Spanish; English)