Skip to content

Data of the project "The Shawi-type Arabic dialects (FWF P 33574)".

License

Notifications You must be signed in to change notification settings

acdh-oeaw/shawi-data

Repository files navigation

SHAWI Transcription Repository

This git repository hosts the transcription data of the project The Shawi-type Arabic dialects (FWF P 33574).

PI: Stephan Procházka (University of Vienna)
National Cooperation Partner: Charly Mörth (Austrian Academy of Sciences)

Status

THIS IS PRELIMINARY DATA AND COPYRIGHTED MATERIAL!

If you want to use any material in this repository please contact PI Stephan Procházka (University of Vienna).

This will change at the end of the project.

Directory Structure

Directory Content Remarks
001_src Original sources Source documents (e.g. raw transcriptions)
080_scripts_generic Conversion Scripts mostly the ELAN2TEI conversion script (implemented in Python) which generates the initial TEI data prior to tokenization based on the ELAN transcription documents in 122_elan
082_scripts_xsl XSLT scripts XSLT scripts
103_tei_w TEI-XML with tokens This is where ELAN2TEI puts its output. Re-running TEI2ELAN will overwrite all content in this directory, so do not do any manual changes here but copy the file to 010_manannot beforehand.
010_manannot manually annotated TEI-XML Tokenized TEI documents from 103_tei_w which are manually annotated.
802_tei_odd TEI customization (ODD) This is the source of truth for the SHAWI Schema and the HTML documentation generated from it.
130_vert_plain NoSketch Engine Verticals NoSketch Engine text verticals
803_RNG-schematron Schemas derived from the ODD in 802_tei_odd
804_xsd Schemas derived from the ODD in 802_tei_odd
850_docs Documentation Further data documentation, esp. the HTML documentation of the ODD

The oXygen project shawi.xpr contains the configuration for various transformation scenarios.

The directories css, html, js and xsl are used by the TEI Enricher.

Other data locations

General Workflow

For more information refer to the SHAWI Data Processing and Curation Document

The following steps happen before data is ingested into this repository:

  • fieldwork (recording audio etc.) – The recordings so far cover only material collected in previous campaigns
  • collecting metadata: – This is collected at curated in [the metadata spreadsheet].

Workflow steps reflected in the data in this repository:

  • Transcription and translation – Curators segment the audio recordings into sensible sets of "utterances" and transcribe and translate them using ELAN. When transcription has finished, the curator adds the ELAN document(s) to 122_elan and pushes the changes to git.
  • Tokenization This push triggers the ELAN2TEI conversion workflow which takes all *.eaf files in 122_ELAN and transforms them into tokenized standalone TEI documents, storing them under 103_tei_w. Additionally, a TEI Corpus file is generated which includes corpus level metadata and controlled vocabularies.
  • Annotation After transformation to TEI, curators annotate the texts using the TEI_enricher and store the results under 010_manannot.
  • Conversion to NoSke Verticals During the tokenization process, a NoSke-compatible vertical is created which incorporates the annotations found in ``010_manannot` .
  • Deployment Inteagration of deployment in the workflow TBD