Saturday, December 29, 2018

Niamey 1978 & Cape Town 2018: 3. Other angles on Wikipedias in extended/complex Latin

Last August, I began a set of three posts marking the 40th anniversary of the Niamey 1978 meting on harmonizing African language orthographies, and associating that with the Wikimedia 2018 conference in Cape Town - the first in sub-Saharan Africa. This post concludes the series.

The central element of this discussion is the extended Latin alphabet, which is used in the orthographies of many African languages to accurately convey meaningfully important sounds, and which often makes input with ordinary keyboards or keypads difficult.
Composite image from welcome page of Eve Wikipedia.

Looking again at Wikipedias in extended/complex Latin

So far, this article has raised some questions, made some admittedly superficial comparisons, and speculated as to factors related to success or not of Wikipedia editions in African languages. What might be ways of improving this analysis?

One way would be different approaches to sorting and categorizing the Wikipedias using the same numbers in the tables in the previous post. And one of those approaches could be to consider the relative degree of complexity of each orthography within a category. For example in category 3, Fulfulde uses 3, 4 or 5 extended Latin characters and Hausa 3 or 4 (depending on the country), and Luganda and Northern Sotho only one. Yoruba, when typed with the precomposed dot-under (aka "subdot") characters and combining diacritics for tones is less complicated than the same language using the classic style small vertical line under - because the latter requires a combining diacritic and that may mean "stacking" two diacritics on a character where a tone is involved.

A different way would be to look at the quality of content in the various Wikipedia editions. In some, the raw numbers of articles (which are the figures used in the tables mentioned above) are inflated by shell articles, by which I mean a stub that may have only a few words of text and perhaps an image. The list of Wikipedias does include a "depth" metric, which might be used (or perhaps adapted) to look for possible correlations between the quantity and quality of the content on the one hand, and the nature of the orthography on the other.

Yet another way would be to consider the numbers of people working on these editions. Wikipedia counts the numbers of users, active users, and administrators per edition. Could one use these figures to better understand whether the more successful Wikipedia editions in extended Latin (in terms of numbers of articles or a depth metric) are so because of the efforts of a relatively small number of users? That's not to imply any negative judgment of such cases, but it would be useful to know if if a complicated writing system (from the point of view of input) is not a hurdle for a large number of contributors (active users), or if it's really a case of a few savvy individuals carrying the load.1

And another approach would be to expand the scope of analysis to consider other factors: How many people speak the language? Is it taught in schools? How much printed material is available? Are there different dialects or written conventions that a contributor to or a reader of a given African language edition of Wikipedia must navigate? Any of these and perhaps others might, individually or in combination, affect the potential production of web content in general, and the success of these Wikipedias in particular.

One could also put the numbers in the background and do a qualitative study focusing on the experience of the editors of African language editions of Wikipedia. What might emerge from such discussions concerning the range of tasks involved in building and maintaining an active Wikipedia?

And then there are some stray questions certainly worth checking out. For instance, does the base interface language on which all of the African Wikipedias are built (English vs French) have any bearing at all on the success of Wikipedia editions? What about the degree of localization of the interfaces (from English or French to the language of content? And does that degree of localization relate at all to the complexity of the script?

Research towards success with African language Wikipedias

Although the number of Wikipedias in African languages is relatively small (about 13% of all editions, and collectively contain less than 1% of the total number of articles in all Wikipedias combined2), there are arguably enough data and diverse user experiences to give us a better idea of both how to develop small Wikipedias in Africa, and how much of a factor the scripts used to write them might be in their relative success.

Looking beyond African languages to the experience of Wikipedia editions in other languages written in extended Latin (and non-Latin scripts) would be instructive. This would likely highlight not only methods to facilitate input of diverse writing systems, but also supportive environments (or "localization ecologies") for these languages in general. 

Success for African language editions of Wikipedia may not be found in imitating work on other editions so much as it would in identifying ways to leverage the strengths and unique resources of African language communities. Nevertheless, facilitating input is fundamental, relying at its most basic level on common technology (for keyboards, etc.) and features of the Mediawiki software.

With rapid advances in language technology, an additional focus should be how to adapt speech-to-text to African languages to facilitate creation of content from oral narratives, interviews, and exposition. This is a topic I hope to return to later.

1. The Yoruba (category 4 orthography) and Northern Sotho (category 3) Wikipedias, for instance, each benefited at different times from large numbers of articles created by a single user in their respective communities.
2. That's 38 of 292 if the tiny Dinka edition is included; 37/291 if not. And about 293k articles out of 48.6 million total. (All as of July 2018.)

Friday, August 31, 2018

Niamey 1978 & Cape Town 2018: 2. Extended Latin & African language Wikipedias

Image adapted from banner on the Yoruba Wikipedia, August 2018
What are the implications of extended Latin characters and combinations for production of digital materials in African languages written with them? The previous post discussed some of the process of seeking to harmonize transcriptions, in which the Niamey 1978 conference and its African Reference Alphabet (ARA) were prominent. That process had a logic and left a legacy for the representation in writing of many African languages. This post asks if there is a trade-off between the complexity of the Latin-based writing system and how much is produced in it using contemporary digital technologies.

One easy, although by no means conclusive, way to consider this question is to look at Wikipedia editions in African languages (those that are written in Latin script). The following table disaggregates 35 African language editions by the number of articles (from the list of Wikipedias, as of 9 August 2018) and the four "categories" of Latin-based orthography1 introduced in African Languages in a Digital Age (ch. 7, p. 58):

Number of articles
Category 1
Category 2
"Category 1" + Latin 1 
Category 3
"Cat. 1" or "2" + any of Latin Extended A, B, etc, Add'l, & IPA
Category 4
"Category 3" + Combining diacritics
< 500
Swati (447)
Sango (255)

Fula (226)
Venda (265)
Chewa2 (389)
Dinka (75)
Ewe (345)
Sotho (543)
Tumbuka2 (562)
Tsonga (563)
Kirundi (611)
Tswana (641)
Xhosa (741)
Oromo (772)
Akan (561)
Twi (609)
Bambara3 (646)
Zulu (1011)
Kinyarwanda (1823)
Kongo (1179)
Luganda4 (1162)
Wolof (1167)
Gikuyu (1357)
Kabiye (1455)
Hausa (1891)
Igbo5 (1340)
Shona (3761)
Kabyle (2860)
Lingala (3028)
Somali (5307)
Swahili (44,375)
Yoruba5 (31,700)
> 50,000
Malagasy (85,033)
Afrikaans (52,847)
# of articles / # of editions = Average
146,190 / 14 =
54,281 / 3 =
20,655 /13 =
36,488 / 5 =
Grouping 1&2, 3&4
168,471 total articles / 17 editions = 
57,143 total articles / 18 editions = 

Looking at the top row with the smallest editions (less than 500 articles), one is tempted to highlight the high presence of African languages whose orthographies include extended Latin - categories 3 & 4. However, in the group of next highest number of articles (500-1000) there are more editions with category 1 orthographies (the simplest) than there are editions with category 3 in the group above that (1000-2000). And the next highest ranges (covering 2000-10,000) are roughly even between category 1 on the one hand, and 3 & 4 on the other. But then the three largest editions (and 3/4 above 25,000) are category 1 & 2.

So with just a visual analysis, there does not seem to be any clear pattern from arraying the editions in this way. Of course there will be other factors than the complexity of the script affecting the success of a Wikipedia edition written in it. But are there ways of looking at this raw data that can give us a clearer idea what might be the effect is of extended Latin - the ARA plus orthographies with other modified letters and diacritic combinations - on the size of Wikipedia editions?

One approach is to consider all the above editions combined, per category of orthography (totaling by column). This puts the focus on the degree of complexity of the writing system, perhaps muting the effect of other language- & location-specific factors. On the second to last row are column totals of the number of articles in all editions listed above, divided by the number of editions, to give an average figure.This yields an uneven pattern (2>1>4>3), since in the cases of 2 & 4, one large edition in a small total number of editions skews the category average up.

By the totals of the two simpler categories (1 & 2) and of the two extended Latin categories (3 & 4), however, one obtains possibly more useful numbers. This aggregation can be rationalized for our purposes here by the fact that the lower two categories are generally supported by commercially available keyboards and input systems,6 while the higher two categories, require a specialized way to input of additional characters and maybe diacritics (such as an alternative keyboard driver, or an online character picker).7

The figures thus obtained show editions written in extended and complex Latin having on average about a third the number or articles as those written in ASCII and Latin-1. Admittedly, this result is in part the result of the way categories have been chosen and figures aligned, but I'm proposing them as a perspective on the use of extended (and complex) Latin, and possible gaps in support. Before considering this in more detail, it is useful to compare with the numbers for non-Latin scripts.

What about non-Latin scripts & African language Wikipedias?

Number of articles
< 500
Tigrinya (168)
Amharic (14,321)
Egyptian Arabic (19,170)
# of articles / # of editions = Average
33,659 / 3 = 11,220
There are only three editions of Wikipedia in African languages written in non-Latin scripts.8 Two of those - Amharic and Tigrinya - are written with the Ge'ez or Ethiopic script unique to the Horn of Africa.

Arabic is the third. How to count this language for the purposes of this informal analysis raises a question. Arabic, of course, is established as a first language in North Africa for centuries, but it is also a world language, spoken natively in the southwest Asia (having originated in Arabia), and learned as a second language in many regions. Drawing users from this wide community, the Arabic Wikipedia is among the top 20 overall, with twice as many articles as all of the editions discussed above combined. It is more than an African language edition. For this analysis, therefore, I have chosen instead to count just the Egyptian Arabic Wikipedia.

Taking these three editions, we then get an average number of articles (11,220), which is close to what is seen for the Latin categories 1 & 2 (11,789). The usual caveats apply for such a small sample, but taking the numbers as they are, it is interesting that Wikipedias in the complex Arabic alphabet and the large Ge'ez abugida (alphasyllabary) are on average much larger than those of the ostensibly simpler extended Latin (3,175).9

Again, script complexity is but one factor, and in this case probably not the most important, since the two non-Latin scripts in question have long histories of use in text in parts of Africa - much longer than any form of Latin script. Nevertheless, from the narrow perspective of what is required for users to edit Wikipedia, the technical issues are in some ways comparable if even more demanding.

Arabic has had standard keyboards since the days of typewriters. The issues there are not so much the input, but whether systems can handle the directionality and composition requirements of the script.

The Ge'ez script on the other hand, does not involve complex composition rules or bidirectionality. However, it has a total of over 300 characters (including numerals and punctuation; more again if extended ranges are added). The good news is that there are numerous input systems to facilitate their input. Literacy in the script and availability of input systems would not be limiting factors for content development in major languages using this script. The difference in development of the Amharic and Tigrinya editions of Wikipedia may relate to both the larger population speaking Amharic (as a first or second language), and its use officially in a relatively large country (Ethiopia). Development of content in Tigrinya - a cross-border language - might also be hindered by issues particular to one of the two countries where it has many speakers (Eritrea).

From the above one might suggest that complexity of the written form (to be taken here as including the nature of the script itself, and the size of the character set) may be a limiting factor on content development, but that other factors, such as a literate tradition, official use, and technical support for digital production may overcome such limitations. In the case of African languages written in Latin script, however, any literate tradition is recent, and they are often marginalized in official and educational contexts. For those written with extended Latin, there is the additional factor of lack of an easy and standardized way of inputting special characters. Paradoxically, it seems, a modification of the most widely used alphabet on the planet may actually hobble efforts to edit in these languages.

Facilitating input in extended Latin for African language Wikipedias?

Wikipedia editing screen with "Special Characters"
drop-down modified to show all available ranges.
Assuming that the inconvenience of finding ways to input extended Latin characters may be a factor in the success of African language Wikipedias written with categories 3 and 4 orthographies, a quick fix might be to add new ranges for the modified letters used in African languages to the "special characters" picker in the edit screens. As it currently structured, the extended characters necessary for a category 3 or 4 orthography might be sprinkled around in up to 3 different ranges (see at right). And within each range, they are not presented in a clear order, so sometimes hard to find.

Since it may be too complicated to have a special range for each language edition, another possibility would be to draw inspiration from the Niamey 1978 meeting's ARA, and combine all extended Latin characters and combinations needed for all current African language Wikipedias into a common new range.

Of course as mentioned above, there are other factors that can contribute to the success or not of Wikipedia editions in African languages written with extended Latin, but this innovation would at least make editing more convenient for contributors to these  editions. And perhaps it might have a positive effect on the quantity and quality of articles in these Wikipedias.

In the third, and concluding article in this series, I'll step back to look at this analysis and consider some other ways to look at the data on African language editions of Wikipedia, and in particular, those written in extended Latin.

1. This categorization was intended to help characterize the technical requirements for display and input of various languages. Although the technology has improved to the point that more complex scripts are generally displayed without the kinds of issues one encountered a even a decade ago, input still requires extra steps or workarounds. The four categories are additive in that each higher category builds on those below, with added potential issues. It is also a "one jot" system in that for example, a single extended Latin character, say š in Northeren Sotho or ŋ in Wolof, makes their orthographies category 3 rather than category 1 or 2 (respectively), and the use of the combining tilde over the extended Latin character for open-o - ɔ̃ - makes Ewe a category 4 rather than 3. In terms of input, the higher the category, the more the potential issues with display and input (although technical advances tend to level the field, esp. as concerns display).
2. The only non-basic Latin character used in Chewa is the w with circumflex: ŵ. Apparently it represents a sound important in only one dialect of the language, and is used infrequently in contemporary publications. On the other hand, there is a proposed (not adopted) orthography for Tumbuka that includes the ŵ. Without this character, either language would be a category 1 orthography; with it, category 3.
3. Bambara is a tonal language. Most often, it seems, tones are not marked in text, however they can be for clarity, and some dictionaries make a point of indicating tone in the entries (not just pronunciation). If tones are unmarked, Bambara would be considered is a category 3 orthography; with tones, category 4. 
4. The addition of the letter ŋ puts Luganda in category 3 rather than 1.
5. The dot-under (or small vertical line under) characters used notably in Yoruba and Igbo are particular to southern Nigeria, and not included in the ARA. Yoruba in Benin is written with characters from the ARA.These are tonal languages, and tone is usually marked.
6. When I first proposed the category (itself a modification of an earlier effort), there were some questions why have a category 2 separate from category 1. That distinction had its origins in the early days of computing where systems used 7-bit fonts, meaning that accented letters (diacritic characters) used in, say, French or Portuguese, could not be displayed. Even as systems using 8-bit fonts enabled use of diacritics commonly used in European languages, display issues would still crop up (as a sequence of characters where an accented letter should be). Nowadays, such display issues are rare, and limited (as far as I can tell) to documents in legacy encodings. On the other hand, input of accented characters used may require, depending on the keyboard one is using, switching keyboard drivers or using extra keystrokes - so one will occasionally see ASCIIfication of text in such languages (apparently as a user choice).
7. The difference between categories 3 (extended Latin) and 4 (complex Latin) once were significant enough from point of view of display that informal appeals to Unicode to change its policy of not encoding new "precomposed" characters were common.
8. The Wikipedia incubator projects.includes several African language projects, which are not covered here. These include some in non-Latin scripts (Arabic versions, N'Ko, and Tamazight) and some in Latin-based orthographies. I mentioned one of the latter - Krio - in a previous post, and hope to do an overview of this space in the near future.
9. Average for all African language editions is 7704. By comparison the average for all Wikipedias is 166k.

Monday, August 13, 2018

Niamey 1978 & Cape Town 2018: 1. Some thoughts about extended Latin & content in African languages

Image features the 31 modified letters & diacritic combinations in
the African Reference Alphabet, 1978. (Nor all are currently in use.)

The world 40 years ago, when the Meeting of Experts on Transcription and Harmonization of African Languages took place in Niamey, and that of the Wikimania 2018 conference in Cape Town (which ended last month) seem very distant from each other. But from the angle of the written form of African languages at least, the concerns of the two events are not so distant.

One of these concerns is the extended Latin alphabets that were on the agenda in Niamey, and which are used in about half of the African language editions of Wikipedia. This post and the next consider these two vantage points, asking whether extended Latin is associated with less content creation, and what might be done to facilitate its use of the longer Latin alphabet.

Adapting the Latin script to African realities

In 1978, representatives of countries that had gained independence no more than a couple of decades earlier, or in some cases only a few years before, met in Niamey to advance work on writing systems for the first languages of the continent. One of the linguistic legacies of the colonial period was the Latin alphabet (even in lands where other systems had been used). But given the phonological requirements sometimes very different than what Latin letters represented in Europe, linguists added various modified letters, diacritics, and digraphs to write African languages (sometimes even a special system for a single publication1.

So, that legacy also often took the form of multiple alphabets and  orthographies for a single language, reflecting the different origins of European linguists (frequently Christian missionaries from different denominations), locations in which they worked (perhaps places where speakers of a language had particular dialects or accents), and individual skills and choices. After independence, many African countries undertook to simplify this situation, but they still often ended up with alphabets and spelling conventions different from those in neighboring countries.

The linguists and language specialists in Niamey, as in other such conferences of that era (many of which, like the one in Bamako in 1966, were supported by UNESCO), were concerned with further simplifying these discrepancies, with accurate and consistent transcription of languages that were for the most part spoken in two or more countries (whose speaker communities were divided by borders). That included adopting certain modified letters and diacritic combinations for sounds that were meaningfully significant in African languages (some of which correspond with characters in the International Phonetic Alphabet).

Language standardization, which is actually a complex set of decisions, was a real concern where there were on the one hand diverse peoples grouped in each state and on the other hand limited resources for producing materials and training teachers. At its most basic level, though, standardization of any sort required an agreed upon set of symbols and conventions for transcription.2

A reference alphabet for shared orthographies

The African Reference Alphabet (ARA)3 produced by the Niamey meeting was an effort in that direction. It built on the longer post-independence process to facilitate use and development of written forms of African languages - a process that had its roots in the early introduction of the Latin script (before the formal establishment of colonial rule) and efforts during the colonial period such as the influential (at least in the British colonies) 1928 Africa Alphabet. The ARA was intended - and to some degree at least still serves - as sort of a palette from which orthographies for specific linguistic, multilingual national, and cross-border language needs could be addressed.4

And that set of concerns - alphabets, orthographies and spelling conventions - turned out to be the starting point for later efforts in the context of information and communication technology (ICT) to localize software and interfaces, including Wikipedia and other Wikimedia interfaces, and to develop African language content online, including for Wikimedia projects. Even if it does not seem as visible as other challenges.

What I haven't seen is an evaluation of the efforts at Niamey and the other expert meetings on harmonization of transcriptions, although the most used of the characters in the ARA can be seen in various publications, and all but perhaps one are in the Unicode standard.

In any event. the situations of the various African languages are diverse, with some having well established corpora while others are "less-resourced," and in the worst case, inconsistently written.

Extended Latin and composing on digital devices

One important element in discussions in the process of which Niamey was part, was the role of modified letters - what are now called extended Latin characters - in transcribing many African languages. The ARA includes no less than 30 of them (22 modified letters and 8 basic Latin with diacritics5). These added characters and combinations are not intended to all be used in any one language, but represent standard options for orthographies. The incorporation of some of these into a writing of a single language makes the writing clearer, and has no drawbacks for teaching, learning, reading, or handwriting (although there are arguments against the use of diacritics). Since the establishment of Unicode for character encoding, the screen display of these characters is not a problem (so long as fonts have been created including glyphs for the characters).

However even the presence of even just one or two extended Latin characters leads to problems with standard keyboards and keypads - where are you going to place an additional character, and how is the user to know how to find it? This is a set of issues that was of course recognized back in the era of typewriters. One of the spinoffs from the Niamey conference was the 1982 proposal by Michael Mann and David Dalby (who attended the meeting) for an all lower-case "international niamey keyboard," which put all the modified characters (of an expanded version of the ARA) in the spots normally occupied by upper-case letters.

While that proposal never went far (I hope to return to the subject later) - due in large part to its abandonment of capital letters - it was but one extreme approach to a conundrum that is still with us. That is, how to facilitate input of Latin characters and combinations that are not part of the limited character sets that physical keyboards and keyboards are primarily designed for. It's not that there aren't ways of facilitating input - virtual keyboard layouts (keyboard drivers that can be designed like and shared, like Keyman, and onscreen keyboards) have been with us for years, and there are other input systems (voice recognition / speech-to-text being one). The problem is lack of standard arrangements and systems for many languages. Or perhaps in the matter of input systems, the old wag, "the nice thing about standards is there are so many to choose from," applies.

The result, arguably, may be a drag on widespread use of extended Latin characters, and as a consequence of popular use on digital devices of languages whose orthographies include them. Or a choice to ASCIIfy text (using only basic Latin), as has been the case with Hausa on international radio websites. Or even confusion based on continued use of outdated 8-bit font + keyboard driver systems, as witnessed in at least one case with Bambara (see discussion and example).

What can the level of contributions to African language editions of Wikipedia tell us about the effect of extended Latin? This will be explored in the next post: Extended Latin & African language Wikipedias.

1. For example some works on forest flora which had lists of common names in major languages of the region.
2. Arguably in the case of a language written in two or three different scripts, one could have a system in each script and an accepted way to transliterate between or among them.
3. The only other prominent use I found of the term "reference alphabet" was that of the ITU for their version of ISO 646 (basically the same as ASCII): "International Reference Alphabet." The concept of reference alphabet seems to be a useful one in contexts where many languages are spoken and writing systems for them aren't yet established.
4. This approach - adopting a standard or reference alphabet for numerous languages - was taken by various African countries, for example Cameroon and Nigeria. These efforts were without doubt influenced by the process of which Niamey and the ARA were part.
5. By comparison, the Africa Alphabet had 11 modified letters and did not use diacritics. All 11 of the characters added in the Africa Alphabet were incorporated in the ARA. It is worth noting that in the range of modified letters / special characters created over the years, some are incorporated into many orthographies, others fewer, and some are rarely used if at all.

Wednesday, July 18, 2018

Wikimania 2018: Sessions on, or of interest to, Wikimedia projects in African languages

The 14th annual Wikimedia conference - Wikimania 2018 - starts today, 18 July, in Cape Town, South Africa, and runs through 22 July. It is the second Wikimania to be held on the African continent - the first being at Alexandria, Egypt in 2008 - and the first in Africa south of the Sahara.

Here is a quick look from afar at what Wikimania 2018 sessions in the conference program might treat questions related to African language editions of Wikipedia, Wiktionary, etc. - what we have previously referred to as "Afrophone Wikis."


According to the program, the first two days - 18-19 July - are devoted to the Preconference, consisting of "various miniconferences and meetings." Among these, I'd make special note of the 2-day Decolonising the Internet Conference - "…the first ever conference about centering marginalized knowledge online!" Run by the NGO Whose Knowledge? (logo at right) as an invitation-only event, it has a theme that I'd consider of interest to increase African language presence on the internet.

Main conference

The main Wikimania conference follows, on 20-22 July. On the morning of the first day, Friday 20 July, there is a track devoted to Africa with three sessions, all of interest (titles link to project pages, which in some instances already have further links to slide presentations):
  • Babel's Tower: South Africa's Wikipedias: An overview and discussion of Wikipedia editions in South Africa's languages (focusing on the 11 official languages), and ways to address the poor development in most of those, including "possible interventions via both educational strategies and technological options." The presentation is by Michael Graaf, who wrote his dissertation at the University of Cape Town on South Africa's Wikipedias.
  • Africa's Wikipedias: "A panel to discuss the interesting challenges and possibilities of the Wikipedia language editions of Africa. Includes review of new tech to amplify efforts of editors." Panel includes several editors of African language Wikipedias (Afrikaans, Arabic, Swazi, Tsonga, and Xhosa).
  • The quotation of oral sources in a decolonization context: Discussion of how to incorporate oral citations in a resource that generally requires citation of written (ideally published) sources. Reference to an oral citations project in Namibia. Presentation by Bobby Shabangu and Stefanie Kastner.
That same morning, there is another session of particular interest from the perspective of working on African language projects (unfortunately conflicting with the Africa track):
 In the afternoon of the same day, another Africa-specific session that might have some content relevant to languages:
  • Coolest African Projects - Be inspired: Spotlights relatively unknown projects and activities by African Wikimedia affiliates. Presentation by Emna Mizouni, Felix Nartey, and User:Thuvack.
On the second day of the main conference, Saturday 21 July, the morning session has several sessions of special interest, including three in the Languages track:
  • Wikipedia for Indigenous Communities: Compares Western and OvaHerero (Namibia) approaches to knowledge, and discusses a project approaching Wiki editing in a way more acceptable to their community. Presented by Peter Gallert.
  • How majorities can support minority languages: Although description does not indicate Africa content, it deals with how people in positions of relative power (in this case speakers of dominant languages) can help those in positions of less power (speakers of "minority" languages) with their Wikipedia projects. Presentation by Jon Harald Søby, Astrid Carlsen, Jean-Philippe Béland, and User:Barrioflores.
  • Including minority languages in Wikimedia projects, a strategic approach: Again, no specific Africa content indicated, but a possibly relevant discussion of how to include minority languages in Wikimedia projects. Presentation by Ahmed Houamel-Bachounda.
Also in the morning, sessions dealing with Africa in the Education track (thus conflicting with the above), but without indication whether African language projects will be discussed, or just major Europhone language projects like English & French:
On the morning of the last day, Sunday 22 July, four sessions in the Communication track look interesting from the point of view of African language projects (even though none of these are specifically mentioned in the session descriptions except for the last one):
  • Working towards Growing Local Language Content on Wikipedia (GLOW): Discusses a 2017 collaboration among Wikimedia Foundation, the Centre for Internet and Society (CIS), Wikimedia India chapter (WMIN), user groups and external partners on a "pilot project in India to encourage local Wikipedia communities to create locally relevant articles in Indian languages." The results will inform development of the GLOW program, which is explained. Presentation by Jack Rabah and Rupika Sharma.
  • Record every language of the world village by village, with Lingua Libre: Discusses project to facilitate "the recording process of words in any language (even minor languages or dialects), uploading them to Wikimedia Commons and reusing them on other projects such as Wiktionary, Wikipedia or Wikidata." Presentation by User:0x010C.
  • Every Language in the World: Introducing Wikitongues: Focuses "on the activities coordinated by Wikitongues, a not-for-profit organization promoting the use and preservation of every language in the world" through collection of oral histories. Presentation by Daniel Bogre Udell.
  • Diglossia and Multilingualism: A help or a Hindrance to Arabic Wikipedians?: Explores "the ways students who are native speakers of Arabic [which has a standard & many vernacular forms] in a multilingual educational system overcome the obstacle of sharing knowledge by using a common idiom while allowing millions of readers engage with the content they create. This session will also suggest solutions for communities with similar language challenges inspired by the educational model used in Arabic-speaking schools that participate in the 'Student Write Wikipedia' program." Presented by Bekriah S. Mawasi.
The above should not be interpreted as meaning that other sessions would not be of interest. This is a subjective selection based on my reading of the descriptions. On the whole it is nice to note the optimism in several cases, with regard to African language projects, and also the efforts to accommodate and integrate oral content and sources.

By coincidence, the timing of Wikimania 2018 corresponds with the 40th anniversary of the Niamey expert meeting on transcription and harmonization of African languages, so I'll draw some connections between the two seemingly very different events in the next post.

Note: The first two images above are from the webpages for the event (Wikimania 2018) or organization (Whose Knowldge?) concerned. Attribution of the third image can be found on the linked Wikimedia Commons page.

Tuesday, July 17, 2018

Expert Meeting on the Transcription & Harmonization of African languages, Niamey, 17-21 July 1978

Niger's National Assembly, where the 1978 meeting was
formally opened. (Source:
Forty years ago today, the Meeting of Experts on the Transcription and Harmonization of African Languages began in Niamey, Niger. Along with the 1966 meeting in Bamako, it was one of the more significant of a series of meetings* organized in Africa with the assistance of UNESCO to deal with questions relating to standardization of the written forms of African languages.

This expert meeting was at once less ambitious than the 1966 Bamako meeting - seeking "harmonization" rather than "unification" of systems for writing - and wider in scope - including representatives from more countries around the continent:Angola; Benin; Burundi; Cameroun; Central African Republic; Guinea Bissau; Ivory Coast; Mali; Niger; Rwanda; Senegal; Tanzania; Uganda; and Upper Volta [now Burkina Faso] (some countries had more than one person). Plus France, United Kingdom, and Yugoslavia. (Representatives from Congo, Ghana, Nigeria, Togo, and Zaire [now DR Congo] were not able to attend.)

This diversity also meant that the number and range of languages considered in Niamey was greater than in Bamako. On the other hand, like Bamako, the Niamey meeting focused only on the Latin-based transcriptions used in educational contexts (notably literacy) by the recently independent governments in sub-Saharan Africa.

This conference was particularly notable for its connection with the African Reference Alphabet, which was intended to provide a common character for each sound encountered in main African languages (rather than each country devising its own symbols or character combinations).

African Reference Alphabet. Source: Proceedings of the Meeting, UNESCO, 1981.
This alphabet was later amended by linguists David Dalby, who participated in the Niamey meeting, and Michael Mann, to include a number of additional characters. They also suggested a lower-case only alphabet,with a keyboard design using both registers to accommodate all the letters. (This keyboard was never adopted as such.)

 This effort was significant in influencing orthographies adopted for many languages (although not all). However it did not seem to be explicitly connected with the contemporaneously emerging digital text standards. Although many of the characters in ISO 6438 "African coded character set for bibliographic information interchange" were the same, there were differences that indicate the latter was the result of a separate process (or perhaps "fork" in today's software development terminology).

A few years ago I had hoped it would be possible to use the occasion of the 40th anniversary of the Niamey expert meeting to organize a conference to review the status and influence of the African Reference Alphabet and its descendants - with particular attention to technical support in ICT - and issues related to non-Latin scripts used for African languages. And perhaps to broach other topics related to use of African languages in the spirit of the efforts of a half-century ago.

Perhaps such a conference will prove useful in the future, but for the moment I'll mark this 40th anniversary with a series of short posts on the 1978 Niamey expert meeting itself and/or contemporary efforts that in one way or another reflect its aspirations.

* Several other expert meetings during this period addressed more specific sets of issues.

Friday, July 13, 2018

A movie on the life of Bishop Samuel Ajayi Crowther?

Bishop Crowther, 1888
Source: Wikipedia

Bishop Samuel Ajayi Crowther (c.1809-1891) was a remarkable figure in West African history, whose life bridged the end of the Atlantic slave trade era and the beginning of the period of European partition of the continent. Could his story be the basis of a major film production?

A compelling life story

Captured at age 12 or 13 in what is today Oyo State in southwest Nigeria, separated from his family, and sold into slavery, he was "recaptured" by a British anti-slaving force and eventually found his way back to his homeland and became prominent as a linguist and a member of the Anglican clergy (becoming in 1864 its first African bishop) in what was to become Nigeria. Towards the end of his life, however, this success collided with increasingly racist attitudes associated with the imposition of colonial rule.

This compelling story deserves more attention, and it is easy to imagine it being the subject of a major motion picture production. The idea is in no way new. From some contacts in Nigeria, notably Dr. Tunde Adegbola, and through him, filmmaker Tunde Kelani, I learned some years ago that there is interest in the production such a drama. There is at least one published dramatization of Crowther's life, a play focusing on his difficult later years, written by Prof. Femi Osofisan

Certainly the most dramatic episode in Crowther's life was the unexpected reunion with his mother in 1846, which he himself recounted in these words (as presented in an 1892 biography):
"August 21. The text for this day in the Christian Almanac, is 'Thou art the Helper of the fatherless.' I have never felt the force of this text more than I did this day, as I have to relate that my mother, from whom I was torn away about five-and-twenty years ago, came with my brother in quest of me. When she saw me she trembled. She could not believe her own eyes. We grasped one another, looking at each other with silence and great astonishment, big tears rolling down her emaciated cheeks. A great number of people soon came together. She trembled as she held me by the hand and called me by the familiar names by which I well remember I used to be called by my grandmother, who has since died in slavery. We could not say much, but sat still, and cast now and then an affectionate look at each other--a look which violence and oppression had long checked--an affection which had nearly been extinguished by the long space of twenty-five years. My two sisters who were captured with us, are both with my mother, who takes care of them and her grandchildren in a small town not far from here, called Absika. Thus unsought for--after all search for me had failed--God has brought us together again, and turned our sorrow into joy."
Although a figure of the 19th century, Crowther's life story resonates beyond that period so critical in African history. The theme of separation and reunion is universal and powerful. His encounter with some of the worst aspects of racism, which unfortunately is still a very present problem (it was not until 2014 that the Church of England finally apologized for the treatment he received). Crowther's efforts in the area of Christian-Muslim dialogue reflect an important ongoing process. And his work with African languages, particularly but not limited to his native Yoruba, have an ongoing influence.

Ajayi Crowther, translation, and transcription

Early in his time in what is now Sierra Leone, where he was settled after reacue by the British, Crowther took an interest in languages and linguistics. Aside from his mother tongue and the English he learned in Freetown, he also learned Temne. In the course of his education leading to clerical vocation, he learned Latin and Greek, which later were essential in translating the Bible into Yoruba. He also learned Igbo (and contributed to work on translating the Bible into that language), Hausa, (in the context of a trip to the north), Nupe (for which he also published a grammar and vocabulary), and Igalla (which is closely related to Yoruba).

Character combos for open-e, open-o, & /sh/
in Yoruba. The small line is the "classic" look;
a dot under seems more often used these days.
Apparently much of the early work on writing Yoruba in the Latin alphabet (an older Ajami transcription already exised, likely used mainly by Muslim Yorubas) was begun by missionaries in Freetown working with Yoruba speakers who had been settled there, and Crowther collaborated with them. The system Crowther used to write Yoruba, including marks under e and o to denote open vowel forms of them (the language having a 7-vowel system), presumably built on those early efforts. With some modifications, such as tone markings, that orthography is still in use today, a significant contribution to the written forms of African languages.

When back in what was to become Nigeria, Crowther worked on A Vocabulary of the Yoruba Language, (1852), which was apparently the first linguistic work published by an African. His translation of the Anglican Book of Common Prayer and Bible into Yoruba came later.

Ajayi Crowther's place in African history

Although I had learned some basics about Bishop Crowther in my limited study of and reading on African history over the years (it is relevant to my work, but not my field), I first took an interest in his story due to his role in establishing the Yoruba orthography.² I hadn't realized until later that opinions of his role in Nigerian and West African history varied. In Prof. Osofisan's words (from prefatory notes to his play):
"Crowther has been much vilified by African scholars and historians, who accuse him of having been merely a lackey to the white colonials. His work as a pioneering missionaery who travelled widely and extensively along the River Niger, establishing missions and, above all, putting up a staunch fight against slavery, is hardly even appreciated. Even his works as translator and scholar, who established schools in many parts of the mission, and worked out the first written alphabets and primers for not only Yoruba, but also Igalla and Igbo, is always glossed over."
Maybe it's time - already a few years after the Church of England's apology for what was done to him late in his career - to "rehabilitate" Bishop Crowther's image in history more generally, and to provide the current generation with a new perspective on a complex and influential life and life's work?

A film on the life of Ajayi Crowther

Crowther achieved several firsts, as a clergy in his adopted religion, and as a prominent early linguist specializing in African languages. He engaged with African cultural, linguistic, and religious diversity. He was one of the first students at the famous Fourah Bay College. He personally experienced both slavery and the institutionalized racism of colonialism, but rose above them. He left a legacy that has perhaps been underappreciated.

So, could this story, to which the above sketch does not do full justice, be the basis for a major feature film production? One involving Nigerian and international producers and actors? Could it fully, accurately, and appropriately treat the linguistic aspects of the story?

The market is there for big Africa-themed productions - including ones in which African languages figure prominently. The action film Black Panther is a recent example. The 1997 historical drama Amistad did well. Language, transcription, and translation may seem like a harder sell, but the recent sci-fi film Arrival had linguistics as a central component of the plot.

Ethnic dimensions might need care to navigate. Crowther was obviously a Yoruba figure, but also spent formative years in the Creole community of Freetown, and later traveled and worked widely in what was to become Nigeria. The episode of his capture into slavery is said to have involved "Muslim Fulani" as well as eventually Portuguese slavers - so how to be historically accurate without feeding stereotypes should be a priority.³

The biggest challenges are that there is no script or even script treatment on the subject, and that research on the subject might yield more than one reasonable plot line to tell the story, with the potential for conflict between commercial imperatives (especially the bigger the production) and the importance of historical and linguistic accuracy, and sensitivity to the people groups treated.

So this post is intended to give a little more lift to an idea that merits consideration and, hopefully, action.

1. Femi Osofisan, Ajayi Crowther: The Triumphs and Travails of a Legend, Bookcraft, Ibadan, 2006. The play was first performed in Lagos in 2002.
2. This was about 10-15 years ago when I was focusing on support for African writing systems on computers and the internet.
3. There are currently reports of conflicts in Nigeria between herders, mainly Fulani (Fulɓe), and farmers of other ethnic groups, which have led to fatalities.