Putting Crowdsourcing In Action


Crowdsourcing is increasing in popularity as a form of distributed problem solving enabled by digital technologies. “The crowd” is invited to contribute towards projects, this contribution potentially being in the form of knowledge or design skills. On June the 3rd this year an interdisciplinary workshop investigating crowdsourcing and citizen science convened. It brought together experts and practitioners from many disciplines that apply a crowdsourcing approach, presenting outputs and how crowdsourcing aids projects from GalaxyZoo (an interactive project for volunteer classification of galaxies), Artmaps (an application for crowdsourcing information on Tate digital artworks) and Taarifa (a platform and community supporting the crowdsourcing of public service issues in the developing and developed world).

My personal presentation was on Taarifa, a project I started in 2011 to support community based public service delivery. Since then I’ve worked in collaboration with the World Bank in Uganda to support the Education and Local Government Ministries with reporting across the country; what started as a pilot was rolled out quickly to cover 111 districts, over a year of an at-scale pilot 14,000 reports were received and acted upon. This resulted into wider research into the wider use of public participatory service delivery in developing countries (FOSS4G Taarifa paper). The uniqueness of Taarifa is that it has been developed and maintained by wholly volunteer contributors, creating free and open source software. The contributors to Taarifa are as diverse as the problems which Taarifa addresses, ranging from PhD Candidates like myself to Physicists, Bankers and Community Organisers. Consequently, Taarifa doesn’t just look after a platform of software; it acts as a forum to share knowledge, experimentations and innovation.

Taarifa was conceived at the London Water Hackathon, as an innovation around water access and quality in Tanzania. Access to water in Tanzania currently covers less than 50% of the country’s population and with 38% of the water infrastructure, like taps, are graded as non-functional. Currently the Ministry of Water has a WPMS (Water Point Mapping System) developed after a countrywide survey. However, the system has is no functionality to update the status of the water point or view a history of service problems. This is combined with poor performance of repairs nationally if water points are repaired; citizens are disenfranchised with current methods of reporting water faults, if they can report at all. The ecosystem around supporting the repair of water points is non-existent; consequently millions of Tanzanians have no access to publicly delivered water.

It is important to stress that there isn’t ‘one’ solution to the problem of water access nor is there ‘one’ platform or software to ‘fix’ the problem. There is no one discipline that can resolve the issue of water access, there has to be a multidisciplinary approach, to a multidisciplinary problem. Cartography, Economics, Engineering, not one discipline can wholly resolve the issue of water access, nor it is an issue which can be researched and resolved through the lens of one discipline. The societal side of technology needs to not be just taken into account, but integrated into the core of the design with the people who face the issue and who will use the technology to resolve it. It is imperialistic and deterministic to assume that technology can just ‘fix’ the kind of complex issue of water access, especially as the technology is, in effect, imposed broadly by outsiders to the community in which it is intended to take root. Hence, an understanding of the community is needed; who the users are, how water access is dealt with currently and the general state of affairs. From this we have created two streams of Taarifa, one that is currently implemented and one that is currently being designed, incorporating lessons learned from initial deployments.

The first iteration of Taarifa’s design story and user action assumed that mobile connectivity wasn’t an issues and that there was an active organisation, be it government or an NGO wanting to resolve water access issues, another predicate was that the water infrastructure was adequately mapped. This led to the following reporting process for a water issue; When a report is made, for example from a Community Water Officer or a concerned citizen, it goes into the Taarifa workflow, which identifies the specific water point from the database. The reporter is then notified, thanked that they have made a report and given an estimation, based on prior time taken to repair broken water points in that district on how long it will take. An engineer is informed what is wrong with the water source. Once an engineer has been selected, a verifier can verify that repair has been completed satisfactorily. Importantly, at each stage the initial reporter is informed about the progress of the repair. This was the version trialled in Uganda.

Subsequently, learning from how Taarifa was deployed and used, the design is now intended to incorporate offline capability and ‘marketplace’ functionality. The offline capability due experiences in Uganda that connectivity wasn’t universal (this was not a surprise, however, improvements to the paradigm should be incremental) the marketplace due to the capacity of local government and organisations. If a district has no capacity to repair a broken water point, the cost could be estimated by a number of engineers receiving information about the problem and they bid using their phones. A micropayment is taken to support the system, providing a surplus, potentially reinvested into creating new capacity. Micropayments are ubiquitous in the developing world, effectively replacing a formal banking infrastructure, hence are familiar to the communities who will use this method. Consequently this should hopefully be viewed as an extension of what already exists, not something completely new.

What does all this mean within the context of the Crowdsourcing in Action workshop? Broadly it allows us, as academic researchers to typify crowdsourcing and understand more. Taarifa acts as a community crowdsourcing code and by extension curating community reporting issues in developing countries. Artmaps develops applications for use on smart phones that will allow people to relate artworks to the places, sites and environments they encounter in daily life. GalaxyZoo leverages the many eyes of the crowd to process space imagery data. Thematically, all the projects presented utilised volunteers to provide information, process it and return it to the user and other interested actors.

After the initial presentations we formed groups, of other experts and practitioners to build a common model of what crowdsourcing means to their projects and work. Then coalescing at periods to feedback practice and information learned from other participants. In doing so, we learn from successes and failures of others, understanding common themes for collaboration.

In identifying these common themes, it hopefully sets an agenda to focus on specific factors and communities under the crowdsourcing agenda. Jeremy Morley and I are planning a future “How to Hackathon” event building “Crowdsourcing in Action”. Hackathons allow volunteers (generally) to co-create ‘hacks’ to problems. In its truest sense you accelerate innovation by combining a random mix of people and skills, providing a set of previously unsolved problems, then observing what happens. As was identified in the Crowdsouring In Action, we can observe the states before crowdsourcing; we can help provide a process for participants; we can observe and process the result. However, an understanding of how participants use the tools to conduct crowdsourcing is scant. By now focusing on hackathons, we hope to discover more on how the design and development of crowdsourcing works.

GISRUK 2013

On the 3rd to the  5th of April I attended GISRUK (Geospatial Information Research in the United Kingdom) to give a paper on Community Mapping as a Socio-Technical Work Domain. In keeping with Christoph Kinkeldey‘s love of 1990s pop stars Vanilla Ice made a second slide appearance, leveraging the fact it’s a very technical academic title. In short I’m using Cognitive Work Analysis (CWA) to create a structural framework to assess the quality (currently defined by ISO 19113:Geographic Quality Principles – well worth a read…) where there is no comparative dataset.

CWA is used to assess the design space in which a system exists, not the system itself. In taking a holistic view and not enforcing constraints on the system you can understand what components and physical objects you would need to achieve the values of the system and vice-versa. In future iterations I’m going to get past first base and look at decision trees and strategic trees to work out how to establish the quality of volunteered geographic data without a comparative dataset. Building quality analysis into day one, as opposed to being an after thought.

Written and submitted from Home (52.962339,-1.173566)

 

A Manifesto for the OSM Academic Working Group

A fellow member of the OSM Foundation replied to a conversation on the mailing list: “As a guerrilla academic…“. The context was around a suggestion for increased academic cooperation within OSM. To this end I proposed a new working group for the OSMF: Academic Working Group. This would have the aim of improving the academic quality and communication of those using OSM in their research and facilitating collaboration.

Below is the start of the manifesto. It’s not complete, but it’s a start.

Background

Academic institutions use OSM data. Be it part of their published research or testing hypotheses. Some of the publications are listed on the wiki: http://wiki.openstreetmap.org/wiki/Research. However within OSM and OSMF this research is undertaken under the researchers own initiative. Researchers are looking at OSM through recommendation (supervision) or self interest within their own academic structures. Given the growth of OSM and the research into it, it seems likely that academic interest will widen and grow.

Goals

Support academic research in OSM, encouraging best practices and acting as a forum for researchers. This has the aim to support researchers starting out with OSM but also to unify a community of existing researchers; collaborations and knowledge sharing will hopefully follow. Identification of areas of research for the community as a whole among potential themes of usability and business models (as a starting point).

Tasks

  • Uniting existing researchers, either at existing institutions or those following independent academic study.
  • Provide documentation (a la learnOSM) but focused for researchers.
  • Provide a forum for researchers to discuss their research and bridge into the community
  • Support and provide problems to the academic corpus.
  • Communicate potential collaborations, needs, wants.
  • More TBD

Working Group vs. Community

I think this is hitting a gap that exists in the community currently. I don’t see potential areas for conflict. However that being said do we have enough members within the OSM(F?) to create and steer the working group?

WWWG vs. other WGs

There is a small amount of overlap in interest between this proposed AWG and other Working Groups.  I can see potential overlap with communications and strategic working group. Communications as this would aim to focus on building up the OSM academic community. Strategic as they may wish to commission studies or at least support them, into critical areas of OSM.

Next steps 

Again, I’ll throw this to the OSMF. Where should we go from here?

Written and submitted from the London St. Pancras to Nottingham Train.

Research Impact

The digital economy program encompasses five universities (Nottingham, Cambridge, Reading, Exeter and Brunel) with numerous Doctoral Training Centres (DTC) training ‘the next generation of researchers’. I’m quite fortunate to be at the Nottingham Digital Economy DTC, which is rare in that it has a combined research hub and DTC. From time to time the hubs and DTCs get together in conference, where the collective research efforts and outputs is demonstrated. However every so often the research council – i.e. the people that write the checks – wish to see the results of their labours.

Due to the nature of the PhD programs – in a cohort, as opposed to individual working – they wished to understand value for money, more specifically the research impact that the programmes have had. The format for this was a poster/live demonstration of gadgets, followed by an interview session with some direct, searching questions. It quickly became apparent the role of the DTC staff was multifaceted, focusing not just on research and supervision but deftly dealing with the work associated with the DTC. Having a window into this world offered a very different perspective on the process of research councils, from the council’s expectations to the reality of research output and the seemingly intangible process of ascertaining ‘impact’ – Impact seemingly being if you’ve done something useful and interesting.

Meeting the other DTCs with the twist of funders and assessors was a good break from the usual, made special that it only happens every few years. The only downside of the process was the EPRSC (Engineering and Physical Sciences Research Council) was based in tragic town of Swindon. However the bods have pulled a bit of a wheeze by placing the building next to the train station, adding a dedicated footbridge. This means, if day tripping, you don’t physically have to enter the town. Instead you walk in the footbridge (from the set of Threads) direct to the centre, unfortunately without seeing the famous Swindon vistas! All in all a good day all around!

Fear and Loathing in Las PhD

This post I guess has been a long time coming, basically hit it’s zenith and then subsided. About a month ago I had serious doubts within the PhD, around whether I was ‘good enough’ to complete. The majority of doubt focused around completing what I had perceived to be an easy task of implementing a ‘simple’ algorithm. This turned into three weeks of nothing. Breakthrough occured on what was supposedly a three day break in Marseille – before a 12 day conference schedule in Avignon (AGILE) and Amsterdam (WhereCampEU).

It would be fair to say I’d hit the lowest point of the PhD then. It was a sequential thought process, if I can’t do this simple thing, how am I prepared for the harder things later. Doubt set in, and the analysis was concluded in that I should quit. Then the break through came and all was good, confidence restored. I then read ‘The Valley Of Shit‘ a blog about going through the same thing; Valleys lead to somewhere else – if you can but walk for long enough. Unfortunately the Valley of Shit can feel endless because you are surrounded by towering walls of brown stuff which block your view of the beautiful landscape beyond.”

Anyhow, I feel out of the valley now. All is good.

Written and submitted from Coffee Company, Amsterdam, Netherlands (52.371554,4.896772)

Wernigerode PhD Summer School

Another hectic few weeks draws to a close. The literature review is more grounded than before nearing 5,000 words and pretty much all of them are ‘good’. The past week has been spent in Wernigerode in the Hars region of Germany on an AGILE (Association of Geographic Information Laboratories Europe) PhD Winter School.

Mixing PhD candidates from research centres across Europe from GIS, Geomatics and other disciplines the event started with the customary ice-breaker. Even over (some very very good) local beer, it was clear that the participants were from very diverse backgrounds, most in the process of doing interdisciplanary research, with projects looking at conflating ontologies, predicting the location(s) of serious criminals, visualising change and crowd sourcing 3D building models among many others.

The first day started with introductions, followed by 10-15 presentations with questions on our topics. Taking all day it was good to see how other geospatial PhDs evolve in differing subjects and countries. During a very German (schnitzel) lunch we wandered in the forest surrounding Wernigerode. Though the place is quite of the beaten track it really is worth a visit if you want to chill out.

The second day started with a very good talk from Bénédicte Bucher from IGN about the differing research groups of the French National Mapping Agency, concluding with her thoughts on the PhD process. She noted that when you first start it’s like being in a Bazaar, you see the different pathways however, eventually, you’ll be forging your own path in the wilderness over tough terrain.

This followed into break out sessions with other participants to start either a paper or initiative on your subjects. Being in the Volunteered Geographic Informations group we went back to basics. Though we were from differing subsections of VGI (Crowd sourcing 3D indoor models, policy of VGI in government and myself in Community Mapping) our common ground was the lack of definitions in VGI, so we proposed an AGILE initiative to fix this.

After putting a 10 minute presentation together (where we managed to get MC Hammer into a slide, under the rather tenuous headline of “Break It Down”) we formally ended the winter school with our proceedings in hand. Then a spot of further networking in the only club in Wernigerode…!

Written and submitted from the DB RegioBahn Magdeburg – Berlin train.

The Role Of Ethics

All academic research should be ethical, guidance comes from the research councils and, generally, from individual departments and faculty ethics boards. Unfortunately the responsibly produce ethical research is on the individual researcher, not with boards and faculty. When I took my first year viva, in it contained some analysis to counter some questions on data and what is possible, knowing it would be a consideration.

In it I produced a time series analysis of the 2010 Egyptian elections with the data from U-Shaid an Ushahidi instance. This hadn’t gone to an ethics board, it had only been seen by my supervision team to allay fears on whether data could be found and analysed (The question on how useful the data is for my research is still being answered). On the day the viva went well with a few discussion points on the data and what it means for future analysis – At this juncture, I’d like to point out that this isn’t a mea culpa of saying I do unethical research!

Because I have decided to use an ‘ethnographically informed’ methodology/action research a lot of the research is based upon my experiences and personal narrative. This then is supported by interviews with experts or people working within the domain. This part of the research isn’t controversial, the ethics around interviews and ethnography are well structured, with clear pathways, dos and don’ts. Fixing these ethics seemingly is based around not doing harm and ensuring informed consent; this being where the subject is being made fully aware of the aims and goals of the experiment before participating. This isn’t cast in stone, especially when trust/deception are part of the experiment, but considerations need to be made. Effectively don’t do the Milgram Experiment, or kill anyone.

At first glance it doesn’t appear that any of that has anything to do with my research. As I’m revolving around community mapping, public participatory GIS (PPGIS), volunteered geographic information (VGI) the ethics are a lot more obfuscated. Something as simple as data collection and its nature can be questioned. Primarily I will use Open Street Map as my VGI source, following Muki Haklay’s work on comparing VGI with an authoritative one, to potentially answer a question around data quality in slum mapping. Informed consent here for mapping parties or the mapping process shouldn’t be too difficult, but what about the pre-existing data?

It’s not realistic to get consent from every person that has contributed to the map. When OSM started blank spots were common, not just in developing nations but in developed nations also. Due to satellite imagery and organisations like HOT (Humanitarian Open street map Team), mappers have also been tracing satellite imagery when a physical survey was impractical. As we want to analyse this information how do we gain consent? Should a clause be put into OSM’s licence (this is bad idea)? Is there a differentiation between mappers which trace and those that physically survey, are all maps created equal?

There is scant research out there, namely “Research Ethics for Studying Open Source Projects” and “Internet Research: Privacy, Ethics and Alienation – An Open Source Approach”. Specifically relating to geospatial data and OSM there is none. The only mention relating to OSM is SK53’s post on research relating to OSM being behind a paywall, which is a fair point but still doesn’t answer whether we should be using data without consent. I sense a PhD chapter forming.

Written and submitted from the Nottingham Geospatial Building (52.953, -1.18405)

Ergonomics of Hackathons and Code Sprints

I recently posted a rather long blog regarding my experience at the London WaterHackathon. Essentially over 48 hours we hacked Ushahidi, adding a triage system for reports. It was fun. Part of my PhD research is about the design and usage of tools in the space of community mapping. People like Muki Hacklay and his extreme citizen science group have made some massive in-roads into verifying the quality of OSM data, with very positive results for OSM. But this doesn’t say anything about the quality of the tools. In a previous life I trained as a computer programmer, I moved into geospatial afterwards, hence it leaves me with a few questions regarding the process and its output.

Designing Processes and Workflow

The software engineering rulebook was thrown out of the window with good enough is perfect taken as the mantra. Although I’m a great believer in the idea that something good and tangible can be created out of an all night – look at any of my coursework from my comp sci bachelors – I’m concerned about the sustainability of the code that it generates. While I have every intention of using the code from the hackathon in something very useful the design came about from a 15 minute brief with around nine whiteboards being used and data structures made on the fly. This especially applies to the SMS solution we wrote, it bellies the importance of good documentation.

From all of this, I would ask the following questions…

  1. How sustainable is the code developed at a hackathon?
  2. Are there any examples of code developed at hackathons being deployed directly with modification?
  3. What steps are needed before hackathon code can be used?
I’m not attacking the hackathon methodology – I do think hackers should be remunerated with some small token of appreciation though – I just question how positive the impact can be if the code isn’t documented. The hackers that came were either students or good industry programmers with a humanitarian interest. Realistically it isn’t feasible for them all to continue working on the project, this is where open-sourcing it all comes in.
From this approach, how would a community be built upon a good project dropping out of a hackathon. Part of our approach meant building upon an already existing open source application; Ushahidi. Should we now consider pushing our edits to the Ushahidi codebase? This has issues as we didn’t create a plugin, we dived right into the code. This was due to the unfamiliarity of hacking Ushahidi on the whole – This also makes a point about decision making, in hindsight with knowing the system better we collectively made a bad design decision.

Being the effect/de facto project/product manager I found the need to keep my team productive involved constant resupply of pizza, coffee and biscuits. The idea being that content and happy coders are productive ones. With the results this is definitely true. It also helped that the team was awesome, throwing themselves into the project with great gusto. Again in hindsight I wonder if a more considered approach was necessary, thinking about sustainability and reuse. We did follow the mantra of JFDI (look it up on urban dictionary), but in forging a massive path ahead have we inhibited further growth and expansion?

Are there design considerations and best practices to follow in hackathons, effectively creating good code that could be reused and used effectively i.e. ergonomics. If not where would be a good place to start?

Written and submitted from the Nottingham Geospatial Building (52.953, -1.18405)

Developing Taarifa; London Water Hackathon 2011

On Friday 21st of October I attended my first hackathon. For those unfamiliar with hackathons they’re events with the idea of sticking a load of intelligent and driven computer hackers and programmers in a room for a period of time and see what drops out. The format of the London Hackathon was that for 48 hours at UCL groups would produce technology demonstrators and designs to solve global water problems.

The execution follows an introduction by the organisers framing the exercise with problem statements and presentations by experts wanting to solve different problems, people skyped in or did them in person. The problems presented all came form the Random Hacks of Kindness website and ranged from a water trading platform to my own issue of public service infrastructure and community mapping – my PhD.

Once the ideas where presented it was up to the developers to choose their projects. I had some awesome and – as my friend Josh Goldstein would say – ‘pumped’ developers with skills ranging from PHP to Python. The idea is essentially ‘Fix My Street’/Open311 for less developed countries. From this I started to hoover up whiteboards like they were going out of fashion. On them trying to get a design spec on workflow through the potential system, how triaging of reports would work. We were left with some outstanding design to be done, but nothing too complicated; or so we thought at the time. The idea was to create this system so it can be used as a technology demonstrator to the Tanzanian Commission of Science and Technology (COSTECH).

When I was last in Tanzania, COSTECH was being bounced around by the IT community. In Nairobi, the iHub acts as a lightning rod for technology and IT supporting local and foreign developers. The iHub model is excellent and really drives technological innovation in East Africa, other initiatives include the Hive Colab in Kampala and Bantalabs in Senegal. The people that run the iHub also run Ushahidi. Ushahidi is a platform/CMS designed for the crowdsourced reporting of issues, its inception was due to the Kenyan election crisis of 2008 and since has been used to report on Flamingos to the recent ‘occupy’ movements. Under the bonnet it’s a PHP website developed with the Kohana framework. It was here where we hit our first issue; none of our developers had worked with Kohana before. We split into two groups, one figuring out Kohana, the other designing workflow.

The first issue faced by the newly developed team was developing environment and server access. While most of the developers (myself included) were familiar with the programming languages needed, it wasn’t our primary choice. Personally I haven’t hacked about with PHP since October – December 2005, for others the experience seemed the same. Installing Ushahidi also proved problematic. Issues with mod_rewrite and other PHP extensions were experienced, but were eventually fixed. This wasn’t a problem which Ushahidi per se, though due to some very strange address rewriting configuring mod_rewrite was necessary for all. As we didn’t have server access we remote hosted, using one of the developers personal servers.

Once the workflow was sketched out we presented back to the developers, who had been exploring the Ushahidi plugin ecosystem. We integrated ‘actionable’, ‘simple groups’. Actionable was used to ‘action’ reports, and place them in a triage system.  Simple groups was used to simulate a ‘team’ of fixers. Fixers was used generically as the people fixing the reported problems, however the dynamic of how this would be worked wasn’t considered at this stage. Currently in Tanzania, in Dar Es Salaam and Zanzibar a number of reporting/monitoring projects are in the pipeline. The next steps are to interface with local government, private companies and citizens to develop public services.

We started to have a good interface, with tasks and problems being received and triaged. The workflow started to come together with reports being verified, on verification being triaged, going to the imaginary team of fixers and finally reaching conclusion or dispute resolution. The tabs to accommodate these were integrated into Ushahidi.

Now reports were able to be triaged we focused on expanding the reporting mechanism. Out of the box Ushahidi supports reporting through a web-based form, twitter and through its mobile applications (iOS, Android, Java and Windows Mobile 6). It can interface with SMS gateways like FrontlineSMS, we wanted to use SMS due to the ubiquitous nature of feature phones in Africa that realistically can only use SMS as a form of reporting. Using SMS presented problems of geo-locating the messages. Theoretically it’s possible for the telecos to triangulate the position of the sender and supply a latitude and longitude but this isn’t practical over a 48 hour hackathon notwithstanding the ethical and privacy concerns.

The solution we came up with, kudos to Dirk Goriseen, was to create a 100m2 grid and have a 10 digit reference for each grid square that people could text in. This would be prefixed by a hash (#) then found through a regular expression in the submitted message. Obviously questions remain when implementing this on a large scale namely ensuring local people know what their code is and creating a reference system that conforms to the human geography not just the physical. Integrating the SMS into the system was more problematic. We found that FrontlineSMS on the face of it doesn’t work in the cloud – if that is wrong please correct me! So Caroline Glassberg-Powell (Caz, our PHP guru) phoned a friend of hers, Daniel Fligg to hack an SMS gateway together.

Now it’s about 2100. We’ve been working for roughly nine hours and out project and code is starting to take shape. Teutonic reinforcements then arrived, all of them with Android mobile phones, none of them came with any Android development experience but did come with a willingness to learn about developing the Android application. This process entailed downloading Eclipse, setting up the Android SDK etc. It was a process that I feel should have been simpler, some issues with devices not being recognised, Eclipse’s idiosyncrasies (like interfacing with GitHub) were encountered.

About two hours later, once an Android development environment had been setup on each of the laptops we forked the Ushahidi Android app on GitHub and started development. While the stock app is very good we wanted to add the functionality of being notified when your report changed its status. To do this, avoiding login and passwords we wanted to use a unique identifier. Android, within its SDK supports a unique identifier and has specific calls for it. We later figured out that it’s not implemented across all versions and devices, with some choosing to not make that part of the SDK available. Quite why this is the case, the rationale doesn’t seem clear and from a design aspect it seems quite stupid, but the situation is as it is so we hack!

Our solution was to create a 256bit random hash assigned when the user opens the application for the first time. This would then be included in the reports to identify which reports come from which device (with the potentially flawed logic of one person uses one device/one device is used by one person). So instead of seeing all the reports in the system you just see the progress of reports that you have submitted. Unfortunately this process went on into the early hours, at which point sleep was necessary.

None of us wanted to use the night buses to go home so we slept in the working space. We fashioned little cubicles from the whiteboards, with chairs providing a bed. The sleep came easy, but when the three that remained woke we were all quite cold, in need of some more sleep and shower. However that doesn’t detract from more hacking!

On waking up the majority of the web-based system was already completed. During the night the SMS gateway had been magically integrated during the night. What remained was obvious bug-fixing, tidying up code and starting documentation.

The Android app had started to encounter real difficulties,  it wasn’t compiling, even when reverting back to code which was known to be ‘fine’. This process continued from 0930 – 1400. The issues with Android turned out to be problems within the Eclipse development environment. The code compiled and ran – with our adjustments – for the first time during the presentation. It was good timing.

We presented the slides above to the judging committee. We won. The credit to this victory goes to the people involved in the project, because of the numbers I guess we had covered the most amount of ground. However this isn’t trying to detract from the Herculean efforts of the programming team and other participants in the hackathon. Participation in our project wasn’t strict, people flitted in and out on the basis of their skills, in all about 4/5 people could be considered to be core/’never say die, I’m here till the end’ members. But the entire project had around 10-12 people that scribbled on the board, gave advice or rolled up their sleeves and hacked a little. This blog is dedicated to all of them.

Photos from the event can be found on Flickr here: http://www.flickr.com/photos/markiliffe/sets/72157628083354348/

Written and submitted from the Nottingham Geospatial Building (52.953, -1.18405)

Water, Sanitation and Geography

I have been surprised with during the Tandale project with how community members are familiar with the geographic boundaries and extent of their community. On touring areas with community members, it was clear how they used geographic features to navigate. Also the administrative boundaries were formed through natural features like rivers, without being imposed upon by an outside force.

Previously when facilitating mapping (essentially “There isn’t anything here, go have a look”) the mappers would have a difficultly in collating the map and their own mental model. Here in Tandale reading of seemed to be a lot easier than when I have previously experienced.

Map reading is a difficult skill, essentially it starts with understanding that the map is an abstract representation of space. As a map is a representation of space; a visualisation of various elements is at the whim of the cartographer (in our case the esteemed people which write the map styles for OSM) and the data that the cartographer/surveyor collects.

The people of Tandale seem to have a spatial awareness down to a very precise art, using landmarks and features with which to demarcate areas. This also relates to official and unofficial landuse within Tandale. Because of the lack of formal solid waste collection, most of the waste is dumped in swampland or wasteland. Unfortunately these waste areas have no buffer with residential homes, further illustrating the potential for disease.

On a tour of the Sokoni sub-ward executive he spoke at length on sanitation and water security. The conversation turned to common diseases and illnesses within Tandale with Malaria and HIV unsurprising common. Also mentioned in the same breath was Typhoid and Cholera, which, according to the officer,  outbreaks are common. Looking at the state of sanitation and drainage this is very believable.

We believe that the first step to solving these problems is to have a map. Hopefully have enough data to give evidence of the problems, to both inside and outside the community. We have also mapped dumping grounds, formal and informal medical facilities, toilets, water points among many other things. Using the map as a basemap in Ushahidi instance allows for the community members to use the map that they have created. Now our focus turns to completing the feedback loop, so there is an interface for the reports. Funnily enough that’s where my PhD comes in…

Written and submitted from the City Style Hotel, Sinza, Dar Es Salaam (-6.47319,39.13199)