Literacy Assessment Using Mobile Technology

Literacy   Assessment  Using   Mobile  Technology   assessment  in  developing  countries.  Wagner   (2003)   identifies   three   major   areas   of...
3 downloads 2 Views 5MB Size
Literacy   Assessment  Using   Mobile  Technology  

assessment  in  developing  countries.  Wagner   (2003)   identifies   three   major   areas   of   concern   for   assessments:   size/scope,   efficiency,   and   cost.   He   aptly   names   his   method   for   reforming   international   assessments  “smaller,  quicker,  cheaper,”  and   defines  each  term  as  follows:  

Sarah  Muffly   The  Earth  Institute  –  Columbia   University   [email protected]  

Overview     Current   educational   monitoring   in   Haiti   is   insufficient   and   doesn’t   give   policy   makers   information   to   make   spatially   relevant   and   targeted  interventions.    The  lack  of  available   results   on   quality   of   literacy   and   numeracy,   key   indicators   of   primary   education,   mean   that   local   schools   don’t   have   basis   to   improve   and   don’t   allow   regional   and   national  ministries  to  optimize  support.         Globally,  India  and  the  east  African  countries   of   Tanzania,   Kenya,   and   Uganda   have   successful   models   for   school   report   cards   that   are   greatly   advanced   with   the   use   by   tools   for   rapid   tabulation   and   review   of   the   results.         This  same  model  could  be  applied  in  Haiti  at   a   lower-­‐cost   by   utilizing   the   cell   phone   network   and   new   mobile   monitoring   technologies.   This   could   be   carried   out   by   existing   employed   government   school   inspectors   to   ensure   a   more   accurate   monitoring,   provide   incentives   for   local   schools   and   communities,   and   support   government  and  funder’s  policy  actions.       Such   a   model   also   aligns   with   current   scholarship  on  effective  practices  for  literacy  

 

Smaller:  Assessment  methods  do  not  need   to   be   major   entrepreneurial   enterprises,   but   rather   just   robust   enough   to   answer   key   policy   questions   at   the   national   and   local  levels  […]   Quicker:   Literacy   assessments   need   to   be   completed   in   ‘real   time’   so   that   results   can   affect   policy   and   spending   in   the   ‘lifetime’   of   current   ministerial   appointments.  […]   Cheaper:   LDCs   [less   developed   countries]   cannot   afford   either   the   fiscal   or   human resources   costs   of   deep   involvement   in   highly   technical   assessment   exercises.   (Wagner,  2003,  p.  294).  

The   literacy   assessment   project   described   in   this  paper  follows  a  method  that  adheres  to   these   guidelines   in   that   it   is   done   at   the   local   level,   disseminates   results   rapidly,   and   is   low-­‐cost.   Most   notably,   it   makes   use   of   mobile  technology  to  record  and  disseminate   results  that  is  reliable  and  easy  to  use.  

Global  examples  of  monitoring   literacy  and  numeracy   Assessment  Case  studies:  India,   East  Africa     Literacy   assessment   endeavors   in   India   and   East   Africa   provide   applicable   examples   for   assessment  in  Haiti.     In   India,   the   Annual   Status   of   Education   Report   (ASER)   conducts   household-­‐based   reading   tests   for   children.   Their   method   serves   to   complement   existing   measures   of   monitoring   progress   in   schools.   The   Indian   government’s  District  Information  System  for   Education   (DISE)   releases   report   cards  

containing   general   information   on   schools,   teachers,   and   students,   such   as   the   state   of   school   infrastructure,   funding,   and   the   figures   on   student   enrollment   and   grade   repetition   (Banerji   and   Wadhwa,   2010).   These   data   are   disaggregated   by   state,   district,  block,  and  cluster,  but  “independent,   national   data   is   scarce,   especially   on   children’s   achievement”   (ibid).   ASER’s   method   is   independent   from   DISE,   and   also   collects   data   on   out-­‐of-­‐school   children.   One   of  the  goals  of  ASER  is  to  survey  households   in   each   rural   district   in   India   (Banerji   and   Wadhwa,   2010).   In   2012,   Results   16,166   villages   were   show  5   surveyed,   in   which   levels  of   596,846   children   in   reading   331,881  households  were   ability:  no   assessed  (ASER,  2012).   reading,   Below   are   the   results   of   letters,   the   2012   literacy   test   in   words,   Hindi.   The   table   provides   paragraph,   an   example   of   how   test   story   results   can   be   organized   and  shared.     Table  1:  2012  ASER  literacy  test  results  by   grade  (Std.)  

plurality   of   children   assessed,   there   are   still   nearly   20%   of   children   who   can   only   recognize   letters,   which   may   lead   some   educators   or   administrators   to   devote   more   effort  toward  getting  children  to  read  words.     Another   educational   organization   following   this   method   of   assessment   is   Uwezo,   which   operates   in   Tanzania,   Kenya,   and   Uganda,   with  the  goal  of  increasing  children’s  literacy   and   numeracy.   Uwezo   conducts   the   same   kind   of   household-­‐based   assessments   that   ASER   does   in   India,   using   the   same   test   format   (the   content   is   translated   and   edited   or   rewritten   for   context).   In   addition   to   administering   the   tests,   Uwezo   writes   annual   reports   on   the   findings   of   the   assessments,   one  for  each  of  the  three  countries  surveyed,   and   one   comprehensive   East   Africa   report.   The   2012   East   Africa   Report   notes   that   in   2011,   350,000   children   from   more   than   150,000   households   were   assessed   (Uwezo,   2012).       The  report  states  that  across  East  Africa,  32%   of  children  in  grade  3  passed—meaning  they   could   read   at   “story   level”—the   Kiswahili   literacy  test,  and  only  16%  passed  the  English   literacy  test.  

Use  of  Mobile  Phones  in  Data   Collection  

  Source:  ASER  2012     The   ASER   method   allows   relevant   stakeholders   and   actors   in   education   to   easily   understand   how   children   are   reading   in   broad   terms.   For   example,   while   “story   level”   competency   claims   the   greatest  

    The   Earth   Institute’s   unique   approach   to   literacy   assessment   incorporates   cutting-­‐ edge   mobile   technology   into   the   process   of   recording   and   sharing   results.   Consequently,   communication   and   action   surrounding   test   results   happens   more   quickly   and   has   the   potential  to  serve  the  very   EI  uses   children  who  are  assessed.   mobile     phones  to   Using   mobile   phones   for   collect  data   data   collection   in   developing   countries   is  

not   a   new   concept,   and   has   been   done   in   many   places   by   various   organizations.   DeRenzi   et   al   (2007)   make   the   point   that   using   mobile   phones   enables   field   workers   to   “stay   in   disconnected   environments   for   extended   periods   of   time”   (DeRenzi   et   al,   2007,   p.   1),   as   workers   can   record   data   on   the   phones   and   then   back   up   or   transmit   them   when   in   a   connected   environment.   Simply   put,   mobile   phones—especially   smartphones,   which   are   essentially   small   computers—are   a   convenient,   reliable   tool   to   collect   and   store   data   when   in   the   field:   they   eliminate   the   need   to   carry   around   sheaves   of   paper,   pens,   and   other   materials   that  are  easily  lost  or  damaged.     Many  documented  instances  of  using  mobile   phones   for   data   collection   involve   research   in   the   health   sector   (DeRenzi   et   al,   2007,   Tomlinson   et   al,   2009),   but   there   is   not   a   great   deal   of   equivalent   research   in   the   education   sector.   There   is,   however,   a   growing   interest   in   using   mobile   technology   for   learning,   called   m-­‐learning   (Traxler   &   Kukulska-­‐Hulme,   2005,   Barker   et   al,   2005).   Using   mobile   phones   to   conduct   literacy   assessments,   as   described   in   this   paper,   is   an   innovative   way   of   combining   the   known   effectiveness   of   mobile   data   collection   with   the   concept   of   m-­‐learning.   This   is   because   assessments  are  more  than  data  points;  they   serve   to   show   educators   and   administrators   how  and  what  children  are  learning.  

Education  Monitoring  in  Haiti:   What  is  being  done,  what  can  be   improved   Current  Educational  Monitoring  in   Haiti     The  MENFP’s  2010-­‐2015  Operational  Plan  for   the   education   system   outlines   its   method   for   monitoring  and  evaluation  in  schools.  

  Large-­‐scale   educational   assessments   in   Haiti   take   the   form   of   summative   exams   to   progress   from   Fundamental   Current   schooling   (primary   and   junior   assessment   secondary)   to   secondary   school,   methods   and   again   to   graduate   from   are  time-­‐ secondary   school.   Both   types   of   consuming   exams  are  given  at  the  end  of  the   school  year,  which  means  that  any   information   useful   for   school   improvement   can   only   be   implemented   during   the   following   school   year   (at   the   earliest)   and   often   does   not   benefit   the   actual   students   who  took  the  exams.     The  Operational  Plan  also  calls  for  reinforcing   current  means  of  monitoring  and  evaluation,   as   well   as   devising   new   methods   for   it.   The   goals   include—among   several   others— making   available   academic   statistics   for   the   whole  country  on  an  annual  basis,  having  all   subsectors   of   education   covered   in   the   MENFP’s   database,   and   creating   a   documentation   center   for   the   Ministry   (MENFP,  2010,  p.  25).       The   MENFP   will   organize   several   different   agencies   to   put   these   M&E   activities   into   effect,   including   Departmental   Directorates,   School   District   Offices,   and   Zone   Inspection   Offices.  In  other  words,  some  M&E  activities   will   be   managed   at   relatively   local   levels.    

The  Earth  Institute  Method:  Pilot   Project     Since   current   and   planned   activities   to   monitor  educational  quality  in  Haiti  still  place   heavy  emphasis  on  end-­‐of-­‐school  exams  and   other   annual   forms   of   evaluation,   a   desirable   shift   would   be   for   future   assessment   endeavors   to   align   with   Wagner’s   “smaller,   quicker,  cheaper”  model  while  also  collecting   much-­‐needed   information   about   other  

elements   of   education,   such   as   student   and   teacher   attendance,   sex   ratios,   and   age   of   students.     The  Earth  Institute’s  literacy  assessment  pilot   project   aimed   to   address   these   issues,   especially   that   of   “quicker”   assessments.   In   the   Port-­‐à-­‐Piment   Watershed   project,   in   early   summer,   2012,   members   of   the   Education   Team   piloted   a   new   literacy   assessment   method   in   the   watershed.   In   total,  50  children  were  assessed,  40  of  whom   were   enrolled   in   traditional   schools   (either   public   or   private)   and   10   of   whom   were   enrolled   in   a   new,   special   program   attached   to  a  public  school  that  caters  to  children  who   have  not  previously  attended  school.       The   tests   used   follow   the   ASER   format.   What   made   the   testing   in   Haiti   unique,   however,   was   the   way   in   which   the   results   were   recorded   and   Mobile   disseminated.   The   Earth   technology   Institute   uses   formhub,   used  for   an   online   platform— faster  data   created   by   the   Modi   collection   Research   Group   at   and   Columbia   University—for   processing   authoring   surveys   that   can   then   be   uploaded   to   Android-­‐powered  mobile  phones  and  used  to   collect  data.  The  surveys  for  the  Haiti  project,   for   example,   recorded   information   about   children’s   age,   sex,   enrollment   status,   and   their   performance   on   the   literacy   test.   Once   the  phones  have  access  to  wireless  Internet,   the  surveys  can  be  sent  electronically  back  to   formhub,   where   the   data   can   be   organized   and   disseminated.   This   type   of   technology   allows   for   automated   generation   of   report   cards,   which   means   that   data   from   assessments   can   be   used   nearly   instantly,   thereby   benefiting   not   just   school   administrators   and   policymakers,   but   potentially   the   students   themselves   who   take  the  tests.  

  Implementation  model   There   are   two   basic   models   for   carrying   out   this   type   of   assessment.   One   option   is   household-­‐based,   in   which   testers   assess   children   in   their   homes,   with   family   members   present   to   observe   the   process.   The   other   typical   model   is   school-­‐based,   in   which   testing   happens   as   part   of   regular   school  activities.     The   Earth   Institute   pilot   project,   because   it   worked   with   a   relatively   small   sample   size,   was  able  to  merge  elements  of  these  models.   The   tests   were   administered   in   schools   to   students   who   had   been   randomly   selected   ahead  of  time  by  school  directors.  The  school   directors   notified   these   students’   families   ahead   of   time,   so   that   a   parent   or   other   family  member  could  attend  and  observe  the   assessment.   Each   school   director   chose   5   students   in   the   3rd   year   of   the   1st   cycle   of   Fundamental   education   (primary   school).   Additionally,   the   Education   Team   met   with   Parent-­‐Teacher  associations  to  introduce  and   explain  the  process.     Testing   The   team   went   to   schools   at   specific   times   arranged   by   the   school   directors   and   by   the   Education   Team.   Many   Family   children   were   members   accompanied   by   their   observed   mothers,   but   others   were   testing     accompanied   by   fathers,   grandparents,   or   in   some   cases,   by   siblings.   Children   were   assessed   one   at   a   time,   in   order   to   make   them   less   nervous   and   distracted.   Testing   spaces   included   the   director’s   office,   an   empty   classroom,   the   yard   behind   the   school,   or   in   one   case,   a   closed-­‐off   corner   of   a   one-­‐room   schoolhouse.   For   each   child   and   chaperone,   a   member   of   the   Education   Team   would  

introduce   us   as   a   group,   explain   what   the   test   was   and   how   it   worked,   and   chatted   with   the   child   to   make   him   or   her   feel   at   ease.  Then  one  of  them  would  ask  identifying   questions:   name   of   the   child,   name   of   the   child’s   mother   (or   the   name   of   the   chaperone   present),   and   the   child’s   age.   Once   this   information   was   recorded   on   the   phone,  the  tests  were  given.     The  tests  and  method  of  administering  them   were   adapted   from   ASER’s   practices.   Students   were   asked   to   read   aloud   from   a   short   test   consisting   of   a   short   story   (about   ten   sentences),   two   short   paragraphs   (four   sentences   each),   a   series   of   words   (three   to   four   letters   each),   and   a   series   of   letters.   The   tests   were   administered   individually.   First,   one   of   the   CRS   animators   held   the   test   in   front  of  the  child  as  he  or  she  read  the  short   paragraphs   aloud.   The   test   was   folded   in   half   so  that  only  the  paragraphs  are  visible.       If   it   was   clear   that   the   child   could   read   the   paragraphs,   the   tester   flipped   the   paper   so   that   only   the   story   is   visible;   the   child   was   then   asked   to   read   the   story.   If   the   child   could  successfully  read  the  story  and  answer   basic   comprehension   questions   about   it,   he   or  she  was  marked  at  “story  level.”  If  not,  he   or  she  is  marked  at  “paragraph  level.”       If   the   child   could   not   read   the   paragraphs— or   read   them   with   clear   difficulty   and   did   not   truly  understand  them—he  or  she  was  asked   to   read   the   series   of   words.   The   tester   repeated   the   action   of   folding   the   test   in   half   so   that   at   first   only   the   words   are   visible.   If   the   child   could   read   the   words,   he   or   she   is   marked   at   “word   level.”   If   not,   the   tester   flipped   the   test   to   the   other   side   and   asked   the  child  to  read  the  series  of  letters.     Results   The  results  of  the  assessment  covered  a  wide   range   of   reading   aptitude.   In   some   schools,  

the   majority—if   not   all—of   the   children   could   read   at   some   level,   but   these   children   were  also  often  overage  for  their  grade  level.   Furthermore,   at   other   schools,   Range  of   the   majority   of   children   chosen   results;   could   not   read   at   all—or   could   majority   read   only   letters—regardless   of   could   age.   There   were,   however,   some   read  story   schools  where  the  majority  of  the   children   were   able   to   read   the   paragraphs   or   the   story.     Of   the   fifty   children   assessed,  14%  were  marked  at  “no  reading,”   but  56%  were  able  to  read  the  story.  

Implications     What   the   pilot   project   shows   first   and   foremost   is   that   this   type   of   literacy   assessment  is  fast,  reliable,  and  easy  to  use.   The   technology   in   question,   while   sophisticated,   is   accessible   to   the   multiple   users.   Therefore,   anyone   from   teachers,   to   school   directors,   to   regional   educational   inspectors   could   be   trained   to   use   it   and   to   conduct   assessments.   It   is   also   a   low-­‐cost   and   rapid   tool.   The   phones   needed   to   conduct   assessments   need   only   be   purchased   once,   and   the   surveys   can   be   created  on  any  computer.       Moreover,   because   tests   can   be   administered   multiple   times   during   a   school   year,   the   Report   Cards   can   be   used   to   monitor   students’   progress   throughout   the   year;   their   information   Frequent   thus   benefits   teachers   testing  =   and   students,   as   well   more   as   officials.   The   opportunities   frequent   testing   could   for   be   a   means   to   measure   improvement     the   effects   of   certain   policy   or   curriculum   changes   in   schools.   The   regular   assessments   can   also   serve   as   a   method   of   monitoring   student  attendance  and  related  details  (such  

as   age   and   sex   of   children   in   school)   and   teacher   presence.   Consequently,   the   assessment  process  can  evaluate  educational   quality   in   terms   of   learning   as   well   as   other   relevant  indicators.    

Scaling  up     The  Earth  Institute  is  already  scaling  up  these   types   of   assessments   in   other   locations.   In   partnership   with   EI  is   Uwezo,   assessments   conducting   have   been   conducted   assessments   in   the   Millennium   in  other  MVP   Villages   of   Mbola,   in   Tanzania,   Ruhiira,   in   Uganda,   and   Dertu   and   Sauri,   in   Kenya.   Additionally,   a   quasi-­‐experimental   study   using   the   mobile   technology   method   is   currently   underway   in   the   Millennium   Village   of   Bonsaaso,   in   Ghana.   Furthermore,   these   assessments   are   administered   to   five   students   per   month   as   part   of   regular   M&E   surveys  in  schools.     As   for   Haiti,   given   the   ease   and   low   cost   of   acquisition   and   use   of   this   technology,   this   type   of   project   has   great   potential   to   be   scaled   up   to   the   whole   country.   Such   an   endeavor   could   be   a   solution   to   the   lack   of   consistent  monitoring  of  schools  in  Haiti.       For   example,   the   country   already   has   a   system   of   regional   and   zonal   inspectors,   responsible  for  monitoring  public  and  private   schools.     These   regional   school   inspectors   could  be  easily  trained  in  the  survey  method   in  order  to  administer  tests  in  all  the  schools   in   their   region.   Report   Cards   would   then   be   generated   from   the   results,   which   could   be   transmitted   to   other   Ministry   of   Education   officials   and   back   to   the   school   administrators.   If   multiple   inspectors   conducted   tests   in   their   regions,   the   Report   Cards   could   serve   as   a   means   of   comparing  

performance   across   regions   and   schools   in   the  country.       Such   a   scale-­‐up   would   require   a   certain   amount   of   advance   planning,   including   developing   tests,   frequent   communication   between   MOE   officials   and   school   directors   (and   possibly   teachers),   buying   and   programming   mobile   phones,   and   training   the   relevant   actors   in   the   process.   Once   the   preparation   is   completed,   however,   the   method  is  self-­‐sustaining,  especially  because   some   of   the   costs   are   one-­‐time   expenses.   That   is,   after   the   initial   preparation,   assessments   can   be   conducted   multiple   times.     Overall,  this  kind  of  endeavor  aligns  with  the   MENFP’s   plan   for   gathering   academic   statistics   and   using   local   government   agencies  to  carry  out  M&E  tasks.  

Conclusion   The   type   of   literacy   assessment   method   described   here   is   useful   in   two   main   ways:   first,   it   is   a   fast,   easy,   and   low-­‐cost   Assessments   way   to   measure   measure   children’s   reading   learning  as  well   skills.   Second,   it   can   as  other  M&E   serve   as   a   tool   for   indicators   monitoring   school   quality,   not   only   as   far   as   learning   is   concerned,   but   in   terms   of   other   important   factors   such   as   enrollment,   attendance,   teacher   presence,   age   and   sex   of   children.   For   Haiti,   this   endeavor   has   the   potential   to   complement   and   promote   existing   M&E   activities   in   schools.   Additionally,   it   can   improve   educational   quality   for   the   children   who   are   assessed   by   attending   to   their   reading  needs  in  real  time.        

   

 

Appendix     A)  Screen  shots  of  the  phone-­‐based  survey  

                             

 

  B)  Photos  from  pilot    

                   

 

                 

 

C)  Sample  Report  Card      

 

       

 

 

Bibliography     ASER.  (2012).  Enrollment  and  learning  report  card.  Retrieved  from     http://www.asercentre.org/education/India/status/p/143.html     Banerji,  R.  &  Wadhwa,  W.  (2010).  Annual  status  of  education  report  2005-­‐2010:  Key     features  of  ASER.  Retrieved  from  www.asercentre.org     Barker,  A.,  Krull,  G.,  &  Mallinson,  B.  (2005,  October).  A  proposed  theoretical  model  for  m-­‐   learning  adoption  in  developing  countries.  In  Proceedings  of  mLearn  (Vol.  2005,  p.   4th).     DeRenzi,  B.,  Anokwa,  Y.,  Parikh,  T.,  &  Borriello,  G.  (2007,  August).  Reliable  data  collection  in     highly  disconnected  environments  using  mobile  phones.  In  Proceedings  of  the  2007   workshop  on  Networked  systems  for  developing  regions  (p.  4).  ACM.     Ministère  de  l’Education  Nationale  et  de  la  Formation  Professionelle  (Ministry  of  Education     and  Professional  Training).  (2010).  Vers  la  refondation  du  système  éducatif  haïtien:   Plan  opérationnel  2010-­‐2015  des  recommendations  de  la  commission  présidentielle   éducation  et  formation  (Towards  the  overhaul  of  the  Haitian  educational  system:   Operational  Plan  2010-­‐2015  from  recommendations  from  the  presidential  education   and  training  commission).  Port-­‐au-­‐Prince,  Haiti:  Ministère  de  l’Education  Nationale   et  de  la  Formation  Professionnelle.     Traxler,  J.,  &  Kukulska-­‐Hulme,  A.  (2005).  Mobile  learning  in  developing  countries.     Uwezo.  (2012).  Are  our  children  learning?  Literacy  and  numeracy  across  East  Africa.     Retrieved  from  http://www.uwezo.net/publications/reports/     Wagner,  D.  (2003).  Smaller,  quicker,  cheaper:  alternative  strategies  for  literacy     assessment  in  the  UN  Literacy  Decade.  International  Journal  of  Educational   Research,  39,  pp.  293–  309.      

Suggest Documents