Home
        Document in Adobe Acrobat format
         Contents
1.    Measuring Adult Victimization in Alaska   Technical Report    Prepared for the  Bureau of Justice Statistics    by the    Alaska Justice Statistical Analysis Center  Justice Center  University of Alaska Anchorage       February 2002  JC 20   0109 011    Measuring Adult Victimization in Alaska   Technical Report    Prepared for the  Bureau of Justice Statistics    by the    Alaska Justice Statistical Analysis Center  Justice Center  University of Alaska Anchorage    February 2002  JC 20   0109 011    This research was supported by Grant No  2000 BJ CX K013 awarded by the  Bureau of Justice Statistics  Office of Justice Programs  U S  Department of Justice     Contents    Tntroductigi ised ate wath see nad ee etn Late eaten hd Se eed Ge and ew ad eee 1  Problems Encountered with the CVS Software        0 0    cece eee een eee 1  Ma Or Activites 2 ica ceed ele ta sg  amp  ions aed aces Re SU AS a eS VARS HEARD EN ORS He 3  Purchasing Software aner oera Dale E A tee Be E A D A E E a AR ee 3  Programming the Victimization Survey          onananaa kB Sig A 0 Sie OE a 4  Testing the Victimization Survey Structure cx PR RE ER ER AE ERROR REE    Developing a Dataset Reduction Strategy         00 cc teens 7  PLOUUCES enera aa ees a o a a aa i a pine eee ean ER O eA 9  Appendix A  Problems Associated with the BJS  Crime Victimization Survey Software Program          0 0    nunna unearen 11  Appendix B  Administering the Victimization Survey             0    cee eee eee eee 14    Appe
2.    yes    box has been checked and a specify text box has  appeared               In order for the survey to work as programmed  the following format must be adhered to when  entering incidents     l brief description  2 brief description    10 brief description    Each incident described will be numbered in sequential order with a maximum of 10 incidents  recorded  Descriptions of the incident must be kept brief since you are limited to a single line of  text  Keep in mind that the descriptions entered are for the interviewer only  The descriptions assist  the interviewer when focusing the respondent on a particular incident  Thus  feel free to use  shorthand or any other method that is brief but allows users to recall the reported incident    Consider the following example  The respondent indicates that both her purse and bicycle were  stolen  The interviewer probes to determine whether the thefts were two separate incidents or part  of a single incident  The respondent indicates that they were part of the same incident  Thus  we  will enter a description as follows on a single line in the text box  see screenshot below      1l purse bicycle stolen     25               If this was the only incident reported for this screen question  press OK and proceed to the next  question  What if  however  the two    crimes    reported were actually separate incidents  Use the  following syntax  see screenshot below      1 purse stolen  2 bicycle stolen              Returning to the fir
3.   age of 18 because of the potential difficulties in obtaining parental consent to interview a minor and  the potential unreliability of using a proxy interviewee  In addition  for practical reasons  the survey  is constructed to capture responses from a single household member  Thus  unlike the NCVS where  all household members age 12 and older are interviewed  the Justice Center survey focuses on a  single household member age 18 or older  The primary consideration here was twofold  First  the  Center speculated that trying to secure interviews from all adult household members would be  extraordinarily difficult  particularly when one considers that most of the calls made to household     5     result in callbacks anyway  Second  the increase in time spent interviewing multiple household  members would greatly extend data collection    The second set of changes can best be described as content modification  These changes are  related to questionnaire ordering and question inclusion or exclusion  These changes are described  below     e Demographic questions were moved to the end of the survey  As such  the survey begins  with a few brief qualifying questions and proceeds directly to the victimization screening  questions  The Justice Center decided to make this change only after beginning the actual  administration of the survey  Respondents were typically most upset about being asked  personal question and  on several occasions  terminated the questionnaire at that point  
4.  and refused options  Ci3 exports the answer to this question as four  separate variables  each taking a value of 1  present  or 2  absent   The four variables would be  named the following     Variable 20_1 corresponds to the    yes    option  20_2 to the    no    option  20_3 to the    don   t know     option  and 20 4 to the    refused    option  If 20_1 assumed a value of 1 while the other three took  the value of 0  the respondent   s answer to question 20_ 1 would have been    yes     20_1 was present  thereby indicating that it was marked while the other three variables were absent indicating that they  were not marked     This method of exporting the data is certainly warranted for answers that require multiple  responses  however  the number of variables unnecessarily increases for single response answers   Thus  the Justice Center created an SPSS syntax file that would combine variables from single   response questions into a single variable  The new variable would simply be called 20 and would  take one of four values  yes  no  don   t know  refused        Of course  the skips patterns which exist throughout the questionnaire reduce the number of questions actually  asked of respondents       Those questions that were eliminated were irrelevant to a victimization survey of residents of the Municipality of  Anchorage  e g   questions about farming       8     The SPSS syntax file generated does more than simply recombine multiple variables into single    variables  
5.  box at the bottom of the screen will assist you  in recalling how many incidents are already reported  In our example  a single incident has been  reported  Therefore  we will describe the break in using the number two  see screenshot     You will again realize the importance of asking whether the break in incident and the  purse bicycle theft incident were one in the same  If they were one in the same and they were not  acknowledged as such  you would have generated two incident reports when  in fact  you would  have only needed one  the single report would capture the break in and thefts   Be sure to determine  whether the incidents are the same or not      28     m           After recording the respondent   s answer  hitting OK and proceeding to the next question   you  will be asked to enter the number of times the incident occurred  Enter the appropriate answer and  proceed to the next screen question  For the sake of consistency  enter one for the number of times  this incident occurred    In our example  the next question deals with the number of cars owned  Assuming any answer  other than    none    is selected  you will be presented with a question about vehicle thefts  Note the  incident count at the bottom of the screen in the screenshot below  Two incidents have been  recorded  purse bicycle theft  break in                Follow the same procedures for describing events if incidents are reported  In our example  we  would begin numbering with three since two inci
6.  four months  identifying flaws and offering suggestions   They checked the logic of every question  verifying skip patterns against paper copies of the  National Crime Victimization Survey   s NCVS 1 and Incident Report  After each survey run   through  they would provide a list of errors suggestions enhancements to the principal investigator   Together the principal investigator and the research assistants engaged in a cyclical process of  program  test  suggestion  reprogram  retest  and so on  This process was extremely advantageous   The most difficult part of developing this survey on the Sawtooth platform was ensuring that the  skip patterns remained and reflected the patterns that existed in the NCVS instrument  The work  of the research assistants was critical to preserving the integrity of the NCVS  By the end of March  2002  the testing phase had run its course  The Justice Center felt confident that the survey was  ready for administration     Developing a Dataset Reduction Strategy    Even after completing the programming and testing phase of the project  the Justice Center was  still faced with a considerable challenge resulting from the complexity of the survey  Although the  survey constructed was more than sufficient for data collection  it was quickly determined that any  dataset created during the administration of the survey was going to be massive  the number of  variables in exported SPSS file was 13 014  Thus  the Justice Center had to develop a procedu
7.  get into a hotel or motel room or  vacation home where you were staying  Did any incidents of this type happen to you     3  V39ab  During the last 12 months   other than the incidents already mentioned   was the  vehicle were any of the vehicles stolen or used without permission  did anyone steal any  parts such as a tire  tape deck  hubcap or battery  did anyone steal any gas from  it them    or did anyone ATTEMPT to steal any vehicle or parts attached to  it them      4  V40ab  Other than the incidents already described  in the last 12 months were you  attacked or threatened OR did you have something stolen from you at home including  while on the porch or in the yard  at or near a friend   s  relative   s  or neighbor   s home     8 The experiences of the Justice Center seemingly indicate that ten incidents is more than enough when using a 12   month reference period and when interviewing only one member of each household  In fact  in a survey of over 800  respondents  only a few respondents reported over three incidents and no respondent reported more than five incidents      22     at work or school  in places such as a storage shed or laundry room  a shopping mall   restaurant  bank  or airport  while riding in any vehicle  on the street or in a parking lot   at such places as at a party  theater  gym  picnic area  bowling lanes  or while fishing or  hunting  OR did anyone ATTEMPT to attack or ATTEMPT to steal anything belonging  to you from any of these places  Did a
8.  in Anchorage  one of which was to  conduct a local crime victimization survey  After learning that the Justice Center had received  funding to conduct a similar project  the Women   s Commission inquired about the possibility of  adding certain questions to the survey that addressed the commission   s key concerns  The Justice  Center willingly agreed to assess the feasibility of adding questions to the prepackaged survey   However  a key problem arose  While the software is advertised as easily customizable  the actual  program is difficult to modify  The principal investigator was told by BJS staff on several occasions  that survey modification requires extensive knowledge of database design  generally  and Microsoft  Access  specifically  Furthermore  they indicated that modifying question phrasing was relatively  easy while altering the arrangement of questions and adding deleting questions was more  challenging  One staff member advised against modifying the software  As a result of these  admonishments  the Justice Center sought alternative methods of including additional questions  without needing to alter the software and  potentially  compromise the integrity of the software  e g    paper addendums     Unfortunately  the difficulty in modifying the software was only one problem encountered when  preparing for the survey  Appendix A details the problems within the software program  The      For example  the first question in the survey asks the respondent to specify 
9.  more than enough to record open ended responses            Numeric  Numeric questions require interviewers to type in some numeric value  Although  similar to open ended questions  numeric questions are different in one important respect  Numeric  questions are treated as a variable in SPSS outputs rather than a line in a text file  To enter data      16     interviewers type the respondents    numerical answer  Most numeric questions have restrictions that  dictate the precise value that can be entered  For example  questions asking about year of  occurrence restrict users to entering a four digit number  The example below illustrates a numeric  question type  Note that values are assigned for    don   t know    and    refused    answers  If a  respondent provides one of these two answers  interviewers would key either 7777 for    don   t know     or 9999 for    refused           Other  Other questions are not truly question types  Instead  they can be considered an  enhancement to the single  and multiple response question types  When the other command is  programmed  a text box appears when certain answer choices are selected  For example  many  questions include an    other    option  When this choice is selected  a text box appears allowing  interviewers to specify exactly what was meant by    other     Please note that the chosen answer  choice is exported as part of an SPSS file while the open ended specify portion of the question is  exported as part of a text file  
10.  software   s sample answer    3 is exported  the program assigns a crime code of 37  Page 270 of the user   s manual does not  list a crime code of 37     oe    Appendix B  Administering the Victimization Survey    This section describes the procedures for administering the crime victimization survey once a  potential respondent consents to participating  The primary focus is on identifying conventions  adopted during the programming phase  describing the process of navigating through the survey  and  generating victimization reports  It is assumed that users are familiar with Ci3 and WinCati to the  extent that the survey syntax is loaded  CATI parameters established  and other preparatory  procedures tended to        Question Types    Information Instructional  Information instruction questions are screens in which no entries  are keyed and no data recorded  Interviewers read the information aloud or follow the instructions  on the screen without reading the remarks to the interviewers  The example below is an example  of an informational instructional question type  This screen  the first to appear when a survey is  commenced  instructs interviewers to read an introductory statement to the respondent from a hard   copy form at their workstation        Single response  Single response questions require interviewers to key a single answer from  a list of possible responses  Single questions are easily identifiable through the circular radio button  to the immediate left of 
11.  would be unequivocal  states  and local communities would finally have a data source that allows them to assess crime through  methods that complement official data sources  In addition  victimization survey results provide  communities with a rich data source that allows them to examine issues relating to community  policing  underreporting of crime  fear  and neighborhood quality of life  Certainly  this information  would be equally valuable to Anchorage  Alaska  However  the Crime Victimization Survey  software  as it existed at the time of the original project proposal  was not a suitable platform from  which to conduct a local survey in Anchorage  This section briefly addresses key problems that the  Justice Center encountered while preparing to administer the survey    The Justice Center began preparations for the victimization survey in January 2001  Initially   the principal investigator familiarized himself with the software  Subsequently  significant amounts  of time were spent attempting to modify the existing survey so that it better fits the geography and  demography of the Municipality of Anchorage  In addition  attempts were made to broaden certain  questions in order to increase the overall survey response rate     Additional efforts to modify the software came after the Justice Center was contacted by the  Anchorage Women   s Commission  The organization had established a domestic violence committee  that made a series of recommendations concerning violence
12. As  a result  data were not collected about victimizations  community policing  or perceptions  of neighborhood  The survey allowed the Center to collect these data even if the respondent  decided to terminate during the demographic portion of the survey     e Reference period was extended from six months to twelve months  Unlike the NCVS  the  Justice Center survey was not going to be administered to the same household every six  months  Thus  the reference period was extended     e Respondents were not asked the county of victimization  NCVS Incident Report question  8a  since Alaska does not have county units     e Respondents were not asked if incident happened on an Indian Reservation  NCVS Incident  Report question 9      e Many of the    check item    questions from the NCVS were eliminated  However  the logic  behind them was programmed into the victimization survey thereby automating the process     e Business related questions  NCVS Incident Report questions 138b through 139  were not  included since they were only recent additions to the NCVS and the survey construction  process began by using questions from the BJS Crime Victimization Survey Software     e Disability health questions  NCVS Incident Report questions 167 through 176  were not  included since they were only recent additions to the NCVS and the survey construction  process began by using questions from the BJS Crime Victimization Survey Software     e Household questions about farm sales and Indian re
13. The example below is a multiple response question  evident by the  square check boxes next to each answer choice  with an added    other    command  If the interviewer  keys choice nine     other  specify      a text box appears  Interviewers type details in the box and click  OK     Vt        m  re      Question Text  amp  Interviewer Instructions    Each question is divided into two components where applicable  question text and interviewer  instructions  Question text  that is the text that is to be read to respondents  is enclosed in quotation  marks  Interviewers are to read question text verbatim in order to ensure that questions are asked  in a consistent fashion  Answer choices are only to be read if indicated in the instructional section    Interviewer instructions are enclosed in boxes with double lined borders  Instructional text is  not to read to respondents unless otherwise indicated  In the example below  the interviewer is  instructed to read answer choices  they are not to be read unless instructed  and mark all answers  that apply  The mark all that apply instruction is simply a reminder  Recall that square check boxes  also denote multiple response items thereby indicating that the interviewers may key all answers that    apply      18        Navigating Through the Questionnaire    Registering answers  Entering numbers or text or clicking answer choices does not result in  a registered answer  Answers are only recorded after they are registered  You can reg
14. The syntax adds variable labels  descriptive names  to all variables in the dataset     Moreover  it produces value labels for each variable  Most importantly  the syntax dramatically    reduces the size of the dataset by more than 50 percent from just over 13 000 variables to slightly    more than 6 000 variables        Products    Several products were borne out of this research project  The deliverables included as hard copy    documents and or electronic files on the enclosed CD ROM are designed to allow any individual    or institution with Ci3 software to work with the victimization constructed by the Justice Center     These products include the following     l     Survey syntax     The text file includes all of the programming for the victimization survey   Users can simply copy the contents of this document and paste them into a new  questionnaire window in the Ci3 software program  Upon compiling this programming  using the compile function in C73  users will be ready to administer this survey by creating  field disks or by preparing a CATI system     Syntax guide  The syntax guide is a printed copy of the survey syntax reduced to include  only a single incident report  remember  all incidents reports use the same programming with  different variable names   Moreover  the syntax guide describes the most commonly used  commands in the syntax  illustrates several examples  and explains other relevant  information relating to qualifying instructions and disposition cod
15. an in Spring 2001  The  Justice Center began by researching available CATI software programs and  upon careful  consideration and testing using demonstration software  it decided that Sawtooth Technologies        3     computer assisted telephone interviewing system was the best available program  The system  contained all of the typical CATI functions and  more importantly  was flexible in its operation and  the learning curve was not very steep  As aresult  the Justice Center purchased the Sawtooth system  in Spring 2001    The software is compatible with various question formats  including all of those included in the  original Crime Victimization Survey software  e g   single response  multiple response  and open   ended questions   Moreover  the software allows users to automatically fill in question text or  responses based on previous responses and skip questions that are not applicable  The software also  has database capabilities  therefore  all data gathered using this software will automatically be  recorded in a format that is compatible with common software applications such as MS Access  MS  Excel  and  most important for the Justice Center  SPSS    The system purchased from Sawtooth is actually two software programs in one  The first  Ci3  for Windows  is the questionnaire authoring application  The Justice Center programmed the  victimization survey using Ci3 software  The second application  WinCati  is Sawtooth   s CATI  software system  WinCati adds functio
16. ated or an extra report is deleted  This involves changing the  incident descriptions and the follow up  how many   questions as noted above    Continue answering questions until you reach the end of the survey  A closing screen will notify  you that the survey is complete  see screenshot   Be sure to click next  ALT N  or enter on this  screen to close out the survey     333     Appendix C  Exploring the Victimization Survey without C73 or WinCati    Readers of this technical report interested in viewing the victimization survey as it would appear  on interviewers    screens can do so on a limited basis without owning a site license for Sawtooth  Technologies    WinCati program  To view the survey  simply read the following description and  limitations and follow the instructions below    Recall that Sawtooth Technologies    Ci3 and WinCati programs are distinct yet interrelated  software applications  Ci3 may be purchased and used without WinCati but WinCati must be used  with Ci3  WinCati extends Ci3  the questionnaire authoring tool  by providing typical CATI  functions such as sample management  callback scheduling  disposition assignment  and other  related features  Nevertheless  Ci3 can be used for questionnaire authoring  data collection  and  analysis without employing a CATI system  Ci3 is able to produce field disks for data collection   Field disks produce stand alone survey administration interfaces on any computer  regardless of  whether or not the computer act
17. back to the screen question  where the purse bicycle description was entered  Click on next and alter the description  You might  say    purse bicycle stolen  break in    in the text box  Register your answer  Proceed forward to the  next unanswered question  You will notice that the counter on each screen question will only report  a single incident instead of two  assuming no additional incident was entered         10 Tt does not matter under which screen question incidents are reported  For example  a sexual assault can be  reported under a screen question for theft  an assault under a screen question for motor vehicle theft  and so on  The  questions are designed to elicit information on all incidents  The precise details will emerge in incident reports      30     If a single incident was determined to be two separate incidents  follow similar procedures  For  example  assume that our purse bicycle stolen incident was actually two incidents  Move back to  the screen question and change the description in the text box to reflect two separate incidents     1 purse stolen  2 bicycle stolen    Register the changed answer  Next  be sure to change the number of times the incident occurred  from one to two in the follow up  how many   question  Move forward to the next unanswered  question    At this point  we are at the point where we have recorded four incidents  purse bicycle stolen   break in  car stolen  car stolen   We will assume that no other incidents were reported in 
18. dents were already reported  Our respondent has     29     indicated that her car was stolen twice  We would describe the events using the same format  see  screenshot      3 car stolen  4 car stolen          ia            fiw      Follow the same format for the remaining seven victimization screen questions  Remember   number new descriptions consecutively using the counter at the bottom of the screen question  display as a reminder    What happens if two incidents were determined to be a single incident or a single incident was  determined to be two incidents  The answer is relatively simple  Let   s examine the first part of the  question  Using our purse bicycle stolen incident as a beginning  assume that a respondent reports  a break in in another question  You follow the appropriate conventions and enter a second incident   the break in   Thus  we have two incidents and have generated two incident reports  Several  questions later you determine that the break in and the purse bicycle theft were both a part of the  same incident  In this case  you must make some changes or else you will be presented with two  incident loops  Back up in the questionnaire to the break in screen question  Click the mouse on     yes    to bring up the text box in which you described the break in  Delete its contents  using the  delete key  so that the box is empty  Change the answer to    no    and register your answer by  pressing return  ALT N  or next  This registers the change  Next  go 
19. e file expformat  txt for more information  on exporting  However  the file is actually a Word document rather than a text file  In addition   the file name is expfrmat doc rather than expformat txt  This mistake makes it  difficult to locate the file using the find search function in Windows     The software is unable to handle multiple incidents  The reports module does not tabulate  multiple incidents nor does the software export data for multiple incidents  This limitation is  extremely problematic  The completed dataset is incomplete and analysis is difficult   Furthermore  asking questions for second and subsequent incidents is inefficient and not  worthwhile     Exported data are not in a user friendly format  For example  multiple response fields are  exported in columns reading    0011     This number indicates that the question had four possible  responses and multiple responses were allowed  In the example case  options three and four  were selected  How are these data analyzed  Frequencies simply provide an indication of all  of the various combinations that were recorded  What if the user wanted to determine what  percent of respondents selected option four  The data could be more easily analyzed if the  responses were in four separate columns  each indicating whether a response was selected or not   Data in this format can easily be analyzed individually or as multiple response categories     The software assigns crime codes that are invalid  For example  if the
20. e is a perfect option when the goal of the survey is to  interview all members of a particular household  Personal victimization entries are variable   depending upon the number of respondents in the household and the number of victimizations  reported  What if  however  the goal was to interview a sole respondent from each household  A  complex one to many database would be unnecessary so long as all victimizations from a single  respondent could be recorded  Since the Justice Center survey is designed to capture responses from  a single household member  this is the type of database underlying the Center   s victimization survey   All respondents are included in a single table  each record in the database represents one household  and one household member  All reported victimizations  up to 10  are recorded within this single  record  Consequently  data analysis is relatively easy  The database could be analyzed as is or  easily converted to a database wherein each record represents a unique incident    Structural changes aside  several other alterations were made to the National Crime  Victimization Survey as it was programmed into Ci3  The first set of changes fall into a category  relating to the eligible sample  While the NCVS is administered to all household members age 12  and older  the Justice Center set a lower age limit of 18  In other words  the survey is constructed  as an adult victimization survey  A conscious decision was made to exclude individuals under the
21. each potential answer choice  Interviewers select an answer choice by    7 Two important factors contributed to our decision to omit instructions for loading the survey  establishing  parameters  etc  First  such a discussion would simply reproduce technical manuals already included with Sawtooth  Technologies    WinCati and Ci3 software programs  Unlike the original BJS Crime Victimization Survey software  a  custom built program requiring detailed instructions  the Justice Center survey was created using a pre existing software  solution  As such  we rely on Sawtooth manuals for basic set up instructions  Second  set up procedures will likely vary  by user  Thus  it is simply not wise to produce a manual informing users of procedures such as where to save the survey  and data  how to produce call lists  and how to establish calling times      14     clicking on the radio button to the left of the chosen response  Upon clicking  you will notice that  the radio button partially fills in with a black circle  Since these questions only allow a single  response  clicking ona different answer after selecting an original choice will automatically deselect  the initial response  In the example below  only two options are available  The interviewer must  key either    yes    or    no           Multiple response  Multiple response questions allow the interviewer to key one or more  responses from a list of possible responses  Multiple response items are indicated by a square check  b
22. ely record information initially    At this point  several points should be measured     e Interviewers should be trained to inquire whether multiple    crimes    reported actually  represent one or more incidents  For example  a respondent might indicate in V37ab that  his her house was broken into and indicate in question V43ab that he she was sexually  assaulted  In both cases  the interviewer might mark each of the follow up questions to  indicate that each incident occurred once  This would result in the generation of two  victimization reports  However  the respondent might not be aware of the term incident  or might have missed the phrasing instructing them to exclude incidents already  mentioned  As such  although it would appear that the break in and sexual assault were  separate incidents  they might actually be part of the same incident  In this case  two  victimization reports would be generated when  in fact  only one is needed  a single  incident report would be able to capture the two offenses  the break in and the sexual  assault      e Interviewers should also be trained to perform a preliminary inquiry about whether an  incident reported as occurring six or more times are part of a series as defined by the  victimization survey  The importance of this preliminary inquiry is clear  If a  respondent reports a threat as occurring 7 times  the software will automatically produce  seven incident report  Moreover  although the actual incident report asks whether a
23. emember  the next unanswered  question might not be exactly where you left off since the change may have resulted in the  presentation of a new series of questions  Important  It is important to make sure changes are  registered  particularly when changing two or more previous responses  For example  suppose you  want to change questions 3 and 7  You back up to question 7 and make the necessary changes  You  then back up to question 3 and again make the necessary changes  Finally  you proceed forward to  the last unanswered question  In this case  the change made to question 7 would not be recorded  unless the change to question 7 was immediately followed by enter  Alt N  or the next mouse click   That is  the new answer must be registered  Thus  the interviewer must either start making changes  at the earliest question and proceed sequentially through the survey or he she must accompany each  change with enter  Alt N  or a next mouse click before backing up even further within the survey    Skipping questions  A question skipping function is not built into the survey  The inclusion  of a skipping function would disrupt the integrity of the survey  That is  a survey such as the crime  victimization survey relies so heavily upon contingent skip patterns  automatically moving  interviewers through the survey based on previous respondent answers  that skipping questions  disturbs these patterns  Yet  failing to include an escape code in each question would create the  possibilit
24. ent used the software but not for a victimization survey  Instead  they used the software as  a platform for constructing surveys for alternative purposes  In conversation  a representative from  that organization indicated that nearly three months were needed to break down and rebuild the  software to meet his organization   s needs  The University of New Hampshire  on the other hand   proceeded to use the software as designed  They acknowledged the limitations of the software and  recognized that some of the data may be compromised  In other words  they ignored the problems  in the software without correcting them    The Justice Center certainly agrees with the desire of the Bureau of Justice Statistics to conduct  local victimization surveys  The topics addressed by such a survey are important  However  the  limitations and problems with the current version of the Crime Victimization Survey software lead  the Center to question the integrity and utility of any data collected through its use  Instead  the  Justice Center proposed to develop the victimization survey by incorporating the questions contained  in the original CVS software into a widely used CATI software program     Major Activities  Purchasing Software    In order to continue with the original proposal   s plan to administer a victimization survey in  Anchorage  the Justice Center proposed to use the current survey questions and construct a  questionnaire using a popular CATI software program  This process beg
25. es  features associated with  WinCati      SPSS syntax for converting variables  This SPSS syntax files transforms variables and  assigns value and variable labels  The importance of transformation was described in more  detail previously but  briefly  the primary reason for transforming the exported dataset is to  reduce the overall number of variables  The SPSS syntax file can be opened in SPSS and  run on datasets collected using the questionnaire syntax developed as part of this project     Codebook  The codebook describes each variable including the value labels assigned  The  codebook variables are described post transformation  That is  the enclosed SPSS syntax  file must be run on exported data prior to using the codebook       Recall that unused victimization report variables may be deleted further reducing the overall size of the dataset   For example  the respondents to a Justice Center survey never reported more than five incidents  The Center was able  to delete nearly 3 000 additional variables corresponding to unused victimization reports six through ten     gt  Electronic file survey_syntax txt on the enclosed CD ROM       Electronic file spss_syntax_for_conversion sps on the enclosed CD ROM      9     5  Demonstration files  The demonstration files allow anyone to view the survey user  interface regardless of whether or not they own the software program on which the survey  was developed  Instructions for viewing the demonstration files are included in Append
26. fect their feeling of safety in  the neighborhood  If the respondent indicates that the conditions do make them feel less safe  they  will be asked to identify the condition that affects their feelings of safety the most  in the example  below  the respondent indicated that trash and panhandling existed within the neighborhood and all  other conditions did not         As you can see  previous answers are automatically entered on the screen  The interviewer will  simply mark the appropriate answer  However  you will note that all answer choices  regardless of  whether or not the choice is populated with a description  have check boxes next to them   Essentially  any box can be checked  However  interviewers should only check boxes with     21     populated answers  in the case above  choices 5  11  15  and 16   Interviewers should be instructed  not to check boxes which are not populated  Doing so will result in the recording of incorrect data     Recording Victimization Incidents    The Bureau of Justice Statistics    Crime Victimization Survey software structure dramatically  departs from the survey created by the Justice Center on Ci3  In the former  each victimization  report is manually added  Interviewers are required to recall the number of victimizations reported  and add the appropriate number of reports  The Justice Center   s version automatically presents  victimization reports based on the sum total of incidents reported in ten survey questions  discussed  in deta
27. his her exact street address  The Justice  Center believed that such a question would reduce the overall survey response rate  particularly among individuals who  are concerned about preserving their anonymity      2     aforementioned problems not only hampered interviewer efficiency  e g   incorrect dates   they  contributed to larger concerns about the accuracy of the data collected  e g   multiple responses  might be inadvertently recorded     These problems were brought to the attention of BJS and JRSA staff members  The original  project   s principal investigator met with Carlena Austin  BJS  and Lisa Walbolt  JRSA  at the 2001  conference of the Academy of Criminal Justice Sciences in Washington  D C  The Justice Center   s  concerns were expressed to both Ms  Austin and Ms  Walbolt and they indicated that others have  experienced similar difficulties  While the Justice Center was informed that there was going to be  an attempt to update the software  no timetable was given    Ms  Austin was extremely helpful during the process and has provided the Justice Center with  contact names and numbers for others who have used are currently using the software  Following  her suggestion  the principal investigator contacted individuals at the University of New Hampshire  and the Sheriff s Department of Pierce County  Washington  Both individuals used the software yet  they reinforced the Justice Center   s concerns about using the software  The Pierce County Sheriff s  Departm
28. il below   This format is arguably more beneficial since it relies less on interviewer recall  ability and automates what was a manual entry process  Nevertheless  the benefits are accompanied  by limitations  First  the survey is limited to recording no more than ten incident reports   Second   the procedures for entering victimizations described below must be strictly adhered to in order to  ensure the accuracy of the data    In order to begin describing the process through which individual victimizations are recorded   we must point out those questions that are designed to elicit incident reporting  They are     1  V36ab  Was something belonging to you stolen  such as  things that you carry like  luggage  a wallet  purse  briefcase  or book  clothing  jewelry  or calculator  bicycle or  sports equipment  things in your home like a TV  stereo  or tools  things outside your  home such as a garden hose or lawn furniture  things belonging to children in the  household  things from a vehicle such as a package  groceries  camera  or cassette tapes   or did anyone ATTEMPT to steal anything belonging to you     2  V37ab  Other than the incidents already described  has anyone broken in or  ATTEMPTED to break into your home by forcing a door or window  pushing past  someone  jimmying a lock  cutting a screen  or entering through an open door or  window  has anyone illegally gotten in or tried to get into a garage  shed  or storage  room  or has anyone illegally gotten in or tried to
29. interviewer gets to incident report Question 1b  stolen car should appear under  the Question 45b label in the left pane  However  there is no 45b label  Instead  the response  is placed under the 44b category  although this refers to a completely different question   Without understanding this error  the interviewer would simply select the corresponding  response category 44b in the right pane  As the interviewer proceeds through the interview   he she will encounter a question that reads     You said that during the last twelve months  no      Obviously  this is incorrect  The question should read     You said that during the last twelve  months someone stole your car        Miscellaneous  1  On some occasions  an interview might be only partially completed  When the interviewer  attempts to schedule a callback  the software displays a default date of January 1980  Although    this may be easily overcome by navigating to the appropriate date  it is certainly inefficient since  it is impossible to call a respondent on a date that has already passed      12     On some occasions  an interview might be only partially completed  The respondent might not  want the interviewer to call him her back  If the interviewer assigns the status    Term Person  Refused to Complete Survey     the phone number is closed with a date of 12 31 29     The software contains a file explaining the process of exporting data  The CVS Software User   s  Manual  p  74  instructs the user to look for th
30. ister an  answer and proceed through the survey using one of three methods     1  Simultaneously press the Alt and    N    keys  in this case the    N    represents next   2  Press the Enter or Return key  3  Click the left mouse button on the Next button at the bottom of the screen    The survey automatically progresses as answer choices are registered until a termination point  is reached  e g   respondent identified as ineligible  survey is completed      Reviewing answers  Interviewers may review previous answers as needed  Backing up in the  survey may be accomplished by     1  Simultaneously pressing the Alt and    P    keys  in this case the    P    represents  previous   2  Clicking the left mouse button the Previous button at the bottom of the screen     Pressing and holding Alt P rapidly moves users back in the survey  After reviewing previous  answers  move forward by following the directions for registering answers discussed above  It is  only possible to move forward to the point of the next unanswered question     Changing previous answers  Backing up within the questionnaire can be used to change  previous answers as well  Using the procedures discussed above  move to the previous answer   Change the answer as needed  Please note that changing an answer might alter the subsequent series  of questions  After changing an answer  register it by pressing enter  Alt N  or next with the mouse     19     button  Hold Alt N to proceed to the next unanswered question  R
31. ix  C     Future Plans    The purpose of developing the victimization survey on a third party software platform was to  allow the Justice Center to administer the survey to a sample of residents of Anchorage  Alaska   Additional funding allowed the Center to begin data collection in April 2002  At the present time   data have been collected for nearly 800 Anchorage residents  The survey worked exceptionally well  and data collection continued with few problems  those that did emerge were simply learning curve  issues   The Justice Center is confident in the survey developed for this project and anticipates  analyzing data collected with the survey sometime in Fall 2002  In the future  the victimization  survey might be administered once again or new surveys will be developed using the Sawtooth  software platform      10     Appendix A  Problems Associated with the  BJS Crime Victimization Survey Software Program    Household CVS1 with COPS Supplement    1  Question 46c  Vandalism   Many questions include an    other    response category  In most  cases  the interviewer selects    other    and a text box is enabled allowing the interviewer to  describe or specify what is meant by    other     However  selecting the    other    category in  Question 46c does not enable the text box  thereby prohibiting the interviewer from providing  additional information     2  Question 17b  What type of activities did you see   The question asks the respondent to  indicate which community ac
32. n  event is a Series  indicating the events as a series will not reduce the number of incident  reports generated  Remember  incident reports are generated based on the sum of all  incidents reported in the follow ups to the questions described above  Therefore  we  suggest that  if a respondent indicates that an incident has occurred more than 5 times   interviewers be instructed to determine whether this is a series or not  They can do so  by asking whether the incidents are similar to one another in detail and whether they are  distinguishable from each other  If the events are similar and indistinguishable   interviewers should enter 1 for the number of times the incident occurred  In this way   only a single incident report will be generated for this series of incidents     These caveats aside  let us now turn to a description of the process of entering incidents  Several  examples will be used throughout to illustrate this process    Recall that incidents are entered in one or more of ten questions above  Each of the questions  is a single response item with an    other    feature added  If the respondent indicates that they have  not been victimized  they don   t know whether they have been victimized  or they refuse to say  whether they have been victimized  the survey will simply move you to the next relevant question   If  however  a respondent indicates that he she has been victimized  click on the yes box  In the     24     example below  you will notice that the 
33. nality that is not available in Ci3 alone including call  management  disposition assignment  callback scheduling  and other CATI features  The  functionality added through the WinCati program is most useful when actually administering the  survey  the actual creation and testing of the victimization survey in this project used the authoring  program  Ci3  much more often than WinCati     Programming the Victimization Survey    The most challenging part of this project was the programming phase  The difficulty lies not in  learning of the Ci3 programming language or in entering question text  Instead  the challenge was  programming the complex skip patterns which exist within the paper and pencil NCVS  In addition   structural changes to the survey as well as modifications were made in order to make the survey  easier to use and more relevant to the needs of Anchorage  This section describes several of these  issues    Learning how to program the victimization survey using Ci3 was initially a daunting task  The  principal investigator  along with a research associate  made great use of the tutorial and reference  manual that accompanied the software program  However  we were only able to proceed so far on  manuals alone  The actual process may best be described as one of trial and error  The survey itself  was separated into modules with each tackled one at a time before being combined with the whole   For example  the community policing questionnaire was treated separately f
34. ndix C  Exploring the Victimization Survey without Ci3 or WinCati                  34    Introduction    In Fall 2000  the Justice Center at the University of Alaska Anchorage received funding from  the Bureau of Justice Statistics under the State Justice Statistics Program for Statistical Analysis  Centers  The proposal originally funded under this program involved administering a local version  of the National Crime Victimization Survey to a sample of residents of Anchorage  Alaska   Moreover  the Justice Center planned on convening a working group tasked with developing a  protocol for administering similar surveys in rural Alaska    The State Justice Statistics program encouraged applicants to address one or more of nine  priority themes topics  Local victimization surveys were one of the themes considered high priority   Specifically  theme 2000A provided funding to conduct victimization surveys     Supplemental funding will be available for SACs to conduct state level and or local  crime victimization surveys  SACs receiving funds under this theme must agree to  use the BJS developed Crime Victimization Survey software  which can be easily  modified to meet State local priorities and requirements     Upon receiving funding  the Justice Center proceeded with the planning stages of the victimization  survey    One of these tasks included testing and modifying the BJS Crime Victimization Survey software   However  the Center quickly learned that the Crime Victimization Sur
35. ny incidents of this type happen to you     5  V4lab  Other than the incidents already described   in the last 12 months has anyone  attacked or threatened you in any of these ways  Excluding telephone  letter fax email  threats  with any weapon  for instance  a gun or knife  with anything like a baseball bat   frying pan  scissors  or stick  by something thrown  such as a rock or bottle  include any  grabbing  punching  or choking  any rape  attempted rape or other type of sexual attack   any face to face threats  or any attack or threat or use of force by anyone at all  Please  mention it even if you are not certain it was a crime     6  V42ab  Other than the incidents already described   in the last 12 months did you have  something stolen from you OR were you attacked or threatened by someone at work or  school  a neighbor or friend  a relative or family member  or any other person you   ve met  or known  Did any incidents of this type happen to you   Excluding  telephone letter fax email threats     7  V43ab  Incidents involving forced or unwanted sexual acts are often difficult to talk  about  Other than the incidents already described  in the last 12 months have you been  forced or coerced to engage in unwanted sexual activity by someone you didn   t know  before  a casual acquaintance  or someone you know well  Did any incidents of this type  happen to you     8  V44a  During the last 12 months   other than any incidents already mentioned  did you  call the police to 
36. on is what you typed in earlier and is visible as number  one toward the middle of the screen  Thus  you would say     Now I will be asking questions about the incidents you described   I ll begin with the first incident mentioned  purse bicycle stolen     At this point  you are ready to proceed with the completion of the first incident report  You will  navigate through the survey as described above  asking questions and entering responses along the  way  There are no tricks within the incident report  The software will automatically move you  through the survey based on answers provided  After completing the first incident loop you will  return to a screen that looks similar to the one that began the first incident loop  see screenshot    You will notice  however  that the screen now states that you are on incident two of four and you  will be asked to read the description for incident two     37        Move through the incident report for the second incident  Repeat for all incidents reported   Upon completion of the incident reports  you will be taken to the community policing and  demographic sections  Keep in mind that  if zero incidents were reported  you would be moved to  the remainder of the questionnaire while bypassing incident reports    If you need to add or delete an incident loop  you must return to the screen questions  pressing  and holding ALT P moves the survey backward rapidly  and make the necessary changes so that  an additional incident report is gener
37. oxes next to the potential choices rather than circular radio buttons  Interviewers use their mouse   left click  to select and deselect answer choices  The example below has 11 possible answer  choices  Users are able to select one or more answer choices by clicking on the answer box  In most  cases  the maximum number of answer choices is programmed within the survey syntax  In the  example below  the interviewer is able to key between one and nine answer choices  We do not  want the choices    don   t know    and    refused    selected along with any of the first nine choices  Note   however  that there is no programming preventing interviewers from entering the    don   t know     and or    refused    choices when other answer choices are keyed  Interviewers must simply be trained  not to enter the last two choices when one or more of the other choices is selected      15        Open ended  Open ended question types require interviewers to type in text responses  Unlike  the responses for single  and multiple response questions  open ended answers are not exportable  by Ci3 in SPSS format  Instead  they are part of a text   txt  file exported separately  In the example  provided below  a curser is positioned below the instruction box  Interviewers simply type text   Interviewers are permitted to enter text on multiple lines  However  there is generally a restriction  on the number of lines of text allowed  experience has demonstrated that the number of lines  permitted is
38. re for  reducing the number of variables into a more manageable dataset  This was accomplished by  developing an SPSS syntax file     Three factors contributed to the large number of variables in the dataset  First  the National  Crime Victimization Survey itself is a complex questionnaire comprising several hundred questions      Since the goal was to maintain the integrity of the original survey as much as possible  the Justice  Center did not eliminate many questions from the original instrument     Second  the survey was  programmed with the capacity to handle ten separate incidents  all of which are included in the  exported dataset regardless of whether or not they are actually used  However  once the dataset is  exported  users can simply delete out those variables for incident reports that were not used  For  example  if no respondents report more than five incidents  users can delete victimization report  variables for the remaining five victimizations thereby reducing the overall number of variables    The final and most important contributor to the large number of variables is the Ci3 program  itself  The vast majority of the questions in the victimization survey require interviewers to select  one or more responses from a list of possible choices  These questions result in dataset variables  equal to the number of possible answer choices    For example  consider a hypothetical question number 20  This question has both yes and no  responses as well as don   t know
39. report something that happened to YOU which you thought was a  crime     9  V45a  During the last 12 months   other than any incidents already mentioned  did  anything which you thought was a crime happen to YOU  but you did NOT report to the  police     10  V46fl  In the vandalism just mentioned  were you attacked or threatened  or was  something stolen or an attempt made to steal something that belonged to you or another  household member  other than any incident s  already mentioned      Each of these questions is followed by a numerical question asking the following  enter the total    of times the incidents just mentioned occurred     Thus  there are ten questions that ask respondents      Questions V44a and V45a are immediately followed by a follow up question asking whether the incident involved  an attack or threat or a theft or attempted theft  The CVS is only concerned with incidents that fit this description  The  question asking about the number of times an incident occurred is presented after this      23     to indicate how many times particular incidents occurred  The number of incident reports generated  by the victimization survey is equal to the sum of all incidents recorded in these ten follow up  questions    Consequently  the importance of accurate and complete victimization reporting cannot be  overemphasized  Although changes to recorded answers can be made at a later point in the survey   see navigation instructions above   it is much easier to complet
40. rom the rest of the  questionnaire  The principal investigator programmed this portion of the survey and then appended  it to the programming for the remainder of the survey     The obstacle causing the most frustration  however  was the complexity of the questionnaire   By complexity  we refer not to the length or content of the questionnaire but  rather  the contingent  relationships between questions resulting in complex skip patterns  A glance through the Syntax  Guide will illustrate this complexity  In some cases  skip pattern programming amounted to more  than 10 lines of text  To say that the principal investigator easily identified the underlying logic of  the skip patterns would be misleading  The skip patterns were programmed and repeatedly tested   see Testing the Victimization Survey Structure below   It was only after research assistants worked  through the survey dozens of times and repeated modifications were made to the programming  syntax that confidence was gained that the Justice Center survey captured the patterns underlying  the original survey    The programming altered the survey in several ways from its original NCVS format   Structurally  the survey was modified to ease use  The NCVS database is constructed as a one to   many database  The one table includes household demographic questions as well as household  victimizations  In contrast  the many table contains all personal victimization data for all members  of the household  This type of structur
41. servations  NCVS 1 questions 8 and 12e   were excluded because they were not relevant to a household survey of Anchorage residents     e The income question  NCVS 1 question 12a  was modified to reduce the number of choices  by using  10 000 increments  During the administration of the NCVS  a card is handed to  respondents with each of the 14 income categories  The respondent then marks points out  the appropriate category  During a telephone survey  reading the answer choices was a     6     tedious process that created confusion for respondents  The Justice Center decided to simply  the categories     Testing the Victimization Survey Structure    The Justice Center produced an initial version of the victimization survey using Ci3 software yet  the construction and revision of the survey considered for nearly eight months  From December  2001 through March 2002  the Justice Center employed two research assistants  Their assignment  was to repeatedly work through the survey using the interviewer interface  they did not view the  programming syntax  while comparing the survey with the paper and pencil NCVS instrument  In  essence  the researchers were advised to    break the survey     That is  they were instructed to locate  typographical errors and structural integrity problems associated with skip patterns and make any  other recommendations that they felt would lead to a better interviewer interface  e g   suggestions  about fonts     The researchers worked tirelessly for
42. st example where the purse and bicycle were stolen during the same incident   After entering in a description  you proceed to the follow up question  how many times did the  incident occur  Let   s suppose the respondent indicates that the incident only occurred once  Enter  one in the space provided  It is important to remember what the implications are of the value entered  in each of the follow up  how many   questions  By entering one  the software is set to generate a     26     single incident report  If we indicated that the purse and bicycle were stolen twice  we would enter  two  Again  the result of entering two would be the generation of two incident reports        As you can see  the number of incidents reported in the follow up  how many   question should  be equal to the number of incidents described in the previous question  To illustrate this point  let  us return to the purse bicycle theft incident  The respondent answers that her purse and bike were  stolen  In the follow up  how many   question  she indicates that this same type of incident  happened twice  Therefore  the incident must be described twice in the previous question  see  screenshot      1l purse bicycle stolen  2 purse bicycle stolen    Dae pe Oe ee pe la       You may ask why the incident needs to be described twice when the circumstances are the same   Besides  we have correctly identified the incident as having occurred twice  although not described  twice  thereby correctly generating two 
43. the seven  additional screen questions  It is now time to begin filling out automatically generated incident  reports    Upon completing the screen and vandalism  if applicable  questions  an  information instruction question screen will appear  see screenshot below   This screen simply  indicates that you are about to start the process of completing incident reports               Pressing enter or next moves you to the initial screen of the incident report  see screenshot       31        You will see the results of your incident descriptions on this screen as well as the sequential  numbers assigned  You may also see the importance of keeping incident descriptions on a single  line  Anything beyond a single line will not appear on the initial incident report screen  Therefore   it is imperative that incident descriptions entered during the screen question section are limited to  one line    Within the double lined box is a counter  In our example  the counter indicates that we are  currently working on incident number one of four  The software has been programmed to recognize  the number of incidents described and automatically generate the appropriate number of incident  reports    You begin the incident loop by stating     Now I will be asking questions about the incidents you described   I   ll begin with the first incident mentioned     The instructions in the double lined box instruct you to follow these remarks by reading the  description of incident one  The descripti
44. tivities they saw police performing  The interviewer is instructed  to enter the appropriate answer for all categories that apply  Yet  the software program only  allows the interviewer to select a single activity  This is an obvious limitation considering that  the question asks respondents to    specify what types of activities     The BJS report  Criminal  Victimization and Perceptions of Community Safety in 12 Cities  highlights the fact that multiple  responses should be allowed     Crime Incident Report    1  Question 42  What did you do   The question asks respondents to indicate which actions they  took to protect themselves  When the software presents this question for the interviewer to read   two categories appear in the response pane  One of the categories is a major heading that is not  a valid data point  Instead  the respondent must open up a list of subcategories  The major  category heading is    attacked offender with weapon     However  only three of the nine  subcategories have anything to do with a weapon  Without clicking on the broad category  heading  the interviewer might be inclined to select    other    if he she believe that the  respondent   s answer has nothing to do with attacking an offender  This would be an inaccurate  assumption     2  Question 43a  Before or after injury   The question asks whether protective actions were  taken before or after injury  although the respondent never indicated that he she was injured   This question appears e
45. ts see  subcategories  the user must either double click the broad category heading or click on the plus  sign to the left of the heading  If the interviewer does the latter  the broad category heading is  highlighted  This should not be problematic since the heading is not a valid data point   However  if the user does not deselect this category  additional  unwanted responses will be  recorded  The software inexplicably moves this broad category highlight up to the next valid  data point  In many cases  this is not an answer given by the respondent or an answer recorded  by the interviewer and results in a compromised dataset     5  Question 1b  identify screen question from CVS1   If the respondent indicates that he she  experienced what was believed to be a victimization and did not report it to the police  Question  45a    Unreported Incidents of Crime    in the Household CVS1 with COPS Supplement   a  problem appears in the incident report  Question 1b is inaccurately constructed  The left pane  shows all screen questions and their corresponding answers  The right pane shows the  corresponding questions that help guide the interviewer through the incident report  However   while the right pane includes ten separate screen question categories  the left pane only includes  nine categories  The discrepancy should not exist  Consider the following example  A  respondent indicates that he she had a vehicle stolen but did not report it to the police  Question  45a   When the 
46. ually has Ci3 installed  Users collect data using the field disks and  the data collected from the field disks are aggregated using the Ci3 program  In other words  data  collection  without CATI capabilities  can be conducted on any computer using Ci3 generated field  disks  However  the survey and field disks must be produced on a computer with C73 installed   Moreover  data collected are only viewable and exportable on the computer with Ci3    The series included on the enclosed CD ROM are essentially survey field disks  Users will be  able to explore the user interface and layout of the survey produced by the Justice Center  In  working through the survey  data are actually being recorded  However  users will not be able to  view the data  These files are simply included to allow users to examine the survey     Instructions  1  Place the CD ROM in the computer CD drive   2  Navigate to the folder   Q  CVS Demonstration  where Q is the drive letter assigned to your computer   s CD ROM drive  3  Inside the    Demonstration    folder you will see the following files   CVSsyn qst  SSIFIELD ini  WinQue exe  Qexit dll  All four files are required for the field disk demonstration program to run  Two additional files  will be created during the first run though of the survey  They are CVSsyn dat and CVSsyn idx    files  These two files store the data collected although  as previously noted  users will not be  able to view the data      34     Start the demonstration program by do
47. uble clicking the left mouse button on the file    WinQue exe     WinQue    The first screen of the demonstration program will appear and should look similar to the  following        6  You are now able to proceed through the survey using directions outlined in Appendix B     Additional Notes    1     When a survey is completed  another survey will automatically open  In order to exit the  program  click the left mouse button on the X in the upper right corner of your screen     Ci3 limits the number of cases that a field disk can collect  Consequently  there is a limit to the  number of times the demonstration program can be viewed  approximately 800 cases   If the  limit is reached  however  you may simply reset the counter by deleting the CVSsyn dat and  CVSsyn idx files  These are the data files and will re regenerated when the WinQue exe file is  opened     35     
48. ven if the respondent answered    no    to Question 24  did the offender hit  you  knock you down  or actually attack you   Question 25  did the offender try to attack you    and Question 26  did the offender threaten you   The question is inapplicable since the  respondent never mentioned any injury     3  Question 96  what items were taken   Question responses are displayed in the answer pane  as seven broad categories  These categories are not valid data points  they cannot be selected  and cannot be recorded as responses  In order to view valid data points see subcategories  the  user must either double click the broad category heading or click on the plus sign to the left of  the heading  If the interviewer does the latter  the broad category heading is highlighted  This     11     should not be problematic since the heading is not a valid data point  However  if the user does  not deselect this category  additional  unwanted responses are recorded  The software  inexplicably moves this broad category highlight up to the next valid data point  In many cases   this is not an answer given by the respondent or an answer recorded by the interviewer  This  problem likely results in a compromised dataset     4  Question 117  why was it not reported to the police   Question responses are displayed in the  answer pane as six broad categories  These categories are not valid data points  they cannot be  selected and cannot be recorded as responses  In order to view valid data poin
49. vey software designed by BJS  was inadequate for its purposes  The problems associated with the software resulted in the recording  of inaccurate data  incorrect dates  and incomplete responses  Furthermore  the software was not  easily modified and exported data was often incomplete and or difficult to analyze    Due to these limitations  the Justice Center submitted a revised proposal outlining a project  designed to ultimately overcome many of the limitations inherent with the Crime Victimization  Survey software  Instead of actually administering a crime victimization survey as originally  planned  the new project involved developing a survey administration platform using a commercially  available computer assisted telephone interviewing  CATI  software program  This report describes  the process of developing and testing the victimization survey using this alternative software  program     Problems Encountered with the CVS Software  Fiscal Year 2000 Theme 2000A required grantees to use the Crime Victimization Survey    software recently developed by the Bureau of Justice Statistics  The computer assisted telephone  interviewing  CATI  software program included questions that were based on those from the     1     National Crime Victimization Survey that has been conducted since the early 1970s  Furthermore   the software was designed to expedite the process of collecting victimization and community  policing data from state or local jurisdictions  The value of these data
50. victimization reports  This requirement is important for a  single reason  it assists the interviewer and the respondent focus at the beginning of each incident  loop  As you will see below  describing incidents more than once  even when exactly the same  is  extremely important and results in a more user friendly survey    For the time being  let us proceed with our example  Assume that our respondent has recorded  a single incident  purse bicycle stolen   We then move on to the next screen question  in this case  one related to break ins        Notice the double lined box on the lower part of the screen above  You will remember that these  boxes represent instructions or notes to interviewers and  unless otherwise indicated  the contents  are not meant to be read aloud to respondents  The screenshot above shows that our respondent has  reported once incident  the purse bicycle theft   At this point  then  we have generated one incident  report  However  we still have nine more screen questions to ask  Suppose our respondent answers  in the affirmative to the question in the screenshot above  She has experienced a break in  You will  probe to determine whether the break in was part of the same incident where the purse and bicycle  were stolen  In this case  our respondent states that they represent separate incidents  Thus  we will  follow our format for entering a description  However  do not begin numbering at one again   Instead  continue numbering in numerical order  The
51. y that interviewers would become fixed on a particular question  In order to overcome this  problem  the vast majority ofall CVS questions include    don   t know    and    refused    options  Thus   if a respondent is unsure of his her response or simply refuses to provide an answer  the interviewer  can mark one of these options and continue through the survey  One may ask how these additional  options are included in the survey and integrated within pre established skip patterns  Common  sense was used to program these skips  The survey skips to the next question that makes sense yet  preserves much of the integrity of the original crime victimization survey  For example  the    don   t  know    and    refused    codes skip interviewers to the same question that appears if a    no    response  is selected  Thus  the skip patterns for the added codes were based on    what made sense        Selecting Responses from Populated Lists    For the purposes of this discussion  populated lists are those answer choices that are  automatically filled in based on answers from previous questions  Before describing how to key  answers to these types of questions  essentially single  or multiple response questions   an example  will be presented  The question depicted below is followed up by a series of questions asking about  other conditions which may or may not exist within the respondent   s neighborhood      20        The respondent is then asked if any of the conditions mentioned af
    
Download Pdf Manuals
 
 
    
Related Search
    
Related Contents
Fujitsu LIFEBOOK NH570  3M™ Nahes Projektion Supersystem SCP713 Betriebsanleitung  COMPRENDRE AGIR  uC-USB-Device User`s Manual v4.05 - Doc  Sling - Etac    819-5236 MAY 09  Samsung PS-50P4HR Инструкция по использованию  Manual Motor Starters      Copyright © All rights reserved. 
   Failed to retrieve file