Step 6 in the scalable risk assessment process is to use the analytic method selected in Step 5 to estimate the desired exposure measure(s). All of the previous steps involve making scoping or planning decisions about how to estimate exposure. Step 6 in the process is when the detailed analysis for exposure estimation occurs, As a result, the Step 6 section in this guide is the largest and has the most content.
Step 6 includes a section for each of the three primary methods that can be used to estimate exposure.
- Site counts: Overview of the manual and automated counting procedures
- Demand estimation models: Overview of the demand estimation models, with a particular focus on direct demand models, in estimating non-motorized exposure.
- Travel surveys: Overview of the most commonly used travel surveys in estimating exposureThis section also introduces an online interactive tool that estimates exposure using national travel survey data from NHTS and ACS.
Site counts are direct measurements of the number of pedestrians or bicyclists at a defined location. The counts may be gathered automatically from various technology-based sensors or manually from human observers. Site counts are taken at a point and are typically used to represent two different scales: point and segmentFor the point scale, counts are most commonly used for intersection crosswalks. For the segment scale, the site count (taken at a point) is applied to the entire length of a street segment (where the counts are not expected to vary significantly along the segment length),
Counts are more commonly used to estimate exposure when the desired facility coverage is limited, as count data collection for all facilities within a large network or region is cost-prohibitive (unless extensive sampling is used)In some cases, it is also a challenge to get representative pedestrian and bicyclist counts due to seasonal and day-of-week variation.
This section of Step 6 provides an overview of counting procedures for pedestrians and bicyclistsIn particular, this section will highlight considerations and issues that are relevant to site counts used for exposure estimation. Analysts will find many more procedural details in the comprehensive guidance documents listed in Table 15.
Typical Applications of Site Counts
In some exposure analyses, site counts may serve as the only source of data for exposure estimationHowever, the collection of site counts on all street segments or at all signalized intersections within a city or region is cost-prohibitive. Therefore, the use of site counts only is most applicable to exposure analyses that are facility-specific (i.e., point or segment scale) and focus on a limited number of intersections or street segments.
In most exposure analyses, site counts are collected at a small but representative sample of locations, and then an estimation model is developed and calibrated from these site counts and used to estimate pedestrian and bicyclist volumes at uncounted locations. The next major section in this chapter describes several different demand estimation models that use a sample of site counts and other street inventory and land use variables to predict pedestrian and bicyclist volumes at all locations citywide.
Some cities, MPOs, and state DOTs have begun to collect pedestrian and bicyclist counts as part of a routine monitoring program, so it may not be necessary to start from the beginning for exposure estimationOne should inquire about existing count data not just with transportation agencies, but also with other groups and agencies that may share your interest in pedestrian and bicyclist activity:
- City or county parks and recreation department (e.g., on shared use paths).
- National or state parks (e.g., on internal or connector paths).
- Public health departments (e.g., monitoring physical activity).
- Retail or business associations (e.g., on pedestrian malls or plazas).
- Pedestrian and/or bicyclist advocacy groups.
Even if other agencies' existing counts can be used, it is likely that additional counts will be desired for the purposes of exposure estimationIn some cities, the existing and ongoing counts are at a very limited number of locations, and may have been chosen because of high pedestrian and bicyclist usage or recent facility improvements. For purposes of exposure estimation, additional counts may be needed at high crash locations or at a broader range of locations that represent a mix of facility types and land uses (for purposes of estimation model development and calibration).
Calculating AADT from Site Counts
Site counts can be conducted for varying durations (e.g., 12 hours, 48 hours, 7 days, etc.) and at different times of the year, but the final count measure most commonly used in both motorized and non-motorized exposure estimation is AADT. In these cases, AADT is applied to a defined segment, and is then multiplied by the defined segment length to calculate average annual daily PMT or BMTThe same process can be used at intersection crossings, whereby the crossing length is used to calculate PMT or BMT at each crossingEarlier in this guide, Step 4 described a simple example of this calculation process.
The estimation of AADT from a short duration site count is a recommended practice for pedestrian and bicyclist exposure estimation. AADT estimates are typically made using what are called factor adjustments (described in detail in Section 4.4 of FHWA's Traffic Monitoring Guide). Aside from being a widely accepted practice in motor vehicle analysis, there are several reasons why AADT values should be estimated and used in exposure analysis. Pedestrian and bicyclist counts can vary dramatically by month of the year, day of the week, and by prevailing weather conditionsSite counts that were collected for a single day during favorable weather are not a representative sample of all other days during the year. Therefore, in simple terms, the factor adjustment process is used to adjust the count samples to represent more accurately the true annual average pedestrian and bicyclist usage.
The factor adjustment process for AADT estimation requires continuous count data from similar location(s) to scale accurately a short duration count to represent an annual average countHowever, some cities or regions may not have the required continuous count data for factor adjustment. At the time of this Guide development, many agencies are still working to implement the factor adjustment processSeveral organizations are currently working to develop default seasonal adjustment factors based on climate zones.1 If factor adjustment and AADT estimation is not feasible for your application, every effort should be made to collect site counts during seasons/months, days of the week, and weather conditions that most closely resemble typical conditions during the year.
Collecting Counts Specifically for Exposure Estimation
The following paragraphs describe key considerations when collecting counts specifically for exposure estimation. Many of these considerations are addressed in more detail in other data collection guides, but are highlighted here to emphasize their importance.
Use automated counter equipment as much as possible: Automated counter equipment allows counts to be conducted for longer periods (multiple days), which reduces error in AADT estimates. Automated counter equipment can also reduce the labor cost of data collection. The resources listed at the beginning of this Site Counts section provide detailed guidance on selecting the appropriate technologies for automated counting.
Avoid very short duration counts (i.e., two-hour counts): The FHWA Traffic Monitoring Guide recommends a minimum duration of seven days for short duration countsIf this duration is not cost-feasible at all of the desired locations, then at least one to two twelve-hour periods should be counted. Two-hour counts should be avoided as much as possible, even it if means reducing the number of count locations to allow for a six-hour or twelve-hour countSeveral research efforts have documented the high error rates that result from estimating AADT from two-hour counts.
Seek balance between number of count locations and duration: As implied in the previous bullet, it may be necessary to balance the number of count locations with the duration of each count. This may be necessary to avoid very short duration counts yet still collect data at a representative number of count locationsIn these cases, one is balancing the temporal error (from sampling short durations of time during the year) with the spatial error (from sampling very few locations of all locations to be analyzed). At this time, achieving a balance between number of count locations and count duration is considered more of an art form and not science.
Select representative months and days of week for your area and count location: Site counts should be collected during typical or normal seasons/days/times, especially if adjustment factors are not feasible to use for estimating AADT valuesEven if adjustment factors are planned for use, site counts should be collected during months/days/times that are considered typical or normal conditionsThis helps to reduce the magnitude of the factor adjustment, and ultimately, reduce the error associated with AADT estimates.
Focus on balance of high-priority yet representative locations: If site counts must be collected specifically for exposure estimation, the count collection effort should focus on a balance of high-priority yet representative locations. High priority locations for exposure analysis are likely to be those locations that have a high crash frequency. However, if an estimation model is to be developed, site counts at high-crash locations are likely to be a biased input for model developmentThe result is an estimation model that only predicts counts accurately at high-crash locations. Therefore, high-crash locations should be balanced with other locations that represent a range of facility types and land use patterns. Note that this balance is considered a more of an art form and not science at this time.
Demand Estimation Models
As described in Step 5, there are several estimation models to estimate pedestrian and bicyclist demand for input to an exposure measureIn this section of Step 6, we provide a detailed overview and key considerations in developing direct demand models since they are the most widely used tools in the literature for pedestrian and bicyclist volume estimation modeling. Specifically, detailed overview of direct demand models is presented first, and then step-by-step instructions are provided (with examples) to develop a direct demand model.
Detailed Overview of Direct Demand Models
Direct demand models are the most widely used tools in the literature for pedestrian and bicyclist volume estimation modeling (especially in supporting traffic safety studies). These models have been primarily used to develop facility-specific demand estimations for the local level of community, project, and facility planning and to evaluate and prioritize projects. The potential usages of direct demand models are listed as follows by Kuzmyak et. al.l (2014):
- Answer questions about facility use or needs that could not be addressed with traditional trip-based regional models because of limitations related to scale and ad hoc treatment of non-motorized modes.
- Address the need for estimates of walk activity on links and at intersections for safety analysis and design.
- Address the need for estimates of bicycle activity to support questions on bike network design and to support decisions on facility needs.
- Provide a better connection between the context of the given built environment and non-motorized travel behavior and demand
The FHWA has sponsored a Non-Motorized Travel Analysis Toolkit, which includes various applications to support non-motorized transportation planning and modeling. This Toolkit includes several direct demand models to estimate pedestrian and bicycle volumes (http://www.nmtk.pedbikeinfo.org/ui/#/). Direct demand models have also been identified as the primary tools to measure bicyclist and pedestrian exposure for safety analysis.
Direct demand models are generally based on different versions of regression modeling to explain “demand levels as recorded in counts as a function of measured characteristics of the adjacent environment” (Kuzmyak et. al. 2014). As indicated by Munira and Sener (2017), “the concept of using a direct-demand model to estimate non-motorized activity is not newStudies dating back 50 years have forecast non-motorized traffic using count and spatial data.” Schmiedeskamp and Zhao (2016) explained such models as following “a similar approach of first proposing a set of explanatory variables, fitting some form of regression model, and then interpreting and justifying the results according to the guiding theory.” Direct demand models are based on variety of data sources such as activity counts, census population and employment characteristics, land use and topography and transportation network characteristics.
Direct demand models are appealing due to their simplicity and convenience in development and application, and since they are generally based on available dataThese models are particularly useful for screening and preliminary analyses especially when the resources are limited and a more comprehensive (and relatively expensive) model is not available or not possible to developHowever, direct demand models are limited in terms of capturing the behavioral structureIn addition, they are usually not transferable due to their strong linkage to local context, activity levels and the characteristics that the models are built onÂ Aoun et. a. l(2015) summarized the advantages and disadvantages of direct demand models as in Table 16.
Source: (adapted from) Aoun et. al. 2015
Kuzmyak et. al. (2014) highlighted the need to be judicious in the development and application of direct demand models, and suggested the following guidelines:
- The models need to be developed from scratch for each study, and well calibrated to existing conditions with the specific area and on the specific facilities under studyModels developed for a specific area cannot be construed as transferable.
- Uncertainties developed due to unaccounted origin-destination, route choice, and trip purpose data may be narrowed down by developing counts and models focused on a specific time period.
- After model calibration, the reliability of the models to predict volume in individual locations and the overall study area should be tested.
- Need to be judicious in the types of applications or decisions to be supported by the models.
The readers are referred to Munira and Sener (2017) for an in-depth review of the available literature associated with direct-demand modeling to estimate bicycle and pedestrian activity.
Development of a Direct Demand Model
This section provides step-by-step instructions to develop a direct demand model.
The generalized approach to develop a direct demand model includes three primary phases as demonstrated in Figure 17
In addition, Figure 18 provides a flowchart of an algorithm to help the analysts walk through the tasks needed to complete the phases when developing a direct demand model
Below, we provide detailed instructions on each phase, and the corresponding tasks involved in each phase as consistent with the flowchart provided in Figure 18Specifically, the main objective and primary tasks of each phase are presented first, followed by discussion of key considerations when processing the phase.
Phase A: Study Identification
Objective: The main objective of Phase A is to identify the study focus.
Primary Tasks: Phase A includes three primary tasks:
- Task 1: Identify study area.
- Task 2: Determine facility locations in the identified study area.
- Task 3: Determine variable of interest (i.edependent variable) of the direct demand model.
Processing the Phase:
Task 1: Phase A starts with identification of the overall study area (or target population) for which the direct demand model is desired to be developedWhen executing the task, it is of utmost importance to identify project objectives to avoid any unnecessary process, optimize the limited resources and limit the bias.
Task 2: The next task is to determine the facility locations at which the exposure measure is desiredExample of facility locations include signalized and unsignalized intersections or midblock locations along the street segmentsStep 3 of this Guide provides information on how to determine desired geographic scale(s) for exposure.
Task 3: Once the location-specific details of the model are identified, the analyst needs to determine the variable of interest, i.ethe model output or outcome whose variation is being examinedFor example, in the context of non-motorized direct demand models, the analyst might be interested in obtaining annual average pedestrian volume and peak hour bicycle traffic volumeIn statistical terms, the variable of interest is named as the dependent variable of the model.
Phase B: Data Preparation
Objective: The main objective of Phase B is to prepare the data needed for model development.
Primary Tasks: Phase B includes five primary tasks:
- Task 1: Identify and compile data needed for the dependent variable of the model.
- Task 2: Process the data and perform data quality checks for the dependent variable.
- Task 3: Identify and compile candidate explanatory (i.eindependent) variables of the model.
- Task 4: Process the data and perform data quality checks for the independent variables
- Task 5: Combine the datasets for dependent and independent variables.
Processing the Phase:
Task 1: The first task in this phase is to compile data needed for the dependent variable of the modelAs aforementioned in the Site Counts part of Step 6 of this Guide, site counts serve as the common source of data for exposure estimation especially when the desired facility coverage is limited and data collection for all facilities within a large network or region is cost-prohibitiveTherefore, site counts are the main ingredient used for creating the dependent variable of the model
Sampling: In the ideal conditions, it is desired to have site counts at all facility locations across the study area, but this is not feasible given the limited amount of resources (i.ebudget, time, equipment and manpower constraints)Therefore, the analyst needs to develop a sampling strategy to select sites from which data will be collectedThere are two different types of sampling technique for data collection: non-probabilistic sampling technique and probabilistic sampling technique
Non-probabilistic sampling techniques are the most commonly used techniques due to their low cost, easy-to-implement methodologyIn this technique, the analyst identifies some subjective criteria (e.gconvenience, engineering judgment, local knowledge, quota, etc.) and collects data based on those criteriaThis technique does not allow the analyst to control for sampling error, and usually a small number of samples are collectedIf the analyst needs to draw conclusions about the entire population, this technique is not recommended since the sites collected might not be representative.
Probabilistic sampling techniques are based on statistical approaches, involves random selection process, and therefore allows the analyst to compute a sample size and to draw conclusions about the entire populationIn general, this technique requires more resources because of the increased sample size needed to obtain a representative sampleThere are various different methodologies that can be selected in applying probabilistic sampling techniques including simple random sampling, strategic sampling, cluster analysis and multi-stage random sampling.
Table 17 is adapted from Greene-Roesel et. al. 2010 and provides summary information on each of these sampling techniques
Source: (adapted from) Greene-Roesel et. al. 2010
Sampling size: Similar to the selection of sampling technique, there are various considerations in determining the sample sizeWhen selecting the sample size, it is recommended that the analyst first identifies what is available, and if the data can be obtained by adjusting/combining readily available data or modifying the existing data collection system. This will help to develop an effective and practical approach in sampling and determination of sample size neededIn addition, it is of utmost importance to continuously evaluate the study objectives to effectively use the existing resources.
The following provides resource examples in determining sample size, which might be helpful in making decisions about the sampling size and technique (adapted from Greene-Roesel et. al. 2010):
Evaluate change over time: The analyst might be interested in understanding the change in pedestrian traffic volume at a particular facility location over timeFor example, if the only focus of a study is to conduct a before and after evaluation of one particular intersection in the region, then there is no need to draw information about the general populationIn such cases, non-probabilistic sampling techniques (e.gprofessional judgement) are commonly usedThe sampling focus should be given to collect representative data at that particular intersection taking into account potential biases regarding the time of data collection (e.gseasonal changes)
Evaluate risk related to infrastructure type: The analyst might be interested in comparing the pedestrian safety between different facility locations, such as signalized intersections versus unsignalized intersections in a cityThis will require one to collect two random samples to determine the pedestrian exposure at signalized intersections and pedestrian exposure at unsignalized intersections, respectivelySimple random sampling technique might be appropriate and are easy to apply assuming that pedestrian exposure will be similar across similar sites (i.eminimal variance across the selected sample) and the complete list of targeted intersections are availableIn this case, the following formulae can be used to compute an approximate value of each sample size (Garder 2004):
where n is the sample size, z is the z-value determined based on the desired level of confidence, CV is the coefficient of variation, and ME is the margin of error.
For example, assuming a 95 percent confidence interval (z=1.96), with low variation (e.g., 10 percent), and with acceptable margin of error (e.g., 5 percent), the minimum sample size is computed as 16For the example described above, this will yield to about 32 intersections with 16 of them signalized and 16 of them unsignalizedIt is also important to reexamine the coefficient of variation during data collection, and re-compute the sample size as needed.
Sampling exposure in geographic area: The analyst might be interested in determining exposure across an area for example to assess pedestrian risk in a cityIn this case, a probabilistic sampling technique is needed since the analyst wants to draw information across the entire area and in need of representative sample of pedestrian volume at different facility types and locations across the area.
The analyst can use different probabilistic approachesFor instance, the analyst can apply stratified sampling by choosing stratification variables and their corresponding categoriesLet's assume the analyst identified intersection type (with 2 categories - signalized versus unsignalized) and geographic area (with four categories - CBD, urban, suburban, and rural)Then, based on the above assumptions, the minimum sample size needed can be computed as 128 (16x2x4)This number may also need to be proportionally adjusted based on the shares of each stratum in the region.
While stratified sampling is an effective method in obtaining observations with different levels for the variables used, the analyst may need to increase the number of strata for better representation, which will eventually increase the sample size neededIn that case, the analyst might consider other probabilistic sampling techniquesFor example, cluster sampling can be used by classifying all the intersections into different clusters with similar characteristicsCluster analysis helps control sample size while adding more variables, but might not be as representative as stratified samplingFinally, the analyst can choose to combine clustering and stratification to obtain more representative samples within the clusters as in the case of multi-stage random sampling.
Task 2: Once the site count data are sampled and compiled, the second task of this phase includes processing the data (e.gfactor adjustment) and conducting data quality checks (for a representative sample) needed. Site counts can be conducted for various different durations and at different times of the yearWhile some studies have used the count data directly (i.e., for the specific collection period) as their dependent variable of the models, other studies have processed and expanded the count data to longer time periodsDirect demand models typically require high-quality volume count information that might be supplemented/validated with travel surveys (such as to account for demographics and trip generators)The analysts are referred to Site Counts part of Step 6 of this Guide, which provides detailed information on site counts and key considerations in processing and obtaining a representative sample of site counts to be used as the dependent variable of the modelThis process completes the preparation of the model's dependent variable.
Task 3: Next, the analyst needs to focus on the preparation of explanatory variables of the modelExplanatory variables represent the cause or reason for the outcomeIn statistical terms, explanatory variables are also named as the independent variables of the model whose relationships with the dependent variable are being examinedAt this stage, the analyst needs to first identify candidate independent variables to be considered in the model, and then compile the data needed for the identified variablesIt is likely that the analyst might have an initial (desired) set of independent variables based on the local knowledge, professional judgement, data availability, practicality in usage, etcHowever, it is also likely that the analyst might not have any preference or knowledge on the candidate independent variables to be considered in the modelIt is important to always keep in mind the goals of developing the model when selecting variablesThe final explanatory variables of the model should be composed of variables that are intuitive, logical and relevant to the action items in the decision making process.
Table 18 provides an overview of the key significant variables used across the studies (based on the extensive literature review conducted by Munira and Sener 2017)The analyst is recommended to review the variables in Table 18 before making a final selection of candidate independent variablesThe table provides information on frequency (i.e., use in the model) and impact (i.e., direction of the variable).
The model variables showed some differences based on the mode (i.e., pedestrian model versus bicycle model) and the analysis methodWhile choosing model variables, it is important to consider the context specific nature of explanatory variables of the direct demand modelsAs indicated by Munira and Sener (2017), “choice of independent variables and their magnitude and direction of impact on non-motorized activity largely depend on community, people, and location”For example, while the availability of sidewalks and land use characteristics might be more influential in motivating walking trips, cycling trips might be more likely to be influenced by various factors across spatial areas beyond the trip origin (Munira and Sener 2017; Winters et. al. 2010).
Legend: = to a small extent (1,2); = to a moderate extent (3,4); = to a great extent (>=5).
Source: Based on the literature review of 22 studies conducted by Munira and Sener 2017.
Task 4: Upon identification and compilation of independent variables, the next task includes processing the data and conducting data quality checks neededSeveral different specifications and alternative functional forms of independent variables might need to be considered to identify the best data fit during the development of direct demand modelsFor example, while some variables may need to be considered in categorical or binary forms, some other might need to be used as continuous variables (e.gdevelopment of income categories versus income as a continuous variable)Similarly, some independent variables might work best if they are transformed into other scales (e.gnatural logarithmic scale), which might be helpful for an easier interpretation of model variables as well as better data fit. In addition, all independent variables should be carefully assessed and statistical tests should be performed to ensure the database compiled for independent variables is logical and free of error (e.gidentification of missing values and outliers)This process completes the preparation of the model's candidate independent variables.
Task 5: The final task of the data preparation phase is to combine datasets prepared for dependent and independent variables to obtain one final dataset to be used in model development.
Phase C: Model Development
Objective: The main objective of Phase C is to develop the model based on the data prepared.
Primary Tasks: Phase C includes four key tasks:
- Task 1: Identify statistical method to be adopted for model development.
- Task 2: Perform statistical checks to identify independent variables to be tested in the model.
- Task 3: Estimate the model.
- Task 4: Perform model validation
Processing the Phase:
Task 1: The first task in this phase is to identify the statistical method to describe the relationship between dependent variable and independent variablesA wide variety of methods have been used in predicting non-motorized activity using direct-demand modelsLinear regression, Poisson regression, and negative binomial regression models are among the most commonly used statistical methods used in direct demand models for bicycle and pedestrian exposure estimationIn order to select the best model to the data, the analyst needs to examine the nature of the dataFor example, Poisson distribution assumes that the mean and variance are the same; however, we often found that count data exhibits over-dispersion with a variance greater than the meanIn that case, it has been shown by many studies that negative binomial provides a better data fit
Task 2: In this second task, the analyst needs to screen the pre-identified model variables and the relationships between variablesThe following describes some of the key considerations in evaluating the variables.
- Compute correlations across variables, and screen all variables for high and low correlation values
- Identify independent variables that are relatively highly correlated with the dependent variableÂ
- Check for multicollinearity to avoid including highly correlated independent variables simultaneously in the same model.
Task 3: Once the statistical method is identified and statistical relationships across variables are initially screened, the analysts can start working on the model estimationIn this task, the analyst may need to conduct several model iterations and different combinations of explanatory variables to find the best model structureIt is recommended that the final model is selected based on both statistical and intuitive considerations (e.gstatistical significance, goodness of fit, insights obtained from the literature, practicality, and engineering judgment)The model performance should be evaluated through statistical checks, such as by overall goodness of fit, residual plots, etcA 5 percent level of significance is recommended in general to include variables in the final model; however, the analyst is also recommended to check logical and intuitive variables that might not be very highly statistically significant to reduce any bias.
Task 4: It is important to validate the model, which is defined as the application of the models and comparison of the results to observed data that was not used to estimate the model It is required that the observed data used for model validation are not the same data used for model estimationModel sensitivity tests will also be useful to determine if the model results are reasonable and sensitive to the changes in explanatory variables.
The analyst is recommended to check the process from the selection of statistical method to the identification and examination of variables included in the model until a good model performance is obtainedAs/if needed, the model should be re-specifiedThis process helps develop more robust models with good fit and intuitive explanatory variables that would be useful for both evaluating risk and informing safety policy and investment decisions.
The completion of this task concludes the development of the modelThe exposure output obtained from the model can then be used in risk analysisThe direct demand models can also be used in predicting volumes at locations where the count data are not available, extending the study to an areawide level in the application process.
Examples of Direct Demand Models
Table 19 provides an overview of example direct demand models from the recent literature. The table provides information on the coverage, data collection scale, analysis methods, and significant explanatory variables of the final estimated modelsNext, we provide example studies that have developed and applied direct demand models for exposure estimation in non-motorized safety analysis.
Source: Adapted from Munira and Sener (2017)
This section provides an example study with details to help analysts follow the process described aboveThe example is based on a study conducted by Schneider et. al. (2012), and presents a typical example for direct demand model development for non-motorized exposure estimation.
Development of Pedestrian Intersection Volume Model in San Francisco, California
Phase A – Study Identification: Schneider et. al. (2012) developed a pedestrian intersection volume model in San Francisco, California, focusing on annual pedestrian crossing intersection volume
Phase B – Data Preparation: The authors identified 50 intersections to collect a sample of counts for the San Francisco pedestrian intersection volume modelThe intersections were selected to represent the range or urban characteristics across the cityThe data collection was conducted at different time periods (2009 and 2010)The authors aimed at increasing the geographic representation of locations across the study area by collecting data at various locations, e.ghigh-crash locations, regional count locations, locations near planned or completed projects, locations near key transit hubs, etcNext, the authors applied automated counter, temporal, and weather adjustment factors to extrapolate an annual pedestrian volume estimate from the two-hour counts at the 50 study intersectionsThe logarithm of the annual pedestrian crossing volume constituted the dependent variableFor independent variables, they considered 16 explanatory variables (e.gtotal number of households within 0.25 mile of the intersection without a car, ratio of population to jobs within 0.25 mile of the intersection, intersection in a high-activity zone, etc.)The authors examined descriptive statistics for all variables (i.emean, standard deviation, minimum and maximum).
Phase C – Model Development: Next, the authors developed a log-linear regression model to identify the relationship between annual pedestrian volume estimate and various different explanatory variables including land use, transportation system, local environment, and socioeconomic characteristics near each sampled intersectionAfter conducting various model runs, the authors identified 12 potential models of annual volumes of pedestrian intersection crossingsThe variables with high level of correlation, without precise estimates, and with counterintuitive relationships with pedestrian volume were excluded from the modelThe 12 potential models were indicated to have good fits (adjusted R2-values between .78 and .83) and were significantly better than a model based only on a constant with no independent variables (F-values between 28.4 and 34.4)The final recommended model was selected based on a combination of good overall fit and intuitive, logical and practical explanatory variablesThe model performance was evaluated by spatially reviewing the difference between predicted and observed countsSensitivity tests were conducted, and the model was validated against pedestrian volumes collected at 49 four-way intersections (different than the intersections used in the model estimation) in 2002, which showed that the model ranked intersections similarly to the previous counts in overall volumeThe authors indicated the model included only six significant factors because of the relatively small sample of intersections, and highlighted the importance of various other variables that might need to be considered in future studies.
The model was then used to evaluate pedestrian crossing risk at each intersection based on the exposure measure of the number of pedestrian crashes per 10 million crossings
This section provides additional examples that are briefly described to present different thought process during the development of direct demand models for non-motorized safety analysis
Molino et. al. (2009; 2012) developed a log-linear regression model (with Poisson distribution) to estimate pedestrian counts at signalized intersections in Washington, D.CWhile 15-min pedestrian counts served as the dependent variable, the independent variables of the model included land use variables (e.g., commercial, residential) and characteristics of the day (e.g., day of the week, time of the day)Using the parameter estimates of the model and follow-up adjustment procedures, a total number of miles traveled were estimated “â€¦by multiplying the total number of pedestrians by the mean width of all the sampled signalized intersections.” This result was then used as an exposure measure in pedestrian crash rate computation.
Using a count database of 954 observations and 471 locations, Hankey and Lindsey (2016) employed a stepwise linear regression model that allowed for varying spatial scale of independent variables including land use and transportation network variablesRelying on the modeled values of bicycle traffic from this work, Wang et. al. (2016) then estimated peak-hour bicycle traffic volumes for many segments in Minneapolis. The model results were then converted to bicycling volume for intersections and used for computing bicycle crash rates by intersections and segments.
Strauss et. al. (2013; 2014) used a relatively improved version of modeling to estimate non-motorized demand. First, Strauss et. al. (2013) developed a bivariate Bayesian Poisson model to simultaneously estimate cyclists' injury occurrence and bicycle activity at 647 signalized intersections on the island of Montreal, Quebec, CanadaIn a follow-up study, Strauss et. al. (2014) applied their Bayesian modeling methodology as part of a multimodal approach aimed at examining the safety at intersections for both non-motorized and motorized traffic. After model calibration, the study compared injury and risk between modes and intersections by using the “expected number of injuries (obtained from the models) per million cyclists, pedestrians or motor-vehicle occupants per year” as the expected risk.
This section of the guide provides an overview of the most commonly used travel surveys in estimating exposure including the ACS, NHTS, and regional household travel surveys. This section includes information on the general purpose of the survey and applicability in estimating exposure for bicyclists and pedestrians, limitations and benefits, and data availability.
American Community Survey (ACS)
The ACS is an important tool for tracking non-motorized (bicycling and walking) travel patternsThis national ongoing survey of a sample of U.Shouseholds conducted by the U.SCensus Bureau gathers a wide variety of information in addition to their primary travel mode from home to work.2 The ACS does not have trip information for non-commute trips (whereas NHTS does but only conducted once a decade)Thus, the ACS can be used to estimate non-motorized exposure expressed as the daily person commute trips by walking or bicycling per the specified areawide geography.
Benefits and Limitations
In many cases, non-motorized trips are secondary modes of travel to the longest distance mode (driving or transit). However, it provides one of the most robust sources of information on non-motorized commuting by bicycle and walking in smaller spatial units like Census block groups Table 20 summarizes the strengths and limitations of the ACS in estimating non-motorized trips.
All survey and census estimates incorporate errors to a certain extentSampling error is the quality measure that can adversely affect any survey resultSampling error usually occurs when data are based on a sample of a population rather than the full populationFor every ACS estimate, margins of error are provided and can be easily converted into confidence intervals.3 For different calculations associated with the ACS, it is important to consider sampling error.
The ACS provides estimates for different levels: a) 1-year estimates, b) 3-year estimates, and c) 5-year estimatesA key U.SCensus document4 lists the distinguishing features of 1-year, 3-year, and 5-year estimatesIt is important to note that using 3-year or 5-year ACS is beneficial due to large sample size relative to 1-year estimates, thus reducing margins of error of estimates for small subpopulationsFor analyzing areas with larger population (e.g., states, congressional districts), 1-year ACS is beneficial.5
For spatial units with smaller populations, the ACS samples may have insufficient numbers of households to provide reliable single-year estimatesFor these spatial units, multiple years (3 or 5) worth of data will be merged together to create more reliable estimatesThe multi-year estimates have advantages of statistical reliability for less populated areas and small population subgroups. The level of precision improves considerably with multi-year estimates.
National Household Travel Survey (NHTS)
FHWA conducts periodically the NHTS to gather detailed information on the travel behavior of the American public. The survey collects a wide array data from respondents, including household characteristics, demographics of each member in the households, vehicle details, and trip attributes (mode used, trip length, trip time, trip purpose). These data are stored in separate files: household, person, vehicle, and travel day (i.e., travel diary).
The basic concept of using travel surveys to calculate the amount travel for an area is based on developing estimated statistics developed from the survey sample and expanding those estimates to the population by using weights (for example, see Schneider et. al. .6). Analysts can use the NHTS to estimate the following exposure measures for pedestrian and bicyclist travel:
- Total estimated annual trips
- Total estimated annual miles traveled
- Total estimated annual hours traveled
Benefits and Limitations
The 2009 NHTS national sample estimates are statistically valid down to the state level. However, if additional add-on samples were purchases by a particular state or MPO, then estimates in those areas may be valid at a smaller geography depending on the methodology used by the analyst. Keep in mind that the NHTS documentation warns that standard errors or margins or error should generally be used when looking at estimates at geographies smaller than the national level.
While providing a rich national sample, the NHTS sample sizes might have sparse coverage at small geographic scales. Transferability of the NHTS results to small geographic area (e.g., census tracts) is limited to estimates of average weekday household person trips, vehicle trips, person-miles traveled, and VMT. Though these estimates could serve as exposure measures for non-motorized travel risk, they are not the best choice since specific mode of travel is not offered or is vehicle-based.
Unfortunately, since the NHTS is not conducted on a more frequent and regular basis, it cannot be used to directly track short-term trends in non-motorized travel exposure. It can, however, be used in sketch planning or travel demand modeling efforts to estimate or predict non-motorized exposure based on more current census demographic information. For example, the generalized daily person trips per person by mode, generated from the NHTS, can be used to estimate the total non-motorized trips produced in subsequent years by using current ACS population estimates. These trip rates need to be updated periodically, and if possible, supplemented by more localized travel data to better reflect local nuances and unique characteristics of the transportation infrastructure and traveling public.
In terms of sketch planning, it is possible to apply generalized person trip rates that were produced to represent the statewide population to local areas. This is done by multiplying the person type trip rate by the total number of corresponding population within the local area. This operation has statistical drawbacks since the generalized person trips rates were produced by using a statewide sample, which is likely to not be statistically representative of the local area population. It is always best to use local data for local purposes; however, the NHTS provides an opportunity to estimate local exposure when local data do not exist.
A NHTS was conducted for 1969, 1977, 1983, 1990, 1995, 2001, and 2009. Most recently, a 2017 NHTS that began in March 2016 was released in early 2018It is comprised of a 26,000 national household sample representing all U.SStates and the District and Columbia along with an additional 103,112 add-on sample households. Additional add-on samples are made available to the states and regional/MPOs for purchase. These add-on samples provide the opportunity to populate different exposure measures at a finer geographic level and to develop more robust safety analyses. The 2017 NHTS add-on sample sizes for the state DOTs and MPOs are listed in Table 21.
Source: NHTS Task C: Sample Design, Dec31, 2015, page 5
The NHTS data can be used to compute or statistically model several different exposure measure estimates (e.g., population, miles traveled, number of trips) nationally, by census region/division, state, and urban/rural area types, depending on the survey year. It is possible to produce these same measures at smaller census geographies like Core Based Statistical Area (CBSA) if the particular location(s) participated in the add-on program and were specified in the sampling design.
Regional Household Travel Survey
A regional household travel survey is typically conducted by an MPO to develop a regional travel demand modelThe frequency of these surveys varies from city-to-city, with some planning agencies conducting household travel surveys every eight-to-ten years or longer. Just like the NHTS, regional household travel surveys collect data from respondents on the household characteristics, demographics of each member in the households, vehicle details, and trip attributes via a travel diary. Figure 19 depicts the relationship between the four separate data components. Exposure measures (e.g., miles traveled or number of trips) can be estimated for household and person types and expanded to the population to provide statistically valid areawide estimates.
Benefits and Limitations
Household travel surveys can be used to measure the population proportion, distance traveled, duration traveled, and number of trips by a specific mode for the survey region. Survey respondents typically fill out a travel diary indicating origins and destinations with the start and end times of trips along with the mode that was used. Since the survey represents only a stratified sample of the population, weights must be applied to expand the survey sample so that it represents the entire population of the study area (see Figure 20). Survey weights indicate how many households each survey observation represents of the total population of households – these weights are typically provided along with the survey data.
For example, the regional household travel survey for Austin, Texas could be used to estimate the total amount of time traveled by walking for the five-county region. The survey sample is comprised of 3,000 households and 8,100 persons and can be expanded to represent the population of the study area by applying the survey weights. To do so, the total duration of trips by mode must be enumerated per household type, as defined in the survey stratification. The totals then must be multiplied by their corresponding survey weight to equal the total daily duration of trips by mode for the entire study area. The result is an estimated total of 189,256 daily walk trips with an average trip duration of 16 minutes equaling approximately 50,437 hours of walking per weekday.
The main limitations of region household travel surveys include their high cost and expertise required to process and analyze the survey data. The data may not be publically available due to survey respondent privacy concerns.
Areawide Non-Motorized Exposure Tool
The Areawide Non-Motorized Exposure Tool described here makes it easier for practitioners to obtain and summarize nationwide travel survey data to estimate pedestrian and bicyclist exposure to risk at statewide and MPO area scales. The first part of the tool titled Statewide Exposure Estimates fills that gap between years when the NHTS is conducted by using the more current ACS data to estimate non-motorized exposure at the state-level. The second part of the tool titled MPO Area Exposure Estimates also uses NHTS and ACS data to estimate non-motorize exposure but for individual MPOs throughout the nationBoth parts of the tool produce annual non-motorized exposure estimates by mode for years 2009 to 2016 in terms of trips, miles of travel, and hours of travel for their respective areawide geography. The results are offered in tabular form along with graphics like the examples shown in Figure 21The following sections describe the tool's capabilities, as well as instructions on how to use the tool.
Statewide Exposure Estimates
The FHWA's Safety Performance Management (Safety PM) Final Rule currently requires each State DOT to report the number of non-motorized fatalities and serious injuries (without considering exposure). To understand the relationship between these crashes and non-motorized risk, exposure is desirable to help measure the magnitude of bicyclist and pedestrian vulnerabilityHowever, users should note that, at this time, the Safety PM Final Rule does not require non-motorized exposure to be reported or considered.
The Statewide Exposure Estimates component offers a method for practitioners to estimate statewide non-motorized exposure in order to calculate non-motorized risk. The tool provides the following exposure measure estimates for both bike and walk travel modes per state for the individual years 2009-2016:
- Total estimated annual trips
- Total estimated annual miles traveled
- Total estimated annual hours traveled
The estimates are based on a combination of the 2009 NHTS and the U.SCensus Bureau's ACS data for each respective year. The 2009 NHTS total annualized trips per state are adjusted to better represent the selected year for analysis by using the more current ACS population and daily commute trip estimates (tables B01003 and B08301, respectively). The adjustment factors account for change in both population and the number of commute trips per mode over time. The population adjustment factor (AFpop,i) is based on the 2009 ACS population estimate since the NHTS data represent 2009It can be written as:
AFpop,i = Population adjustment factor in ith year (i = 2009 to 2016) for state
POPi = ACS population estimate in ith year for state
POP2009 = ACS population estimate in 2009 for state
In order to expand daily person commute biking and walking trips, the relationship between commute and total trips is required. The commute trip adjustment factor is based on the 2009 NHTS annualized person trips per person by mode (bike and walk) and the annualized ACS daily persons commuting by mode (bike or walk) for the selected year (i.e., ith year) for analysis. The equation is as follows:
AFCT = Commute trip adjustment factor by mode for state
PT2009 = NHTS annualized person trips by mode in 2009 for state
PC2009 = ACS daily persons commuting by mode in 2009 for state
The adjustment factors (AF) are applied to the selected year ACS commute trips per person by mode to provide estimated annual person trips:
PTi = Estimated annual person trips by mode (biking or walking) in ith year for state
PCi = ACS daily persons commuting by mode in ith year for state
AFpop,i = Population adjustment factor in ith year for state
AFCT,i = Commute trip adjustment factor by mode for state
Finally, to calculate the estimated total annual miles and hours traveled, the 2009 NHTS average trip durations and trip lengths per state were then applied to the total trips to estimate the amount of hours and miles traveled annually per mode for each state.
HTi = Estimated annual hours traveled by mode (biking or walking) in ith year for state
PTi = Estimated annual person trips by mode in ith year for state
TD2009 = 2009 NHTS average trip duration (in hours) by mode for state
MTi = Estimated annual miles traveled by mode in ith year for state
TL2009 = 2009 NHTS average trip length (in miles) by mode for state
Data sources for the above variables are as follows:
|POPi & POP2009||ACS 1-year estimate, table B01003 - Total Population|
|PCi & PC2009||ACS 1-year estimate, table B08301 - Means Of Transportation To Work|
This method assumes that the average trip durations and lengths remain constant between year 2009 and 2016 due to the lack of more current data. However, the tool does provide the user the option to input their own values if available In addition, the tool should be updated with the newly published 2017 NHTS data to produce the 2017 estimates based on current travel behavior data.
The NHTSA Fatality Analysis Reporting System (FARS) person data were used to calculate the total annual non-motorized fatalities per state from 2009 to 2016. The totals are provided in the spreadsheet tool along with total annual risk per state that is based on the total annual non-motorized fatalities per million hours of travel. The total annual non-motorized fatalities are defined as individuals classified as a bicyclist or pedestrian that sustained a fatal injury in a motor-vehicle crash. Â As of May 2018, the 2016 FARS data were incomplete. The data can be found online: https://www.nhtsa.gov/research-data/fatality-analysis-reporting-system-fars
The interface of the Statewide Exposure Estimates component is shown in Figure 22.
Figure 22. Interface of the Statewide Exposure Estimates Tool Component
- 1: Open the spreadsheet containing the tools.
- 2: Read the “Introduction” page for an overview of the available tools.
- 3: Click on the “Statewide Exposure Estimates” tab to select the tool.
- 4: Select the state of interest from the drop down menu titled “State” (also marked by Â in the spreadsheet tool).
- 5: Select the source (default or user input) of the required inputs, “Average Trip Length (Miles)” and “Average Trip Duration (minutes)”, from the dropdown menus highlighted in green for each year (also marked by Â in the spreadsheet tool).
- 6: For those required inputs where the default option is chosen, no further action is required.
- 7: For those required inputs where the user input option is chosen, enter the desired user input value into the appropriate cell (no color highlights) below the source dropdown menu.
MPO Area Exposure Estimates
The MPO Area Exposure Estimates component offers a method for practitioners to estimate MPO-wide non-motorized exposure for calculating non-motorized risk. The tool provides the following exposure measure estimates for both bike and walk travel modes per MPO for the individual years 2009-2016:
- Total estimated annual trips
- Total estimated annual miles traveled
- Total estimated annual hours traveled
Non-motorized exposure estimates at the MPO level are derived from a combination of ACS and the 2009 NHTS. The Census data offers estimates at relatively small geographies that can be interpolated to the MPO leve. lThe 2009 NHTS data provides information on travel behavior for a sample of the travel public from around the nation and can be used to calculate average person trip rate, average trip length, and average trip duration per mode.
Total person trips by bike and walk can be estimated with a generalized 2009 person trip rate per mode applied to the total population of the year of interest and then annualized (365 days). The product is then adjusted to account for any change in the mode-specific commuting population between 2009 and the year of interest. However, this adjustment does not capture any change in non-motorized recreational travel that may be induced from communities investing in bicycle and pedestrian infrastructure. Total person trips are then be applied to the average trip length and average trip duration to equal total miles and total hours traveled per mode, respectively. It also important to note that any error in the 2009 NHTS estimates of walking or bicycling is carried through to the subsequent years.
The MPO-level estimated annual person trips by mode equation is as follows:
PTiÂ = Estimated annual person trips by mode (biking or walking) in ith year for MPO
PTR2009Â = 2009 NHTS average daily person trip rate by mode for CBSA peer group
POPiÂ = ACS 5-year population estimate in ith year for MPO
PCiÂ = ACS 5-year daily persons commuting by mode estimate in ith year for MPO
PC2009Â = ACS daily persons commuting by mode 2009 for MPO
To calculate the estimated total annual miles and hours traveled, the 2009 NHTS average trip durations and trip lengths per MPO are applied to the total trips to estimate the amount of hours and miles traveled annually per mode for each MPO.
HTiÂ = Estimated annual hours traveled by mode (biking or walking) in ith year for MPO
PTiÂ = Estimated annual person trips by mode in ith year for MPO
TD2009Â = 2009 NHTS average daily person trip duration (in hours) by mode for CBSA peer group
MTiÂ = Estimated annual miles traveled by mode in ith year for MPO
TL2009Â = 2009 NHTS average daily person trip length (in miles) by mode for CBSA peer group
Data sources for the above variables are as follows:
|POPi||ACS 5-year estimate, table B01003 - Total Population|
|PCi & PC2009||ACS 5-year estimate, table B08301 - Means Of Transportation To Work|
Several caveats do apply to this methodThe end user should keep in mind that the MPO-level estimates for their area are based on an estimated average for their CBSA (Core Based Statistical Area) peer group. Also, like the statewide tool, the MPO-level method assumes that the average trip durations and lengths remain constant between year 2009 and 2016 due to the lack of more current data. However, the tool does provide the user the option to input their own values if available.
The NHTSA FARS person data were used to calculate the total annual non-motorized fatalities per MPO from 2009 to 2016. The crashes were plotted in a GIS based on the coordinated provided and spatially joined the underlying MPO layerThe totals along with total annual risk per MPO that is based on the total annual non-motorized fatalities per million hours of travel are provided in the spreadsheet tool. The total annual non-motorized fatalities are defined as individuals classified as a bicyclist or pedestrian that sustained a fatal injury in a motor-vehicle crash. As of May 2018, the 2016 FARS data were incompleteThe data can be found online: https://www.nhtsa.gov/research-data/fatality-analysis-reporting-system-fars.
The interface of the MPO Area Exposure Estimates component is shown in Figure 23.
- 1: Open the spreadsheet containing the tools.
- 2: Read the “Introduction” page for an overview of the available tools.
- 3: Click on the “MPO Area Exposure Estimates” tab to select the tool.
- 4: Select the state of interest from the drop down menu titled “State” (also marked by Â in the spreadsheet tool).
- 5: Select the state of interest from the drop down menu titled “State” (also marked by Â in the spreadsheet tool).
- 6: Select the source (default or user input) of the required inputs, “Person Trip Rate”, “MPO Population Estimate”, “Population Adjustment Factor”, “Average Trip Length (Miles)”, and “Average Trip Duration (minutes)”, from the dropdown menus highlighted in green for each year (also marked by Â in the spreadsheet tool).
- 7: For those required inputs where the default option is chosen, no further action is required.
- 8: For those required inputs where the user input option is chosen, enter the desired user input value into the appropriate cell (no color highlights) below the source dropdown menu.
Census Data for MPOs
The Census does not offer data specific to MPO geographies; therefore, tract-level ACS Census data are used to provide the finest resolution to areal interpolate the population and commuter population for the MPOs. Only ACS 5-year estimates are available at the tract level; therefore, the estimates represent a given year within the five-year period as opposed to any individual year. 1- and 3-year estimates are unavailable due to inadequate ACS sample sizes at small geographies (i.e., tracts and counties). Figure 24 offers a visual comparison example of the Census Core-base Statistical Areas (CBSA), MPO and tract geographies for Memphis, TN.
Variables that require ACS data:
POPiÂ = MPO population in ith year (derived through areal interpolation of tract-level ACS data)
PCiÂ = MPO commuter population in ith year (derived through areal interpolation of tract-level ACS data)
PC2009Â = MPO commuter population in 2009 (derived through areal interpolation of tract-level ACS data)
CBSA Peer Grouping Methodology: The 2009 NHTS survey data represent only a sample of the traveling public from both rural and urban areas. A portion of the 2009 NHTS data are labeled as being located within Census Core-base Statistical Areas (CBSA) that represent metropolitan areas. The CBSA geography is the smallest geography in the 2009 NHTS data and the only way to locate a portion of the survey sample.
The survey sample data are grouped by their CBSA as indicated in the original 2009 NHTS data to represent metropolitan areas around the nation. The CBSA metropolitan areas serve as proxies for MPOs in terms of developing of travel estimates from the 2009 NHTS; however, sample sizes vary between CBSAs and are possibly not statistically representative of the local populations. CBSAs are then grouped together based on 2009 ACS 1-year estimates for bicycle and walk commute percentages to increase samples sizes. Tables 22 (bike) and 23 (walk) list the ACS commute percentage ranges for the initial CBSA peer groupings along with their corresponding 2009 NHTS trip sample and the generalized travel estimates.
Generalized Travel Estimates Applied to MPOs: The 2009 NHTS survey data are used to generate an average person trip rate, average trip length (miles), and average trip duration (minutes) for bicycling and walking per CBSA grouping. These generalized travel estimates are applied to the MPOs that possess similar bicycle and walk commute percentages as their peer 2009 NHTS CBSAsThe MPOs area assigned a CBSAs peer grouping for every year between 2009 and 2016 based on the annual release of ACS 5-year estimates for bicycling and walking commute percentages. Refer to the Appendix for a list of the MPOs with their corresponding ACS population and commuter information along with their CBSA bike and walk grouping assignments.
In developing the generalized travel estimates, the NHTS survey weights are not applied because:
- Estimates are for CBSAs with unrepresentative samples
- Unrepresentative samples do not include all segments of the population
- NHTS weights are to replicate the national population (meaning, each person is weighted to be represented within the national population)
Manual Data Extraction
The following section offers details on how to the manually extract data from the ACS and NHTS sources. These data sources offer a variety of additional demographic and travel behavior information that may be of value to safety analysis projects or outreach efforts.
ACS Data Extraction
ACS provides pedestrian and bicycle commuting estimates in geodatabase and CSV formats.7 Data are available for different measures related to pedestrian and bicycle commutingUsers can download data based on their requirementsTo get estimate the number of workers that commute to work by walking or bicycling, four major variables can be used (see Table 24).
To create maps, users can use ESRI ArcGIS software to join TIGER/Line (Topologically Integrated Geographic Encoding and Referencing) shapefile with the ACS data tables.
Example Problem: Determine recent bicycle commuting estimates for Census Tracts in Texas
With the following steps, users can determine bicycle-commuting estimates for census tracts in TexasIt is important to note that manual data extraction requires expertise in ArcGIS tool.
2: By using ArcMap 10.3, users can join ACS shapefile with ACS commuting data provided in the geodatabaseShapefile's variable 'GEOID' is required to be joined with 'GEOID' variable in ACS commuting data (see Figure 26).
3: Export data to text file format from the joined ACS shapefile attribute table.
4: To generate a choropleth map (see Figure 27), users need to follow the following steps:
- Right click on the generated shapefile to select properties
- Under the 'Symbology' tab, select Quantiles and Graduated Maps
- Select variable 'B08301e18' from the drop-down lists in values.
NHTS Data Extraction
The following section highlights NHTS online tools that provide relevant pedestrian and bicyclist travel summaries at the state and CBSA geographies based on the NHTS data files as presented in Table 25.
Since these data are collected via a survey, and not a census, they must be weighted according to ACS demographic information to represent the entire population and produce valid estimates. Survey weights for households and persons are available for all useable households in the NHTS databases.
The 2009 and 2017 NHTS national sample datasets (i.e., adjusted for oversampling due to add-ons) offer the ability to estimate non-motorized travel exposure by person trips at the state level. Figure 28 shows the online analysis tool with the total annual person trips by mode per state from the 2009 NHTS. The tool provides choropleth maps of 2009 NHTS person trips per state with drill-down capability to state statistics on trips, mode, and purpose. For example, Colorado produced approximately 7 billion person trips in 2009 with 9.7% (approximately 675 million) being walking trips.
Another online analysis tool based on the 2009 and 2017 NHTS national sample datasets is the table designer shown in Figure 29 and Figure 30. The table designer tools allow users to build customized data tabulations quickly and easilyThe tabulation outputs are offered in either HTML, Excel, or CSV formats.
The NHTS attributes can also be tabulated for various census-based geographies, such as state, CBSA, and rural/urban (http://nhts.ornl.gov/2009/pub/UsersGuideClaritas.pdf)For example, the exposure measure of annual person miles of travel can be calculated by mode per state with these toolsTable 26 shows that in 2017 Arizona generated approximately 114 billion person miles of travel, of which 0.5% (approximately 585 million) were by walking.
1 Rails-to-Trails Conservancy is developing nationwide seasonal adjustment factors in its Trail Modeling and Assessment Platform (T-MAP), as is Portland State in its PORTAL non-motorized count database
7 American FactFinder https://factfinder.census.gov/faces/nav/jsf/pages/index.xhtml
8 TIGER/Line® with Selected Demographic and Economic Data https://www.census.gov/geo/maps-data/data/tiger-data.html Accessed on Sept 29, 2017