Driving pattern interpretation based on action phases clustering

Xue Yao, Corresponding Author
Department of Transport & Planning, Delft University of Technology, Delft, the Netherlands
Email: X.Yao-3@tudelft.nl
Simeon C. Calvert
Department of Transport & Planning, Delft University of Technology, Delft, the Netherlands
Email: s.c.calvert@tudelft.nl
Serge P. Hoogendoorn
Department of Transport & Planning, Delft University of Technology, Delft, the Netherlands
Email: s.p.hoogendoorn@tudelft.nl


1 Abstract↩︎

Current approaches to identifying driving heterogeneity face challenges in comprehending fundamental patterns from the perspective of underlying driving behavior mechanisms. The concept of Action phases was proposed in our previous work, capturing the diversity of driving characteristics with physical meanings. This study presents a novel framework to further interpret driving patterns by classifying Action phases in an unsupervised manner. In this framework, a Resampling and Downsampling Method (RDM) is first applied to standardize the length of Action phases. Then the clustering calibration procedure including “Feature Selection”, “Clustering Analysis”, “Difference/Similarity Evaluation”, and “Action phases Re-extraction” is iteratively applied until all differences among clusters and similarities within clusters reach the pre-determined criteria. Application of the framework using real-world datasets revealed six driving patterns in the I80 dataset, labeled as “Catch up”, “Keep away”, and “Maintain distance”, with both “Stable” and “Unstable” states. Notably, Unstable patterns are more numerous than Stable ones. “Maintain distance” is the most common among Stable patterns. These observations align with the dynamic nature of driving. Two patterns “Stable keep away” and “Unstable catch up” are missing in the US101 dataset, which is in line with our expectations as this dataset was previously shown to have less heterogeneity. This demonstrates the potential of driving patterns in describing driving heterogeneity. The proposed framework promises advantages in addressing label scarcity in supervised learning and enhancing tasks such as driving behavior modeling and driving trajectory prediction.

Keywords: Driving behavior interpretation, Driving pattern, Action phase, Clustering analysis, Clustering calibration procedure

2 Introduction↩︎

Driving heterogeneity, recognized as the differences in driving behaviors exhibited by different driver/vehicle combinations under similar conditions, is widely acknowledged [1]. Studies have shown that heterogeneity in driving behavior can lead to a rise in traffic accidents, congestion, and emissions [2], [3]. Further, user acceptance of autonomous vehicles (AVs) has been found to hinge on accurately comprehending and emulating driving heterogeneity of human-driven vehicles (HDVs), such as human drivers’ preferred driving styles [4]. Thus, understanding driving heterogeneity plays a significant role in enhancing traffic operations and enabling manufacturers to design safe and efficient automated vehicles at various levels.

Existing studies have addressed driving heterogeneity by personalizing driving styles based on driving behavior data such as vehicle kinematics variables (e.g., velocity and headway) and vehicle dynamics variables (e.g., braking and throttle opening), which enable categorizing drivers into several groups [5]. For instance, [6] classified drivers into two categories (i.e., normal, and aggressive) based on velocity and throttle opening data. In another study, overtaking maneuvers were identified as low-medium-high risk levels based on speed and distance between vehicles [7]. Other studies have utilized car-following model parameters to distinguish driving styles [2]. While this method captures drivers’ static driving characteristics, it is not capable to describe the inherent traits of driving behavior. This is because driving behavior is a dynamic decision-making process [5], drivers may exhibit heterogeneous driving styles in different traffic scenarios. Even under the same traffic scenario, the same driver’s behaviors might vary at different time intervals.

Some studies found that driving heterogeneity can be derived by decomposing driving behavior into distinct driving patterns. As driving behavior displays certain characteristics during the transition of driving maneuvers [8]. As such, some researchers [9], [10] segmented driving behavior data into primitives with unique characteristics. Typically, these primitives referred to distinct driving patterns, specifically known as primitive driving patterns. By doing so, the characteristics of driving behavior can be accurately captured by corresponding the traffic environment and driving maneuvers. Using supervised learning approaches, different patterns were extracted and assigned with semantic labels (e.g., rapidly closing in, falling behind) by learning sample features like vehicle operating data [5]. However, pre-labeling tasks are labor-intensive, limiting the implementation of supervised learning technologies in driving pattern recognition [11]. As a result, there is a growing interest in semantic analysis using unsupervised techniques. [12] identified a specific set of state-action clusters and employed them to characterize potential driving patterns of passenger car and truck drivers. Employing a hierarchical Dirichlet process-Hidden semi-Markov Model (HDP-HSMM), [13] extracted 75 primitive driving patterns from time series driving data. This method allows for the identification of a wider variability in driving behavior by encompassing different driving characteristics. However, an excessive number of patterns, for example, 75, may limit the categorization’s effectiveness due to reduced interpretability. This expansive classification has limitations in fully clarifying fundamental driving behaviors and understanding driving heterogeneity. As a result, continued research in this field is required to overcome these challenges.

In our previous research [14], the concept of Action phases was introduced to capture driving characteristics with physical meanings, aiming to pave the way for identifying driving heterogeneity. It expanded the scope of the “action point” [15] by incorporating additional variables to provide more comprehensive information about driving behavior. The action trend space for each driving behavior variable is represented as \(S = \{I, D, H, L\}\), which denotes ‘Increasing’, ‘Decreasing’, ‘Stable in a high value’, and ‘Stable in a low value’, respectively. One Action phase carries a label name with each variable having a single action trend in it. All the Action phases obtained from a certain dataset constitute the Action phases Library of the dataset, representing all driving behavior characteristics under this specific traffic flow condition. However, the dispersion of variables can complicate the uniform interpretation of driving behavior. Furthermore, when additional variables are considered, the quantity of label names escalates rapidly. Consequently, many Action phase, despite having different label names, might exhibit minor differences in driving behavior. In this situation, consolidating Action phase with similar characteristics into a smaller number of patterns can assist in interpreting driving behavior through the analysis of group-specific characteristics.

To bridge these research gaps, this study presents a novel framework to enhance the comprehension of driving behavior by classifying Action phases into various driving patterns. The unique contributions of this study include: (i) By using Action phases with physical meanings, this unsupervised learning method holds dual advantages in eliminating pre-defined bias and ensuring behaviourally interpretable results. (ii) The clustering calibration process greatly assists in deriving driving patterns with clear categorization and high internal similarity. Evaluation using real-world datasets demonstrates various driving patterns with unique characteristics, accurately reflecting empirically observed driving behaviors. The findings also indicate the prospective advantages of using driving patterns to illustrate the heterogeneity in driving behavior.

3 Methodology↩︎

In this section, we first give an overview of the framework for interpreting driving patterns. Then the methods and techniques employed in each step are explained.

3.1 Overall framework↩︎

The objective of the framework is to categorize Action phases into several driving patterns by utilizing an Action phase Library. These patterns consist of different Action phases that exhibit common traits. The Action phase Library for each dataset was built in our previous work, which is briefly shown as the Data Preprocessing step in Figure 1. Due to the limitations of the high-dimensional data used as input [2], Feature Selection is first conducted. If there is insufficient information regarding feature importance, particularly during the initial analysis phase, an unsupervised feature extraction method is utilized. After the variable importance is evaluated, the features are selected according to the weights of variable importance. The extracted features serve as input for the subsequent Cluster Analysis with an unknown cluster number \(k\). The clustering results are evaluated from both inter-class and intra-class perspectives, demonstrated in the Similarity/Difference Evaluation step. In inter-class evaluation, the differences (df, indicating the distance) among clusters are calculated. As for intra-class, the shape similarities of variables between Action phases within the same cluster are assessed using a dissimilarity index (dSI). Specifically, the inter-class evaluation takes precedence over the intra-class, implying that only clusters with substantial differences from each other are subjected to internal similarity evaluation. If the differences are less than a specific threshold \(\delta\), the clustering results are directly accepted for driving pattern description. In the intra-class evaluation, Action phases that display a dSI greater than the threshold \(\epsilon\) will be re-extracted and re-analysis in the next analysis round. The remaining Action phases, which do not exceed this threshold, are maintained within the current cluster, contributing to the interpretation of driving patterns.

Figure 1: General framework of driving pattern interpretation.

Notice that certain Action phases are re-extracted based solely on one variable, while others are determined by multiple variables. This implies that some variables contribute significantly to the dissimilarity between Action phases and have a more substantial role in the Action phases Re-extraction process. We hypothesize that high-frequency variables carry more informative content reflecting driving behavior characteristics. These variables, therefore, take greater importance in explaining driving behavior. An importance score (IS) is subsequently computed to quantify the importance of driving behavior variables. Feature Selection, commencing from the second round of analysis, is guided by the variable IS obtained in the previous round, giving this process clear interpretive significance.

The clustering calibration process of “Feature Selection”, “Clustering Analysis”, “Difference/Similarity Evaluation”, and “Action phases Re-extraction” (marked as gray background in Figure 1) is iterated until all difference and similarity indices meet the pre-defined criteria. Ultimately, the optimal number of clusters is determined. Each cluster, representing a driving pattern, is interpreted based on the characteristics of the Action phases it contains. This framework enables the transformation of extensive sets of Action phases into more concise driving patterns, providing a semantic description of various driving behaviors. The techniques adopted at each stage will be detailed in the following subsections.

3.2 Data preprocessing↩︎

The Action phases accommodated driving trajectory segments with varying lengths in order to provide a more detailed representation of driving behavior characteristics. This refinement adds a layer of complexity to subsequent analyses, as most algorithms necessitate input data of equal length. Various methods have been suggested to mitigate this issue. One prevalent approach is padding, where sequences are extended with a specific value (commonly zero) to match the length of the longest sequence. Alternatively, truncation is utilized, wherein sequences are grouped into “buckets” of similar lengths (for instance, 1-10, 11-20, 21-30, etc.), followed by padding within each bucket. There are also studies that employed variable-length Recurrent Neural Networks (RNNs) to convert sequences into fixed-length representations.

Given that the time series data in Action phases holds interpretable physical significance, the padded zeros also convey explicit meanings, such as \(0 \mathrm{m/s}\) or \(0 \mathrm{/s^{2}}\), which can significantly change the original characteristics of Action phases. It is important to note that as the fundamental unit for describing driving behavior, the lengths of Action phases can range from 2 seconds to 100 seconds, representing the duration of the driving behavior. Under this circumstance, the truncation method risks the loss of substantial information. The data processed by RNNs also suffers from a significant drawback - limited interpretability, which makes it unsuitable for data processing in this study.

To standardize the length of input data while preserving the information of the original Action phases to a large extent, the Resampling and Downsampling Method (RDM) is utilized. Initially, the median lengths of all Action phases are determined to serve as a reference value. Then Action phases shorter than this reference value are resampled to match the reference length using Fast Fourier Transform (FFT) and Inverse Fourier Transform (IFFT) [16]. In parallel, isometric extraction is implemented to truncate Action phases exceeding the reference length down to the standard length. An example of fixing Action phase length with RDM is illustrated in Figure 2.

Figure 2: Fix Action Phase length with Resampling and Downsampling Method (RDM).

3.3 Feature Selection↩︎

In the initial stage of clustering, an unsupervised feature extraction approach is employed due to the lack of clear information about the significance of variables. Principal Component Analysis (PCA) [17], Kernel PCA, t-SNE, among others, are frequently utilized methods for this purpose.

Upon completion of the first analysis round within the framework, the importance of variables can be evaluated. Consequently, for the subsequent stages of clustering, feature extraction is guided by the variable importance score (IS) acquired from the preceding analysis round. Borda Count [18], a highly recognized example of weighted scoring rules, is commonly applied in multi-candidate, single-winner electoral procedures. This method, which also provides the votes of each candidate during the selection of the winner, is adopted here for determining the importance of driving behavior variables.

Suppose the election has \(m\) candidates (variables), \(m = 1, 2, ..., M\), The ballots cast information is shown in matrix \(X\),

\[X = \begin{bmatrix} x_{1,1} & x_{2,1} & \cdots & x_{1,u} \\ x_{2,1} & x_{2,2} & \cdots & x_{2,u} \\ \vdots & \ddots & \vdots \\ x_{m,1} & x_{m,2} & \cdots & x_{m,u} \end{bmatrix}\]

here, \(x_{m,u}\) signifies the count of the \(u\)-th ballots secured by the candidate \(m\). Notably, ballot types are determined by the number of combined variables that re-extract the Action phases demonstrating low similarity, hence \(m=u\) in this situation.

The weight of each ballot type is represented by a score vector \(\alpha = (\alpha_1, \alpha_2, ..., \alpha_u)\) that satisfies \(\alpha_1 \ge \alpha_2 \ge \ldots \alpha_u\). It is normalized and specifically set as \(\{1, \frac{1}{2}, \frac{1}{3}, ..., \frac{1}{u}\}\) in the proposed framework.

Utilizing the ballot information from matrix \(X\) and the score vector \(\alpha\), the weighted Borda Score [18] of variable \(m \in V\) is computed by Equation 1 . Then the importance score (IS) of variable \(m \in V\) is obtained by normalizing the results.

\[\label{eq:bordaScore} Sc(m) = \sum_{i=1}^{u} \alpha x_i (m)\tag{1}\]

3.4 Clustering Analysis↩︎

The clustering analysis in our framework aims at finding typical driving patterns in a given dataset. Various clustering approaches with inherent techniques have been proposed, this is due to the fact that there is no such precise definition to the notion of “cluster” [19]. According to [20], the clustering approaches are divided into two different groups: hierarchical and partitioning techniques. Hierarchical clustering is chosen as the principal clustering method for two main reasons. The first is its capability to form a hierarchical depiction of the provided dataset, which in essence gives an outline of the distribution of driving patterns. The second advantage is that hierarchical clustering offers reproducibility of the resulting clusters [21]. This mitigates the sensitivity to random initializations commonly encountered by partitioning clustering techniques such as k-means.

Hierarchical clustering can function in two ways - agglomerative (bottom-up) and divisive (top-down) - both of which are intrinsic strategies for building a binary tree. An agglomerative strategy is employed here, as it begins with each pattern as an individual cluster and inspects the connections between patterns or intermediate clusters. The fundamental concept is to merge the two closest patterns or intermediate clusters into a new, larger cluster. The proximity between any two patterns is determined using the Manhattan distance, also known as the city block distance [22]. When assessing the distance between two clusters, the average-link scheme is implemented, which measures the mean distance across all pairs of Action phases from those clusters (refer to Equation 2 ). This procedure is repeated until only a single cluster is left, indicating all Action phases have been grouped into the same cluster.

\[\label{eq:hierarchicaltheory} d(p, q) = \sum_{l=1}^{D} |p(l) - q(l) |\tag{2}\]

\[\label{eq:distance} d(R, \mathcal{S}) = \frac{\sum_i^{N_{\mathcal{R}}} \sum_{j}^{N_{\mathcal{S}}} d \left(x_i^{\mathcal{R}}, x_j^{\mathcal{S}}\right)}{N_{\mathcal{R}} \times N_{\mathcal{S}}}\tag{3}\]

here, \(p\) and \(q\) denote two feature vectors that each represent a unique Action phase, with \(D\) as the total number of dimensions. \(p(l)\) denotes the \(l\)-th element present within vector \(p\). \(\mathcal{R}\) and \(\mathcal{S}\) symbolize two distinct clusters. \(N_\mathcal{R}\) and \(N_\mathcal{S}\) are indicative of the quantity of Action phases in clusters \(\mathcal{R}\) and \(\mathcal{S}\) respectively. \(x_l^\mathcal{R}\) stands for the feature vector that represents the \(i\)-th Action phase located within cluster \(\mathcal{R}\). \(d(p, q)\) calculates the total sum of the absolute differences between corresponding elements of vectors \(p\) and \(q\), which is the standard computation for Manhattan distance.

The term \(d(R, \mathcal{S})\) in Equation 3 computes the average distance between all possible pair combinations of Action phases from clusters \(\mathcal{R}\) and \(\mathcal{S}\). This is achieved by adding up the distances between every possible pair and then dividing by the total number of such pairs. The total number of pairs is given by multiplying the count of Action phases in each of the two clusters.

3.5 DTW-based Similarity Evaluation↩︎

The key to effective clustering is ensuring significant distances between clusters while also maintaining high internal similarity. Each variable in the Action phases can be considered as a set of time-series data. However, the challenge with comparing the similarity of two time-series data points is that even if they share similar characteristics or trends, they might not align along the time axis. For instance, two sets of velocities could demonstrate similar trends over time, but one may occur at a faster or slower rate than the other. This is where Dynamic Time Warping (DTW) proves useful.

The main idea behind DTW is to compare the distances of two sequences under all possible “warpings”, and then identify the optimal match among these warpings [23]. Consider two sequences, \(X=[x_1, x_2, ..., x_n]\) and \(Y=[y_1, y_2, ..., y_m]\) of lengths \(|X|\) and \(|Y|\) respectively. A warp path \(W\) is then created, as shown in Equation 4 .

\[\label{eq:warpPath} W = w_1, w_2, ..., w_K \; \; \max(|X|,|Y|\leq K < |X|+|Y|)\tag{4}\]

here, \(K\) signifies the length of the warp path, and the \(k\)-th element of the warp path is \(w_k = (i,j)\), where \(i\) and \(j\) are indices of time series \(X\) and \(Y\), respectively. The warp path initiates at the beginning of each time series at \(w_1 = (1, 1)\) and finishes at the end of both time series at \(w_K = (|X|, |Y|)\). A constraint on the warp path mandates \(i\) and \(j\) to be monotonically increasing in the warp path. Every index of both time series must be engaged. This constraint can be expressed more formally as follows:

\[w_k = (i, j), \;w_{k+1} = (i', j') \; \; i \leq i' \leq i+1, j \leq j' \leq j+1\]

The optimal warp path is the one with minimum distance, where the distance (or cost) of a warp path \(W\) is

\[Dist(W) = \sum_{k=1}^{k=K} Dist(w_{ki}, w_{kj})\]

\(Dist(w_{ki}, w_{kj})\) represents the distance between the two data point indices (one from \(X\) and one from \(Y\)) in the \(k\)-th element of the warp path.

Dynamic programming is then deployed to identify this minimum-distance warp path between the two time series, providing the best match between them. This can be expressed as:

\[D(i, j) = Dist(i, j) + \min[D(i-1, j), D(i, j-1), D(i-1,j-1)]\]

However, the complexity of this algorithm is \(O(N^2)\). When the time series is considerably long, the efficiency of the DTW reduces, thus being unable to meet the needs. Consequently, FastDTW was developed, providing a linear and accurate approximation of dynamic time warping. FastDTW utilizes a multilevel strategy that recursively projects a warped path to a higher resolution and fine-tunes it. The three critical operations in this process include [23]:

1) Coarsening – Reducing a time series into a smaller time series that accurately represents the same curve with fewer data points.

2) Projection – Identifying a minimum-distance warp path at a lower resolution, and utilizing it as an initial guess for a higher resolution’s minimum-distance warp path.

3) Refinement – Refining the warp path projected from a lower resolution by locally adjusting the warp path.

These operations effectively lower the time complexity to \(O(N)\). While the strategy to reduce the search space may increase the error, these errors usually remain within an acceptable range [23].

4 Evaluation of the clustering calibration process↩︎

Evaluation of the clustering calibration process is designed and presented in this section. First, data preparation including Action phases review and feature selection results is introduced. Then, hierarchical clustering is conducted, and the results are evaluated by using a qualitative measure for inter-class difference and a quantitative measure for intra-class similarity. Action phases demonstrating low intra-class similarities are extracted, during which process the variables’ importance is calculated.

4.1 Data and Feature Selection↩︎

As proposed in [14], driving behavior trajectories are segmented to yield Action phases, with each driving variable in these phases displaying a single trend. For example, \((D, L, L, I)\) indicates the vehicle’s velocity (\(v\)) has a trend of decreasing, the acceleration (\(a\)) and distance (\(d\)) are keeping a low value, and the speed difference (\(\Delta v\)) is increasing. All the Action phases extracted from one dataset (representing a specific traffic flow condition) constitute the Action phase Library of this dataset, which is adopted as the initial data in this study. The total size of the Action phase Library amounts to 1764 for the I80 dataset and 13564 for the US101 dataset.

a

I80 dataset.

b

US101 dataset.

Figure 3: Distribution of PC1’s cumulative contributions.

Action phases consist of trajectory information using four variables, making feature extraction crucial for high-quality cluster analysis. This study employs Principal Component Analysis (PCA), an unsupervised feature selection method, to cohere variables and extract significant features. The results are displayed in Figure 3. Notice that in the I80 dataset, the cumulative contribution of the first Principal Component (PC1) exceeds 80% for 90% of Action phases, as indicated by the red dotted line in Figure 3 (a). Similarly, PC1 has a considerable cumulative contribution of over 75.82% in the US101 dataset, as shown in Figure 3 (b). Consequently, PC1s are selected and used as the input for subsequent clustering analysis.

4.2 Clustering Analysis↩︎

Determining a suitable validation strategy for an unsupervised learning problem is acknowledged to be a complex task. The literature generally offers two validation criteria: internal and external [24]. The external criterion compares the clustering outcome with existing knowledge of the dataset’s structure (commonly referred to as true labels), whereas such information is usually subjective or unavailable. As a result, this study concentrates on the internal criterion, which assesses clustering results based on the inherent properties of the dataset. The evaluation undertaken in this study includes i) a qualitative measure for inter-class by visually observing the dendrogram to ascertain the differences between clusters, and ii) a quantitative measure for intra-class by calculating the similarity of Action phases within a cluster.

a

I80 dataset.

b

US101 dataset.

Figure 4: Dendrogram representations of hierarchical clustering results..

4.2.1 Qualitative evaluation↩︎

Clustering results are evaluated by examining different (intermediate) clusters, i.e., branches, demonstrated by the two dendrograms shown in Figure 4. The metric is based on the similarities among the Action phases within the same cluster. Focusing on the dendrogram for the I80 dataset depicted in Figure 4 (a), the first analysis involves the two highest sub-trees (or branches): The left sub-tree comprises small, simple patterns of Action phases, as colored by orange; conversely, the right sub-tree displays a variety of patterns. Thus, further examination is conducted on the right sub-tree, revealing that (i) the left sub-subtree contains small, straightforward patterns of Action phases; (ii) despite the right sub-subtree demonstrates a diversity of patterns, they occur at close distances. Following this analysis, three clusters are identified in the I80 dataset (marked as Cluster 3, 2, and 1), and two clusters are observed in the US101 dataset (marked as Cluster 2 and 1). As shown in Figure 4 (a) and Figure 4 (b), the distance between clusters obtained from the two datasets exceeds 2, surpassing the pre-established threshold \(\delta = 1\). Hence, the analysis continues with evaluating the similarity within the cluster.

4.2.2 Quantitative Evaluation↩︎

FastDTW compares the shape similarity by calculating the minimum-distance warp path between two variable data sets. This minimum distance is used as a dissimilarity index (dSI) to provide a quantitative assessment of the clustering results. A value closer to 0 implies higher similarity between the two variable sequences. Conversely, a larger dSI suggests less similarity in their shape. The similarity assessment of the I80 dataset is depicted in Figure 5. (a) - (c) represent the three clusters obtained from hierarchical clustering, while 1 - 4 denotes the four variables \(v\), \(a\), \(d\), and \(\Delta v\) considered in this study. All figures use a uniform color bar as a reference, with darker colors indicating greater distances or less similarity. As represented by the deeper blue lines, larger dSI values are observed in both Cluster 2 and Cluster 3, suggesting these Action phases have a relatively low similarity. In contrast, Cluster 1 generally has low dSI values, signifying a high degree of similarity between Action phases in this cluster. The same analysis is also carried out on the US101 dataset, as shown in Figure 6.

Figure 5: Dissimilarity Index according to fastDTW - I80 dataset.

Figure 6: Dissimilarity Index according to fastDTW - US101 dataset.

To improve the intra-class similarity, Action phases with a dSI index for variables exceeding a threshold value \(\epsilon\) will be re-extracted for the next round of analysis. In this context, the threshold \(\epsilon\) is defined as the 99th percentile of all dSI values in a given dataset. Consequently, a total of 1054 and 11342 Action phases are re-extracted from the I80 and US101 datasets, respectively. More specifically, from the I80 dataset, 1089 Action phases are extracted from Cluster 1, while 126 and 133 Action Phases are from Cluster 2 and Cluster 3 respectively. As for the US101 dataset, the numbers are 9258 and 2084 for Cluster 1 and Cluster 2, respectively.

4.3 Variable Importance Evaluation↩︎

Figure 7 (a) and Figure 8 (a) display the count of each variable involved in the re-extraction process. The variable \(v\) appears to be the most recurrent one used in Action phases re-extraction in both datasets, succeeded by \(a\), \(d\), and \(\Delta v\), in that sequence. It is worth noting that an Action phase could be selected due to the marked dissimilarity in one or multiple variables, as the statistical data shown in Figure 7 (b) and Figure 8 (b). In the I80 dataset, the selection of most Action phases relies on two variables, whereas in the US101 dataset, all four variables commonly participate. Figure 7 (c) and Figure 8 (c) illustrate various combinations of these variables used in the re-extraction process. The combination of all four variables displays high frequency in both datasets. When it comes to univariate extraction, velocity \(v\) exhibits the highest frequency, and for bivariate extraction, the combination of \(v\) and \(a\) prevails.

In this study, the \(m\) candidates in Borda count are the four variables, namely \(v\), \(a\), \(d\), and \(\Delta v\). It is assumed that a variable that is individually involved more in the re-extraction process holds greater significance. As such, four different types of ballots are generated, leading to the score vector \(\alpha = \{1, \frac{1}{2}, \frac{1}{3}, \frac{1}{4}\}\). The ballots matrix \(X\) for these variables is displayed in Table 1 and Table 2. Subsequently, the weighted Borda Score (wBS) for each variable is calculated and normalized to yield the Importance Score (IS) of driving behavior variables. The outcomes are \(IS_{80} = [1.0, 0.833, 0.831, 0.223]\) and \(IS_{101} = [1.0, 0.674, 0.532, 0.411]\). Higher scores represent higher importance. This result will serve as a guide for selecting features of Action phases in the subsequent round of analysis.

Figure 7: Statistics of variable contribution in re-extracting Action Phases - I80 dataset.

Figure 8: Statistics of variable contribution in re-extracting Action Phases - US101 dataset.

Table 1: Variable ballot count - I80 dataset
Combination Num. Velocity(\(v)\) Acceleration(\(a\)) Distance(\(d\)) Speed difference(\(\Delta v\))
1 84 1 0 0
2 451 451 0 0
3 231 231 231 0
4 440 440 440 440
Table 2: Variable ballot count - US101 dataset
Combination Num. Velocity(\(v)\) Acceleration(\(a\)) Distance(\(d\)) Speed difference(\(\Delta v\))
1 1199 0 1 0
2 1046 1046 0 0
3 1325 1328 1325 0
4 6050 6050 6050 6050

4.4 Analysis on Re-extracted Action phases↩︎

Different from the initial round of analysis where feature selection is based on an unsupervised learning method, this round of analysis employs features selected based on the variable importance score (SI) computed in the prior round. Hierarchical clustering is also executed, with results depicted in Figure 9. An observation of the dendrograms reveals three clusters in the I80 dataset and two clusters in the US101 dataset (see the sub-trees with different colors). Notice that sub-trees in each cluster are with small distances, or df, significantly smaller than the threshold \(\delta = 1\), indicating high similarity among Action phases within the cluster. Therefore, the results are directly considered as patterns obtained in this round of analysis. Combined with the updated cluster from the previous iterative analysis, the total number of clusters is obtained for each dataset, with each cluster representing a unique driving pattern.

a

I80 dataset.

b

US101 dataset.

Figure 9: Dendrogram representations of hierarchical clustering results..

Finally, 6 clusters in I80 dataset and 4 clusters in US101 dataset are found through the clustering calibration process. As the finding mentioned in [14], the traffic flow in US101 dataset exhibits less heterogeneity than that in the I80 dataset. Likewise, a smaller number of driving patterns are identified in the latter dataset compared to the former. This consistency provides a connection between driving patterns and the description of driving heterogeneity.

5 Driving Pattern Interpretation↩︎

Driving patterns are represented by the final results of the clustering calibration process, and the interpretation for each pattern is provided in this section. Based on a measurement to distinguish different driving patterns, i.e., the slope, a general analysis of the results is first presented. Then driving patterns obtained from I80 dataset and US101 dataset are analyzed, respectively.

5.1 General Analysis of Driving Patterns↩︎

The concept of “Trend” has consistently been a key of our Action phases analysis, and it forms the basis for characterizing driving patterns. Trends are illustrated through the slope, symbolizing the rate of change in a given variable. A positive slope portrays an increasing trend, such as an increase in velocity, while a negative slope implies a decreasing trend. The magnitude of the slope reflects the rate of change. Especially, several adjacent gentle slopes form fluctuations, representing a ‘Keeping’ trend of variables. Given that variables usually manifest identical trends in different ways, such as linear increase, convex/concave progression, or slightly fluctuating increase, linear regression may struggle to precisely identify variable trends. To capture local trends within specified intervals in the dataset and retain overall trend accuracy, we employ a ‘sliding window’ method. This technique involves forming a ‘window’ of a particular size (e.g., 5 periods, 10 periods, etc.) which ‘slides’ over the data points in the series [25]. A linear regression is computed at each window position and the slope of the regression line is recorded. The final slope of the variable data, which serves as the trend index for each variable, is derived by averaging the slopes of these windows.

Figure 10 and Figure 11 use boxplots to illustrate the statistical data of the variable trend index for both datasets, respectively. Each boxplot displays the upper quartile, lower quartile, and median of the trend indexes. The whiskers extend to the farthest data points not deemed outliers, while outliers (if present) are denoted by asterisks. As previously mentioned, velocity holds the highest importance score when it comes to identifying dissimilarities within Action phases, thus it is prioritized during analyzing the characteristics of driving patterns. Given that the speed difference \(\Delta v\) signifies the interaction between vehicles and their preceding vehicles, it is also taken into account when describing driving patterns. Based on domain knowledge, if the velocity increases while the speed difference decreases, this represents the situation where the target vehicle is catching up with the vehicle in front. Conversely, a decrease in velocity accompanied by an increase in speed difference represents a situation of keeping away from the preceding vehicle. If both the velocity and speed difference remain roughly constant, it indicates that the vehicle is maintaining its current state.

5.2 Driving Patterns in I80 dataset↩︎

Figure 10 displays the trend index of variables for each cluster using the I80 dataset, where each sub-figure corresponds to a driving pattern. Specifically, Figures 10 (a)-(c) represent patterns derived from the first round of clustering, while Figures 10 (d)-(f) represent those from the second round. Notably, in the cluster illustrated by Figure 10 (c), the trend index of velocity for all Action phases exceeds 0, indicating an increasing trend. Concurrently, the speed difference demonstrates a pronounced downward trend. Given the aforementioned domain knowledge, we label Action phases in this cluster as the “Catch up” pattern. Conversely, the cluster of Action phases depicted in Figure 10 (b) exhibits a completely reverse trend, that is, a negative velocity trend index and a positive speed difference trend index, representing a driving pattern named “Keep away”. In Figure 10 (f), the trend indexes of velocity and speed difference generally fluctuate around 0, denoting a “Maintain distance” pattern.

In the same way, the three aforementioned driving patterns can also be identified in the other three figures. More specifically, Figure 10 (e) illustrates an uptrend in velocity and a downtrend in speed difference, indicating a “Catch up” pattern. The “Keep away” and “Maintain distance” patterns can be discerned in Figures 10 (a) and 10 (d), respectively. It is worth noting that the patterns manifest with more instability in these three clusters, thus being labeled as an Unstable state. On the contrary, the driving patterns detected in Figures 10 (c), 10 (b), and 10 (f) reflect a Stable state. Ultimately, the driving patterns identified in the I80 dataset are interpreted as “Stable catch up”, “Stable keep away”, “Stable maintain distance”, “Unstable catch up”, “Unstable keep away”, and “Unstable maintain distance”, as summarized in Table 3. The shade of color corresponds to the size of each pattern. Generally, Unstable patterns considerably outweigh Stable ones in size. Among the Stable patterns, “Maintain distance” exceeds the other two dynamic patterns in size. This observation aligns with empirical knowledge: driving is a dynamic process and contains stochastic, maintaining the status is the simplest approach to perform a Stable state.

Figure 10: Trend index of variables in each cluster - I80 dataset.

Table 3: Overview of driving patterns - I80 dataset
Catch up Keep away Maintain distance
Stable (c) (b) (f)
Unstable (e) (a) (d)

5.3 Driving Patterns in US101 dataset↩︎

Corresponding driving patterns are also identified in the US101 dataset as illustrated in Figure 11. In Figure 11 (c), most of the Action phases display an increasing trend in velocities. Conversely, the overall speed difference exhibits a decreasing trend. Considering the limited outliers in both variables’ indexes, this pattern is interpreted as a “Stable catch up” pattern. Figure 11 (b) demonstrates an obvious downward trend in velocity and an upward trend in speed difference, coupled with numerous outliers. Hence, this pattern is recognized as “Unstable keep away”. Both Figure 11 (a) and Figure 11 (d) exhibit a “Maintain distance” pattern. As the outliers in Figure 11 (d) are significantly more than those in Figure 11 (a), they are labeled as “Unstable” and “Stable”, respectively. As the statistics presented in Table 4 reveal, the “Maintain distance” pattern has the largest size, where the Unstable pattern significantly outweighs the Stable one. The driving pattern with the smallest size in this dataset is “Stable catch up”. Notice that patterns of “Stable keep away” and “Unstable catch up” observed in the I80 dataset are absent here, making the “Maintain distance” pattern significantly surpasses others in frequency. This is consistent with driving behavior in the relatively heavy traffic observed during morning peak hours.

Figure 11: Trend index of variables in each cluster - US101 dataset.

Table 4: Overview of driving patterns - US101 dataset
Catch up Keep away Maintain distance
Stable (c) (a)
Unstable (b) (d)

6 Concluding Remarks↩︎

To capture representations of driving characteristics and facilitate a comprehensive understanding of driving behavior, this study proposed a framework to cluster Action phases and interpret those clusters as several driving patterns. This section presents a summary of the key findings and main conclusions derived from this study. Then discussion and outlook of the proposed framework are provided to shed light on its implications and future directions.

6.1 Findings of the Study↩︎

As clustering algorithms need input data arrays of equal length, a Resampling and Downsampling Method (RDM) was first adopted to standardize the various length of Action phases in this framework. Then, the clustering calibration process of “Feature Selection”, ‘Clustering Analysis’’, “Difference/Similarity Evaluation”, and “Action phases Re-extraction” was iterated until all differences among clusters and similarities within clusters meet the pre-determined criteria. Finally, six clusters were observed in I80 dataset, indicating six driving patterns, which have been labeled as “Catch up”, “Keep away”, and “Maintain distance”, each with “Stable” and “Unstable” states. These driving patterns were also identified in the US101 dataset, while the patterns “Stable keep away” and “Unstable catch up” were absent.

6.2 Conclusions↩︎

The main conclusions of this study are summarized below:

(i) Velocity \(v\) exhibits the highest importance score among the considered four variables, suggesting that it reflects more characteristics of driving behavior.

(ii) In general, Unstable patterns significantly outnumber Stable ones in terms of size. Among the Stable patterns, “Maintain distance” exceeds the other two dynamic patterns in size. This observation aligns with empirical knowledge: driving is a dynamic and stochastic process, and maintaining the status is the simplest way to achieve a Stable state.

(iii) Comparable driving patterns have been detected in both the I80 and US101 datasets. Notably, the patterns “Stable keep away” and “Unstable catch up” are missing from the US101 dataset. As previously identified by [14], the traffic flow in US101 dataset displays less heterogeneity compared to the I80 dataset. This consistency indicates the prospective advantages of using driving pattern to illustrate the heterogeneity in driving behavior.

6.3 Discussion and Outlook↩︎

This framework introduces an unsupervised learning method to improve the manual categorization of heterogeneous driving behaviors, thereby addressing the pitfalls of depending solely on experiential knowledge. By incorporating a wide array of driving behavior characteristics, such as stability and driving state, more accurate and justifiable labels result. This will help to alleviate the scarcity of labels in supervised learning and consequently bolster its performance in tasks such as driving behavior modeling and driving trajectory prediction, among others. However, the framework does have its limitations, which are the focus of our future research. First, the framework needs further validation to strengthen the credibility of the derived driving patterns. Additionally, the methods and techniques employed in each element of the framework could be further optimized and justified.

7 Acknowledgements↩︎

This work is supported by the Department of Transport &Planning at Delft University of Technology and Data Analysis & Traffic Simulation Lab (DiTTLab).

8 AUTHOR CONTRIBUTIONS↩︎

The authors confirm their contribution to the paper as follows: study conception and design: Xue Yao, Simeon C. Calvert, Serge P. Hoogendoorn; data collection: Xue Yao; analysis and interpretation of results: Xue Yao, Simeon C. Calvert; draft manuscript preparation: Xue Yao, Serge P. Hoogendoorn, Simeon C. Calvert. All authors reviewed the results and approved the final version of the manuscript.

References↩︎

[1]
Ossen, S., S. P. Hoogendoorn, and B. G. Gorte, Interdriver differences in car-following: A vehicle trajectory–based study. Transportation Research Record, Vol. 1965, No. 1, 2006, pp. 121–129.
[2]
Sun, Z., X. Yao, Z. Qin, P. Zhang, and Z. Yang, Modeling car-following heterogeneities by considering leader–follower compositions and driving style differences. Transportation research record, Vol. 2675, No. 11, 2021, pp. 851–864.
[3]
Kerner, B. S. and S. L. Klenov, Spatial–temporal patterns in heterogeneous traffic flow with a variety of driver behavioural characteristics and vehicle parameters. Journal of Physics A: Mathematical and General, Vol. 37, No. 37, 2004, p. 8753.
[4]
Tavakoli, A. and A. Heydarian, Multimodal driver state modeling through unsupervised learning. Accident Analysis & Prevention, Vol. 170, 2022, p. 106640.
[5]
Zou, Y., T. Zhu, Y. Xie, Y. Zhang, and Y. Zhang, Multivariate analysis of car-following behavior data using a coupled hidden Markov model. Transportation research part C: emerging technologies, Vol. 144, 2022, p. 103914.
[6]
Wang, W., J. Xi, A. Chong, and L. Li, Driving style classification using a semisupervised support vector machine. IEEE Transactions on Human-Machine Systems, Vol. 47, No. 5, 2017, pp. 650–660.
[7]
Figueira, A. C. and A. P. C. Larocca, Proposal of a driver profile classification in relation to risk level in overtaking maneuvers. Transportation research part F: traffic psychology and behaviour, Vol. 74, 2020, pp. 375–385.
[8]
Terada, R., H. Okuda, T. Suzuki, K. Isaji, and N. Tsuru, Multi-scale driving behavior modeling using hierarchical PWARX model. In 13th International IEEE Conference on Intelligent Transportation Systems, IEEE, 2010, pp. 1638–1644.
[9]
Bender, A., G. Agamennoni, J. R. Ward, S. Worrall, and E. M. Nebot, An unsupervised approach for inferring driver behavior from naturalistic driving data. IEEE transactions on intelligent transportation systems, Vol. 16, No. 6, 2015, pp. 3325–3336.
[10]
Liu, H., T. Taniguchi, T. Takano, Y. Tanaka, K. Takenaka, and T. Bando, Visualization of driving behavior using deep sparse autoencoder. In 2014 IEEE Intelligent Vehicles Symposium Proceedings, IEEE, 2014, pp. 1427–1434.
[11]
Ackerman, E., How drive. ai is mastering autonomous driving with deep learning. IEEE Spectrum Magazine, Vol. 1, 2017.
[12]
Higgs, B. and M. Abbas, Segmentation and clustering of car-following behavior: Recognition of driving patterns. IEEE Transactions on Intelligent Transportation Systems, Vol. 16, No. 1, 2014, pp. 81–90.
[13]
Wang, W., J. Xi, and D. Zhao, Driving style analysis using primitive driving patterns with Bayesian nonparametric approaches. IEEE Transactions on Intelligent Transportation Systems, Vol. 20, No. 8, 2018, pp. 2986–2998.
[14]
Yao, X., C. S. Calvert, and P. S. Hoogendoorn, Identification of Driving Heterogeneity using Action-chains. In IEEE 26th International Conference on Intelligent Transportation Systems (ITSC), 2023, Bilbao, Spain.
[15]
Knoop, V. L. and S. P. Hoogendoorn, Relation between longitudinal and lateral action points. In Traffic and Granular Flow’13, Springer, 2015, pp. 571–576.
[16]
Liu, Q., N. Nguyen, and X. Tang, Accurate algorithms for nonuniform fast forward and inverse Fourier transforms and their applications. In IGARSS ’98. Sensing and Managing the Environment. 1998 IEEE International Geoscience and Remote Sensing. Symposium Proceedings. (Cat. No.98CH36174), 1998, Vol. 1, pp. 288–290 vol.1.
[17]
Abdi, H. and L. J. Williams, Principal component analysis. Wiley interdisciplinary reviews: computational statistics, Vol. 2, No. 4, 2010, pp. 433–459.
[18]
Kilgour, D. M., J.-C. Grégoire, and A. M. Foley, Weighted scoring elections: is Borda best? Social Choice and Welfare, 2022, pp. 1–27.
[19]
Rokach, L. and O. Maimon, Clustering methods. Data mining and knowledge discovery handbook, 2005, pp. 321–352.
[20]
Fraley, C. and A. E. Raftery, How many clusters? Which clustering method? Answers via model-based cluster analysis. The computer journal, Vol. 41, No. 8, 1998, pp. 578–588.
[21]
Nguyen, T. T., P. Krishnakumari, S. C. Calvert, H. L. Vu, and H. Van Lint, Feature extraction and clustering analysis of highway congestion. Transportation Research Part C: Emerging Technologies, Vol. 100, 2019, pp. 238–258.
[22]
Stuart, R. and N. Peter, Artificial intelligence-a modern approach 3rd ed, 2016.
[23]
Salvador, S. and P. Chan, Toward accurate dynamic time warping in linear time and space. Intelligent Data Analysis, Vol. 11, No. 5, 2007, pp. 561–580.
[24]
Rendón, E., I. Abundez, A. Arizmendi, and E. M. Quiroz, Internal versus external cluster validation indexes. International Journal of computers and communications, Vol. 5, No. 1, 2011, pp. 27–34.
[25]
Chu, C.-S. J., Time series segmentation: A sliding window approach. Information Sciences, Vol. 85, No. 1-3, 1995, pp. 147–173.