This paper adds analytical kinematic priors for uncertainty across timesteps in trajectory forecasting, boosting performance and stability in traffic tasks.This paper adds analytical kinematic priors for uncertainty across timesteps in trajectory forecasting, boosting performance and stability in traffic tasks.

From Guesswork to Ground Truth: Making Traffic Forecasts Physically Feasible

:::info Authors:

(1) Laura Zheng, Department of Computer Science, University of Maryland at College Park, MD, U.S.A ([email protected]);

(2) Sanghyun Son, Department of Computer Science, University of Maryland at College Park, MD, U.S.A ([email protected]);

(3) Jing Liang, Department of Computer Science, University of Maryland at College Park, MD, U.S.A ([email protected]);

(4) Xijun Wang, Department of Computer Science, University of Maryland at College Park, MD, U.S.A ([email protected]);

(5) Brian Clipp, Kitware ([email protected]);

(6) Ming C. Lin, Department of Computer Science, University of Maryland at College Park, MD, U.S.A ([email protected]).

:::

Abstract and I. Introduction

II. Related Work

III. Kinematics of Traffic Agents

IV. Methodology

V. Results

VI. Discussion and Conclusion, and References

VII. Appendix

\ Abstract— In trajectory forecasting tasks for traffic, future output trajectories can be computed by advancing the ego vehicle’s state with predicted actions according to a kinematics model. By unrolling predicted trajectories via time integration and models of kinematic dynamics, predicted trajectories should not only be kinematically feasible but also relate uncertainty from one timestep to the next. While current works in probabilistic prediction do incorporate kinematic priors for mean trajectory prediction, variance is often left as a learnable parameter, despite uncertainty in one time step being inextricably tied to uncertainty in the previous time step. In this paper, we show simple and differentiable analytical approximations describing the relationship between variance at one timestep and that at the next with the kinematic bicycle model. In our results, we find that encoding the relationship between variance across timesteps works especially well in unoptimal settings, such as with small or noisy datasets. We observe up to a 50% performance boost in partial dataset settings and up to an 8% performance boost in large-scale learning compared to previous kinematic prediction methods on SOTA trajectory forecasting architectures out-of-the-box, with no fine-tuning.

\

I. INTRODUCTION

Motion forecasting in traffic involves predicting the next several seconds of movement for select actors in a scene, given the context of historical trajectories and the local environment. There are several different approaches to motion forecasting in deep learning, but one thing is in common— scaling resources, dataset sizes, and model sizes will always help generalization capabilities. However, the difficult part of motion forecasting is not necessarily how the vehicles move, but rather, figuring out the why, as there is always a human (at least, for now) behind the steering wheel.

\ As model sizes become bigger and performance increases, so does the complexity of learning the basics of vehicle dynamics, where, most of the time, vehicles travel in relatively straight lines behind the vehicle leading directly in front. Since training a deep neural network is costly, it is beneficial for both resource consumption and model generalization to incorporate the use of existing dynamics models in the training process, combining existing knowledge with powerful modern architectures. With embedded priors describing the relationships between variables across time, perhaps a network’s modeling power can be better allocated to understanding the why’s of human behavior, rather than the how’s.

\ Fig. 1: Motivating example for probabilistic kinematic priors. In real-world traffic, vehicles have a constrained range of behaviors. For instance, it is not possible for a vehicle to move side to side directly, and driving in reverse is highly unlikely. Without kinematic priors, neural networks may search the space of all trajectories, possible or impossible. Without accounting for analytical variances in trajectories as in previous work, the range of possible future trajectories may also be unrealistic.

\ Kinematic models have already been widely used in various autonomous driving tasks, especially trajectory forecasting and simulation tasks [1], [2], [3], [4]. These models explicitly describe how changes in the input parameters influence the output of the dynamical system. Typically, kinematic input parameters are often provided by either the robot policy as an action, or by a human in direct interaction with the robot, e.g. steering, throttle, and brake for driving a vehicle. For tasks modeling decision-making, such as trajectory forecasting of traffic agents, modeling the input parameters may be more descriptive and interpretable than modeling the output directly. Moreover, kinematic models relate the input actions directly to the output observation; thus, any output of the kinematic model should, at the very least, be physically feasible in the real world [1].

\ While previous works already use kinematic models to provide feasible output trajectories, none consider compounding uncertainty into the time horizon analytically. Previous works apply kinematic models deterministically, either for deterministic trajectory prediction or to predict the average trajectory in probabilistic settings, leaving the variance as a learnable matrix.

\ Kalman filtering is a relevant and classical technique that uses kinematic models for trajectory prediction. Kalman filters excel in time-dependent settings where signals may be contaminated by noise, such as object tracking and motion prediction [5], [6]. Unlike classical Kalman filtering methods

\ Fig. 2: Qualitative example. We visualize an example of the mean trajectory of the highest-scored Gaussian from the results of Table II. Standard deviations are visualized as ellipses at each second into the future, and ground truth trajectories are drawn in green. Our method (blue) not only predicts a smoother, less jagged mean trajectory but also provides a more realistic spread of trajectories into the future. The baseline method (red) shows uneven speeds (ellipses are at uneven intervals) in addition to uncertainty extending far beyond the road boundaries.

\ for motion prediction, however, we focus on models of learnable actions and behaviors, rather than constant models of fixed acceleration and velocity. We believe that, with elements from classical methods such as Kalman filtering in addition to powerful modern deep learning architectures, motion prediction in traffic can have the best of both worlds where modeling capability meets stability.

\ In this paper, we go one step further from previous work by considering the relationship between distributions of kinematic parameters to distributions of trajectory rollouts, rather than single deterministic kinematic parameter predictions to the mean of trajectory rollouts [1], [4]. We hypothesize that modeling uncertainty explicitly across timesteps according to a kinematic model will result in better performance, realistic trajectories, and stable learning, especially in disadvantageous settings such as small or noisy datasets, which can be common settings, especially in fine-tuning. We run experiments on four different kinematic formulations with Motion Transformer (MTR) [7] and observe up to a 50% performance boost in partial-dataset settings and up to an 8% performance boost in full dataset settings on mean average precision (mAP) metrics for the Waymo Open Motion Dataset (WOMD) [8] compared to with kinematic priors enforced on mean trajectories only.

\ In summary, the main contributions of this work include:

\

  1. A simple and effective method for incorporating analytically-derived kinematic priors into probabilistic models for trajectory forecasting (Section IV), which boosts performance and generalization (Section V) and requires trivial additional overhead computation;

    \

  2. Results and analysis in different settings on four different kinematic formulations: velocity components vx and vy, acceleration components ax and ay, speed s and heading θ components, and steering δ and acceleration a (Section IV-B).

    \

  3. Analytical error bounds for the first and second-order kinematic formulations (section IV-C).

\ Fig. 3: Linear approximation of position distributions via kinematic priors.

\

II. RELATED WORKS

A. Trajectory Forecasting for Traffic

Traffic trajectory forecasting is a popular task where the goal is to predict the short-term future trajectory of multiple agents in a traffic scene. Being able to predict the future positions and intents of each vehicle provides context for other modules in autonomous driving, such as path planning. Large, robust benchmarks such as the Waymo Motion Dataset [8], [9], Argoverse [10], and the NuScenes Dataset [11] have provided a standardized setting for advancements in the task, with leaderboards showing clear rankings for state-of-the-art models. Amongst the top performing architectures, most are based on Transformers for feature extraction [7], [12], [13], [14], [15]. Current SOTA models also model trajectory prediction probabilistically, as inspired by the use of GMMs in MultiPath [16].

\ One common theme amongst relevant state-of-the-art, however, is that works employing kinematic models for time-integrated trajectory rollouts typically only consider kinematic variables deterministically, which neglect the relation between kinematic input uncertainty and trajectory rollout uncertainty [1], [4], [17]. In our work, we present a method for use of kinematic priors which can be complemented with any previous work in trajectory forecasting. Our contribution can be implemented in any of the SOTA methods above, since it is a simple reformulation of the task with no additional information needed.

\

B. Physics-based Priors for Learning

Model-based learning has shown to be effective in many applications, especially in robotics and graphics. There are generally two approaches to using models of the real world: 1) learning a model of dynamics via a separate neural network [18], [19], [20], [21], or 2) using existing models of the real world via differentiable simulation [22], [23], [24], [25], [26], [27], [28].

\ In our method, we pursue the latter. Since we are not modeling complex systems such as cloth or fluid, simulation of traffic agent states require only a simple, fast, and differentiable update. In addition, since the kinematic models do not describe interactions between agents, the complexity of the necessary model is greatly reduced. In this paper, we hypothesize that modeling the simple kinematics (e.g., how a vehicle moves forward) with equations will allow for greater modeling expressivity on behavior.

\

:::info This paper is available on arxiv under ATTRIBUTION-NONCOMMERCIAL-NODERIVS 4.0 INTERNATIONAL license.

:::

\

Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact [email protected] for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.

You May Also Like

The Channel Factories We’ve Been Waiting For

The Channel Factories We’ve Been Waiting For

The post The Channel Factories We’ve Been Waiting For appeared on BitcoinEthereumNews.com. Visions of future technology are often prescient about the broad strokes while flubbing the details. The tablets in “2001: A Space Odyssey” do indeed look like iPads, but you never see the astronauts paying for subscriptions or wasting hours on Candy Crush.  Channel factories are one vision that arose early in the history of the Lightning Network to address some challenges that Lightning has faced from the beginning. Despite having grown to become Bitcoin’s most successful layer-2 scaling solution, with instant and low-fee payments, Lightning’s scale is limited by its reliance on payment channels. Although Lightning shifts most transactions off-chain, each payment channel still requires an on-chain transaction to open and (usually) another to close. As adoption grows, pressure on the blockchain grows with it. The need for a more scalable approach to managing channels is clear. Channel factories were supposed to meet this need, but where are they? In 2025, subnetworks are emerging that revive the impetus of channel factories with some new details that vastly increase their potential. They are natively interoperable with Lightning and achieve greater scale by allowing a group of participants to open a shared multisig UTXO and create multiple bilateral channels, which reduces the number of on-chain transactions and improves capital efficiency. Achieving greater scale by reducing complexity, Ark and Spark perform the same function as traditional channel factories with new designs and additional capabilities based on shared UTXOs.  Channel Factories 101 Channel factories have been around since the inception of Lightning. A factory is a multiparty contract where multiple users (not just two, as in a Dryja-Poon channel) cooperatively lock funds in a single multisig UTXO. They can open, close and update channels off-chain without updating the blockchain for each operation. Only when participants leave or the factory dissolves is an on-chain transaction…
Share
BitcoinEthereumNews2025/09/18 00:09
Craft Ventures Opens Austin Office

Craft Ventures Opens Austin Office

AUSTIN, Texas–(BUSINESS WIRE)–Craft Ventures, the venture capital firm co-founded in 2017 by David Sacks and Bill Lee, has opened a new office in Austin, Texas,
Share
AI Journal2026/01/01 08:00
Paxos launches new startup to help institutions offer DeFi products

Paxos launches new startup to help institutions offer DeFi products

PANews reported on June 19 that according to The Block, the stablecoin issuer Paxos launched a new startup Paxos Labs, which aims to help institutions integrate DeFi and on-chain products
Share
PANews2025/06/19 00:04