Vision-based navigation can be categorized largely as relative and absolute navigation. The relative navigation based on a vision sensor has the disadvantage of accumulating errors over time caused by various factors. In contrast, Scene matching based absolute navigation has the advantage that the position of the vehicle can be independently calculated without needing external information, making it possible to calculate a stable navigation solution without cumulative errors. However the conventional image based navigation based on scene matching algorithm has a disadvantage also. Which is mismatching problem between large database and camera input image. Mismatching problem is a dangerous to navigation system. Because matching result of the image based navigation can be estimated far from the true position of the vehicle, like an outlier. The large position errors caused by unnecessary feature point clusters that have not been removed during outlier rejection. Extracted feature points would be created, disappeared, or tracked in next time step. The created or disappeared features are not our interesting points and can be filtered out by matching consecutive images. We propose the robust scene matching algorithm with only time invariant feature points via relative velocity model and uncertainty. Only trackable feature points are needed for a matching with database and improve the matching result. So that we use the relative velocity model. Which represents the feature's moving motion by opposite direction of the vehicle's velocity. Proposed relative velocity model consist of the acceleration of the vehicle. For a short time, the velocity from accelerometer is very accurate and precise. Because an accumulated error is relatively small. Also we propose a pixel boundary. Which expresses the uncertainty that the feature point is expected to be located. If the feature point extracted by next time-step is in the proposed boundary, the bounded feature point is used for matching with the database. The algorithm is summarized as follows. First, the process of determining the time-invariant feature points extracts and matches the feature points in a continuous camera input image. Through this process, feature points that disappear or generated can be removed. The second step is to propagate the position of the feature point used at the previous point to the position expected to be located at the current point through the proposed relative velocity model. The third step is to calculate the uncertainty of the propagated feature point, that is, feature point that are expected to be present. Finally, it is determined whether the feature point detected at the present exists within the uncertainty bound of the predicted feature point. If extract points are in uncertainty boundary, they are the feature points to be matched. After the proposed filtering process, feature points are very reliable and enough to mitigate mis-matching problem. Because terminal feature candidates are propagated by accurate velocity model and bounded its uncertainty. Proposed algorithm is verified by simulation using real flight experimental data. Flight vehicle is DJI-Company Mavic-pro model with embedded down looking gimbaled camera. The simulation result include extracted feature point and its boundary. And the navigation accuracy with its covariance. We combine advantages of both inertial navigation system and vision-based navigation to mitigate mismatching problem in scene-matching. Because inertial navigation system is very accurate in a short time and scene-matching has bounded solution. We defined the trackable and bounded features as time-invariant feature point. And we use relative velocity model to estimate the location of the feature point. Selected feature points are used for the scene-matching algorithm and are bounded by model uncertainty Finally, the proposed algorithm is verified by a simulation using real flight experimental data.