![]() Localization method and apparatus, mobile terminal and computer-readable storage medium
专利摘要:
Disclosed are a positioning method and apparatus, and a mobile terminal. The method includes: determining (S102) an actual distance between a first position and a second position of a target object according to a preset model, and determining a heading angle corresponding to the actual distance according to a direction sensor, where the first position is a position corresponding to a current sampling instant, and the second position is a position at a previous sampling instant adjacent to the current sampling instant; and determining (SI04) the first position of the target object according to a predetermined reference position, the actual distance and the heading angle, and taking the first position as a predetermined reference position for calculating a third position, wherein the third position is a position at a next sampling instant adjacent to the current sampling instant. Further provided is a computer storage medium. 公开号:EP3680616A1 申请号:EP18852966.3 申请日:2018-08-02 公开日:2020-07-15 发明作者:Shijun Chen;Xinxin LIU;Xiaolong Xu;Dawei Chen;Yuanyuan Wang 申请人:ZTE Corp; IPC主号:G01C21-00
专利说明:
[0001] This application is based on and claims priority to a Chinese patent application No. 201710805524.1 filed on September 08, 2017 , disclosure of which is incorporated herein by reference in its entirety. TECHNICAL FIELD [0002] The present application relates to, but is not limited to, the positioning field, and in particular, relates to a positioning method and apparatus, a mobile terminal and a computer-readable storage medium. BACKGROUND [0003] With the rapid development of the Internet, position-based services have become an indispensable part in people's life. As one of the key technologies of position-based services, the positioning technology has gained more and more attention. The positioning technology as a whole can be divided into indoor positioning and outdoor positioning. Global positioning system (GPS) is generally used for the outdoor positioning, but once indoor, GPS signal attenuation rapidly attenuates due to the shielding of buildings, and accurate positioning services cannot be provided for human beings at all. [0004] However, in large indoor places such as shopping malls, underground garages and exhibitions, the need for accurate indoor positioning is becoming more and more urgent. For a general user, in a strange and complex indoor environment, accurate positioning can help the user to know the position of the user in real time, the user can plan the advancing route under the condition that the destination is known, and the situation that the user is lost is avoided. The more valuable situation of indoor positioning is also reflected in the disaster occurrence time, for rescuers of disaster rescue and victims of self-rescue, the accurate indoor positioning can better assist the rescuers and the victims in positioning and planning feasible rescue paths and escape paths in real time under the condition that sight lines of the rescuers and the victims are blocked, and larger damage caused by environmental disturbance is avoided. [0005] At present, the common indoor positioning technologies include wireless local area network (WLAN) positioning, Bluetooth positioning and the like, but for these positioning technologies, positioning and navigation can only be performed with a network. In actual use, it is possible that currently no network exists, and the positioning cannot be performed. [0006] Therefore, all related positioning technologies need a network and cannot be performed without the network. Further, for the related positioning technologies, external auxiliary equipment is used for performing positioning and the energy consumption of a mobile phone is too large, which is not practical for users. SUMMARY [0007] Embodiments of the present application provide a positioning method and apparatus and a mobile terminal. [0008] In one aspect, an embodiment of the present application provides a positioning method. The method includes: determining an actual distance between a first position and a second position of a target object according to a preset model, and determining a heading angle corresponding to the actual distance according to a direction sensor, where the first position is a position corresponding to a current sampling instant, and the second position is a position at a previous sampling instant adjacent to the current sampling instant; and determining the first position of the target object according to a predetermined reference position, the actual distance and the heading angle, and taking the first position as a predetermined reference position for calculating a third position, where the third position is a position at a next sampling instant adjacent to the current sampling instant. [0009] In another aspect, an embodiment of the present application further provides a positioning apparatus. The apparatus includes a first determination module and a positioning module. The first determination module is configured to determine an actual distance between a first position and a second position of a target object according to a preset model, and determine a heading angle corresponding to the actual distance according to a direction sensor, where the first position is a position corresponding to a current sampling instant, and the second position is a position at a previous sampling instant adjacent to the current sampling instant; the positioning module is configured to determine the first position of the target object according to a predetermined reference position, the actual distance and the heading angle, and take the first position as a predetermined reference position for calculating a third position, where the third position is a position at a next sampling instant adjacent to the current sampling instant. [0010] In another aspect, an embodiment of the present application further provides a mobile terminal including the positioning apparatus described above. [0011] In another aspect, an embodiment of the present application provides a computer-readable storage medium. The computer-readable storage medium stores computer-executable instructions. After being executed, the computer-executable instructions are capable of implementing the positioning method provided in any aspect described above. [0012] According to the embodiments of the present application, the actual distance between two positions when the target object moves is determined through the preset model, the heading angle of the first position at the current sampling instant is determined according to the direction sensor, the current position can be determined through calculation combined with the predetermined reference position, and during the continuous movement of the target object, each position in the moving process can be determined. The positioning method does not need the support of a network, can achieve accurate positioning, can at least solve the problem in the related art that positioning must depend on the network, and achieves indoor positioning and accurate positioning without network coverage or with a poor network coverage effect. BRIEF DESCRIPTION OF THE DRAWINGS [0013] FIG. 1 is a flowchart of a positioning method according to an embodiment of the present application; FIG. 2 is a structural diagram of a positioning apparatus according to an embodiment of the present application; and FIG. 3 is a flowchart of another positioning method according to an embodiment of the present application.DETAILED DESCRIPTION [0014] The present application will be further described below in detail with reference to the drawings and embodiments. It is to be understood that specific embodiments described herein are merely intended to explain the present application, and not to limit the present application. [0015] An embodiment of the present application provides a positioning method. The method includes steps S102 to S104 as shown in FIG. 1. [0016] In S102, an actual distance between a first position and a second position of a target object is determined according to a preset model, and a heading angle corresponding to the actual distance is determined according to a direction sensor, where the first position is a position corresponding to a current sampling instant, and the second position is a position at a previous sampling instant adjacent to the current sampling instant. [0017] In some embodiments, if the target object is a person, the preset model may be a pedestrian step size model determined for a person; and if the target object is a machine device such as a vehicle and a movable robot, the preset model may be a model determined for a machine. [0018] The preset model may be any pre-generated positioning model, for example, a positioning model based on a neural network, a learning model based on a learning machine such as a vector machine, etc. [0019] The preset model may be used for obtaining, according to historical data of the movement of the target object, a model capable of predicting the movement rule of the target object between the adjacent two sampling instants, so that the position of the target object in the next sampling instant can be predicted according to the preset model according to the second position in the previous sampling instant without a network. Therefore, movement positioning or trajectory tracking of the target object is achieved. The embodiment is illustrated by taking a person as an example, and in S102, the actual step size of the current step of a user is determined according to a preset pedestrian step size model, and the heading angle of the current step is determined according to the direction sensor. [0020] The preset pedestrian step size model in the embodiment is obtained by training the current user for multiple times. For example, the distance traveled by the user is specified, the user walks at different walking speeds for multiple times, the number of training walks is n, the walking time of the user is measured each time, the number of steps of the user is determined, the interval and variance of each step are recorded, the average step size is obtained by dividing the distance by the number of steps, the average value of frequencies and the average value of variances in all steps of the pedestrian in all measurements are respectively taken as the frequency and the variance of the measurement, so that n step size samples {L 1, L 2,..., Ln }, n frequency samples {f 1 , f 2,...,fn } and n variance samples {v 1, v 2,...., vn } are obtained, and different step size models of pedestrians can be obtained by performing linear fitting on the n samples, so that the step sizes of different pedestrians at different instants can be estimated in real time during positioning. [0021] A preset pedestrian step size model after training may be: Lk =b + cf(k) + dv + δ, where Lk is an actual step size of the user at a k-th instant, f(k) is a step frequency of the current step of the user, v is a variance of acceleration data generated by the target object (user) in the current step, b may be a constant, c and d are parameter values of corresponding variables respectively, b, c and d are obtained from training the user, and δ is model system noise. In some embodiments, c is the parameter value of corresponding, f(k) and d is the parameter value of corresponding v. [0022] In S104, the first position of the target object is determined according to a predetermined reference position, the actual distance and the heading angle, and the first position is taken as a predetermined reference position for calculating a third position, where the third position is a position at a next sampling instant adjacent to the current sampling instant. This embodiment is illustrated by a taking a person as an example, and the person may be replaced with other target objects. The predetermined reference position in this embodiment is the current position obtained, through calculation, from the previous step (the previous step may be the previous movement of the target object) of the current step (the current step may be the current movement of the target object). When the whole system is initialized, a user may manually position an initial position on a certain off-line map and take the initial position as a first predetermined reference position, and then after a current position has been calculated, the calculated current position may be taken as a predetermined reference position, and the current position of the next step is calculated according to the reference position. Here, one movement may be a movement in physical space, and may also be a movement within a preset duration. [0023] According to the embodiments of the present application, the actual distance between two positions when the target object moves is determined through the preset model, the heading angle of the first position at the current sampling instant is determined according to the direction sensor, the current position can be determined through calculation based on the predetermined reference position, and during the continuous movement of the target object, each position in the moving process can be determined. The positioning method does not need the support of a network, can achieve accurate positioning, and solves the problem in the existing art that positioning needs to be performed in an environment with the network and cannot be performed without network in the related positioning technologies. [0024] During the continuous movement of the target object, the target object will pass through many positions. Therefore, the predetermined reference position and the first position may also be connected to draw a motion route after the actual distance between the first position and the second position of the target object is determined according to the preset model and after the heading angle corresponding to the actual distance is determined according to the direction sensor. In some embodiments, if the target object is a machine device, the drawn motion route can be stored in the system of the machine device and can also be viewed by a user of the machine device. If the target object is a person, the drawing of the motion route can facilitate observation, by the user, of the route that the user walks through. After a motion route is drawn, the user can clearly observe the distance traveled by the user. In order to more clearly present the distance traveled, the initial position initially selected by the user may also be marked with a color different from the color of the movement route to facilitate user identification. [0025] Before the actual distance between the first position and the second position of the target object is determined according to the preset model and in order to prevent misjudgment, whether the target object actually moves needs to be determined according to a certain detection method, and in more cases, the process is more suitable for the situation that the target object is a person. The detection method may include the process described below. [0026] Accelerations in three direction axes of a three-dimensional coordinate system are obtained through a three-axis sensor, and an average acceleration at the current sampling instant is determined according to the accelerations in the three direction axes (an x-axis, a y-axis, and a z-axis which are mutually perpendicular). First acceleration data is obtained by passing average acceleration data of the current sampling instant through a first preset sliding time window, and second acceleration data is obtained by passing the average acceleration data of the current sampling instant through a second preset sliding time window, where processing time of the second preset sliding time window is shorter than processing time of the first preset sliding time window. If the second acceleration data is larger than the first acceleration data, it is detected whether the average acceleration data of the current sampling instant is the same as average acceleration data of the previous sampling instant. In a case where the average acceleration data of the current sampling instant is not the same as the average acceleration data of the previous sampling instant, it is determined that the target object actually displaced. Additionally/ alternatively, if the second acceleration data is smaller than or equal to the first acceleration data, the average acceleration data of the previous sampling instant is updated with the average acceleration data of the current sampling instant. [0027] In order to more accurately determine whether the target object has displaced, in the case where the average acceleration data of the current sampling instant is not the same as the average acceleration data of the previous sampling instant, it may also be determined whether a time interval between an acquisition time of the average acceleration data of the previous sampling instant and an acquisition time of the average acceleration data of the current sampling instant is within a preset time range. In a case where the time interval between the acquisition time of the average acceleration data of the previous sampling instant and the acquisition time of the average acceleration data of the current sampling instant is within the preset time range, it is detected whether a first difference value between the maximum acceleration among original acceleration data corresponding to the average acceleration data of the current sampling instant and the average acceleration data of the current sampling instant is larger than a first preset difference value, and it is detected whether a second difference value between the average acceleration data of the current sampling instant and the minimum acceleration among the original acceleration data corresponding to the average acceleration data of the current sampling instant is smaller than a second preset difference value. In a case where the first difference value is larger than the first preset difference value and the second difference value is smaller than the second preset difference value, it is determined that the target object actually has displaced. Through the process, whether the user walks one step can be determined more accurately. [0028] An embodiment of the present application further provides a positioning apparatus which may be arranged in a mobile terminal, and the structure of the apparatus is schematically shown in FIG. 2. The apparatus includes a first determination module 10 and a positioning module 11. [0029] The first determination module 10 is configured to determine an actual distance between a first position and a second position of a target object according to a preset model, and determine a heading angle corresponding to the actual distance according to a direction sensor, where the first position is a position corresponding to a current sampling instant, and the second position is a position at a previous sampling instant adjacent to the current sampling instant. The positioning module 11 is coupled to the first determination module 10 and configured to determine the first position of the target object according to a predetermined reference position, the actual distance and the heading angle, and take the first position as a predetermined reference position for calculating a third position, where the third position is a position at a next sampling instant adjacent to the current sampling instant. [0030] In some embodiments, if the target object is a person, the preset model may be a pedestrian step size model determined for a person; and if the target object is a machine device, the preset model may be a model determined for a machine. [0031] The embodiment is illustrated by taking a person as an example, and the first determination module described above determines the actual step size of the current step of a user according to a preset pedestrian step size model, and determines the heading angle of the current step according to the direction sensor. [0032] The preset pedestrian step size model in the embodiment is obtained by training the current user for multiple times. For example, the distance traveled by the user is specified, the user is required to walk for multiple times with different walking speeds, the number of training walks is n, the walking time of the user is measured each time, the number of steps of the user is determined, the interval and variance of each step are recorded, and the average step size is obtained by dividing the distance by the number of steps. The average value of frequencies and the average value of variances of all steps of the pedestrian in all measurements are respectively taken as the frequency and the variance of the measurement, so that n step size samples {L 1 ,L 2 ,...,Ln }, n frequency samples {f 1,f 2,...,fn } and n variance samples {v 1,v 2,...,vn } are obtained. Step size models of different pedestrians can be obtained by performing linear fitting on the n samples, so that the step sizes of different pedestrians at different instants can be estimated in real time during positioning. [0033] A preset pedestrian step size model after training may be: Lk = b + cf(k) + dv + δ, where Lk is an actual step size of the user at a k-th instant, f(k) is a step frequency of the current step of the user, v is a variance of acceleration data generated by the target object (user) in the current step, b may be a constant, c and d are parameter values of corresponding variables respectively, b, c and d are obtained from training the user, and δ is model system noise. In some embodiments, c is the parameter value of corresponding, f(k) and d is the parameter value of corresponding v. [0034] The step frequency here may include: the number of steps of a user per unit time; the step frequency of a pedestrian is one of moving times per unit time of a target object. [0035] The predetermined reference position in this embodiment is the current position obtained, through calculation, from the previous step of the current step. When the whole system is initialized, a user may manually position an initial position on a certain off-line map and take the initial position as a first predetermined reference position, and then after a current position has been calculated, the calculated current position may be taken as a predetermined reference position, and the current position of the next step is calculated according to the reference position. [0036] According to the embodiments of the present application, the actual distance between two positions when the target object moves is determined through the preset model, the heading angle of the first position at the current sampling instant is determined according to the direction sensor, the current position can be determined through calculation combined with the predetermined reference position, and during the continuous movement of the target object, each position in the moving process can be determined. The positioning method does not need a network, can achieve accurate positioning, and solves the problem in the existing art that positioning needs to be performed in an environment with the network and cannot be performed without network in the related positioning technologies. [0037] During the continuous movement of the target object, the target object will successively pass through many positions, and therefore, the predetermined reference position and the first position may also be connected to draw a motion route after the actual distance between the first position and the second position of the target object is determined according to the preset model and after the heading angle corresponding to the actual distance is determined according to the direction sensor. In some embodiments, if the target object is a machine device, the drawn motion route can be stored in the system of the machine device and can also be viewed by a user of the machine device. If the target object is a person, the drawing of the motion route can facilitate observation, by the user, of the route that the user walks through. After a motion route is drawn, the user can clearly observe the distance traveled by the user; in order to more clearly present the distance traveled, the initial position initially selected by the user may also be marked with a color different from the color of the movement route to facilitate user identification. [0038] In order to prevent misjudgment, whether the target object actually moves needs to be determined according to a certain detection method, and in more cases, the process is more suitable for the situation that the target object is a person. Therefore, the above apparatus may further include a second determination module, a detection module and an update module. The second determination module is configured to acquire accelerations in three direction axes of a three-dimensional coordinate system through a three-axis sensor, and determine an average acceleration of the current sampling instant according to the accelerations in the three direction axes; and obtain first acceleration data by passing average acceleration data of the current sampling instant through a first preset sliding time window, and obtain second acceleration data by passing the average acceleration data of the current sampling instant through a second preset sliding time window, where processing time of the second preset sliding time window is shorter than processing time of the first preset sliding time window. The detection module is coupled to the second determination module and configured to: in response to determining that the second acceleration data is larger than the first acceleration data, detect whether the average acceleration data of the current sampling instant is the same as average acceleration data of the previous sampling instant, and in a case where the average acceleration data of the current sampling instant is not the same as the average acceleration data of the previous sampling instant, determine that the target object actually has a position movement. The update module is couple to the second determination module and configured to: in response to determining that the second acceleration data is smaller than or equal to the first acceleration data, update the average acceleration data of the previous sampling instant with the average acceleration data of the current sampling instant. [0039] In order to more accurately determine whether the target object has displaced, the detection module described above may further be specifically configured to: in the case where the average acceleration data of the current sampling instant is not the same as the average acceleration data of the previous sampling instant, determine whether a time interval between an acquisition time of the average acceleration data of the previous sampling instant and an acquisition time of the average acceleration data of the current sampling instant is within a preset time range. In a case where the time interval between the acquisition time of the average acceleration data of the previous sampling instant and the acquisition time of the average acceleration data of the current sampling instant is within the preset time range, the detection module described above may further be specifically configured to detect whether a first difference value between the maximum acceleration among original acceleration data corresponding to the average acceleration data of the current sampling instant and the average acceleration data of the current sampling instant is larger than a first preset difference value, and detect whether a second difference value between the average acceleration data of the current sampling instant and the minimum acceleration among the original acceleration data corresponding to the average acceleration data of the current sampling instant is smaller than a second preset difference value. In a case where the first difference value is larger than the first preset difference value and the second difference value is smaller than the second preset difference value, the detection module described above may further be specifically configured to determine that the target object actually has the position movement. Through the process, whether the user walks one step can be determined more accurately. [0040] In some embodiments, the various modules described above may be implemented by a processor in a mobile terminal. [0041] A third embodiment of the present application provides a real-time off-line indoor map positioning method. The method performs real-time off-line accurate positioning and moving trajectory drawing for a pedestrian (a person is taken as an example in the embodiment) indoors by processing data acquired by a sensor built in a smart phone, thereby achieving low-cost, low-power consumption and high-precision indoor positioning. [0042] The embodiment adopts the technical solution including the steps described below.(1) In a step of identifying a pedestrian gait, acceleration data of a pedestrian during walking is acquired according to a three-axis accelerometer built in the smart phone and is smoothed, each step of the pedestrian is accurately identified through a double sliding time window average algorithm, and the occurrence time of each step of the pedestrian is recorded. Here, every step of the pedestrian may be understood as a movement of the target object, and/or a movement within a preset duration. (2) In a step of estimating a pedestrian step size, the step size model of the current pedestrian is trained in an off-line stage according to different acceleration characteristic data generated by different pedestrians, so that the step size of the current pedestrian at each instant is estimated in real time according to the step size model corresponding to the current pedestrian. (3) In a step of determining a pedestrian heading angle, the walking direction of each step of the pedestrian is obtained according to a direction sensor (i.e., an electronic compass) of the smart phone. (4) In a step of positioning the pedestrian in real-time and drawing the moving trajectory, the position coordinate of the pedestrian at each instant is calculated according to the position coordinate, the step size and the heading angle data of each step of the pedestrian at the initial instant, that is, the real-time positioning of the pedestrian is completed, and thereby the position coordinate of the pedestrian at each instant from the starting time to the ending time can be connected to obtain the moving trajectory of the pedestrian. [0043] Optionally, the pedestrian gait identification step may include the following process. [0044] In step A, a mobile phone three-dimensional coordinate system is established for a built-in three-axis sensor of the mobile phone, where the measured sensor data is relative to the three-dimensional coordinate system. [0045] In step B, the collected three-axis acceleration data in the smart phone is processed, and the overall acceleration is calculated, that is, a =a x 2 + a y 2 + a z 2, [0046] In step C, in a walking process of the pedestrian, each walking step corresponds to an instant, and sliding time window average filtering is performed on acceleration data generated by the pedestrian during walking so as to remove some noise. [0047] In step D, according to acceleration data generated from walking of the pedestrian, the double sliding time window average algorithm is used to determine whether the pedestrian has taken a possible step. [0048] In step E, on the basis that the pedestrian has taken a possible step in step D, it is determined whether the pedestrian really takes one step according to the walking period of the pedestrian and whether the maximum acceleration value and the minimum acceleration value in one step meet a preset threshold value range, and if the pedestrian is detected to have taken one step, the occurrence time of each step of the pedestrian is recorded, thereby finishing pedestrian gait identification. [0049] Optionally, the step of estimating the pedestrian step size may include the following process. [0050] In step F, the pedestrian step size model Lk = b + cf(k) + dv + δ is established, where Lk is a real-time step size of the pedestrian at the k-th instant, f(k) is the step frequency of this step of the pedestrian, v is a variance of acceleration data generated in this step by the pedestrian, b is a constant, c and d are parameter values of corresponding variables, b, c and d can be obtained from training different pedestrians, and δ is model system noise and may generally be set as δ = 0. [0051] In step G, after a pedestrian step size model is established, off-line training needs to be performed for a pedestrian to obtain parameter values in the model. When off-line training is performed to collect data, the distance traveled by the user is specified, the user is required to walk at different walking speeds, the number of training walks is n, the walking time and the number of steps of the user are measured each time, the interval and variance of each step are recorded, the average step size is obtained by dividing the distance by the number of steps, the average value of frequencies and the average value of variances in all steps of the pedestrian in all measurements are taken as the frequency and the variance of the measurement, so that n step size samples {L 1,L 2,..., Ln }, n frequency samples {f 1,f 2,...,fn } and n variance samples {v 1,v 2,...,vn } are obtained, and the step size models of different pedestrians can be obtained by performing linear fitting on the n samples, so that the step sizes of different pedestrians at different instants can be estimated in real time during positioning. For different training subjects, the step size of any training subject can be estimated with the step size model described above. [0052] Optionally, the step of determining the pedestrian heading angle may include the following process. [0053] In step H, the included angle between the horizontal projection in the y-axial direction of the mobile phone on the horizontal plane in the geographic coordinate system and the magnetic north direction is obtained through a direction sensor, i.e., an electronic compass, of the smart phone, so that the direction angle of the pedestrian relative to the magnetic north is obtained. [0054] In step I, the local magnetic declination of the pedestrian is obtained through table lookup. [0055] In step J, at the ending time when it is determined through pedestrian gait identification that the pedestrian really walks one step, a real-time heading angle of the pedestrian is obtained according to calculation of θk = βk - ε, where θk is the heading angle of the step of the pedestrian, βk is the direction angle of the pedestrian relative to the magnetic north measured by the direction sensor of the mobile phone at the ending time of the step, and ε is a local magnetic declination. [0056] Optionally, in the pedestrian real-time positioning and moving trajectory drawing step, the following process may be included. [0057] In step K, the indoor plan view of the place where the pedestrian is located is displayed on the mobile phone in a certain proportion, and coordinates (x 0, y 0) of the initial position of the pedestrian are obtained. [0058] In step L, the position (xk, yk ) of the pedestrian at any instant k is calculated as x 0 + ∑ i = 1 kL i sin θ i , y 0 + ∑ i = 1 kL i cos θ i [0059] In step M, the positions of the pedestrian at all instants are connected and drew on the terminal map in proportion, thus completing the drawing of the moving trajectory of the pedestrian. [0060] In some embodiments, in step C, the sliding time window average filtering is performed on acceleration data of the pedestrian in the following manner. [0061] In step C1, for the acceleration data collected in the walking process of the pedestrian, each acceleration collected by the accelerometer corresponds to an instant, and the acceleration value of the instant i is defined as ai = (a i-1 , + ai )/2, where ai is the acceleration value of the instant i. [0062] In step C2, the sliding time average filtering is performed, according to the formulaa i = ∑ a j ∗ l j ∑ l j, [0063] In some embodiments, in step D, the double sliding time window average algorithm is used to determine whether the pedestrian has taken a possible step in the following manner. [0064] In step D1, a long time window and a short time window are separately used for performing sliding average on the original acceleration data, i.e., the preset duration of one sliding time window is a little longer, the preset duration of the other sliding time window is a little shorter, and the sizes (e.g., durations) of the two windows are determined through training, so that the real peak of the original acceleration data is located at the intersection point of the acceleration data subjected to average of the two sliding windows. [0065] In step D2, after the acceleration data is averaged according to the preset two sliding time windows, two monitoring quantities oldSwapState and newSwapState are set for separately representing a state of the magnitude comparison of the acceleration value a 1 obtained by averaging two adjacent detected acceleration data through the long sliding time window with the acceleration value a s obtained by averaging two adjacent detected acceleration data through the short sliding time window. If a s > al at the current time point when the acceleration data is detected, newSwapState is set to true, otherwise, newSwapState is set to false. oldSwapState is a state of the magnitude comparison of al and a s at the previous time point when acceleration data was detected. If values of oldSwapState and newSwapState are different and newSwapState is true, it indicates that a possible step is detected, and the process goes to step E; otherwise, the value of newSwapState is assigned to oldSwapState to update oldSwapState, and al and a s at the next instant continue to be compared in magnitude. [0066] In some embodiments, in step E, it is determined whether the pedestrian actually takes one step in the following manner: on the basis that it is detected in step D2 that a pedestrian may take one step, if the following three conditions are satisfied, it indicates that the pedestrian actually takes one step and the time at the instant is recorded; otherwise, the pedestrian is not detected to take one step. The three conditions are as follows: (1) The time interval between the previous detection of a possible step and the current detection of a possible step is 0.2s to 2s; (2) In the process of detecting a possible step, the difference between the maximum acceleration value among the original acceleration data and the current acceleration value averaged over the long sliding time window is larger than 0.2; (3) The difference between the current acceleration value averaged over the long sliding time window and the minimum acceleration value among the original acceleration data is greater than 0.7. [0067] In some embodiments, in step F, the step frequency f(k) for each step of the pedestrian is calculated by the formula f k = 1 t k − t k − 1 , [0068] In some embodiments, in step F, the variance v of the acceleration data generated by the pedestrian in each step may be calculated by the formula v = E(ak 2) - (E(ak ))2, where ak is all acceleration data obtained in a step of the pedestrian, E(ak ) is the expectation of the acceleration data generated in the step of the pedestrian, and E(ak 2 ) is the expectation of the square of the acceleration data generated in the step of the pedestrian. [0069] An embodiment of the present application provides a real-time off-line indoor positioning mechanism and system. The mechanism and system adopts a double sliding time window average algorithm to perform pedestrian gait identification, which has high accuracy and improves the stability of pedestrian gait identification. [0070] An embodiment of the present application provides a real-time off-line indoor positioning mechanism and system. The mechanism and system do not depend on a satellite, a network or an external device and merely depend on an acceleration sensor and a direction sensor which are built in a widely popularized smart phone to implement off-line real-time positioning, with low cost, high positioning accuracy, good practicability and strong popularization. [0071] An embodiment of the present application provides a real-time off-line indoor positioning mechanism and system, which can solve the problem that accurate positioning of a pedestrian is difficult in the indoor environment without network at present. [0072] An embodiment of the present application provides a real-time offline indoor positioning mechanism and system which can be fused with other positioning solutions to implement positioning with higher accuracy. [0073] An embodiment of the present application further provides a positioning method, and the problem of the existing art to be solved by the method is that a pedestrian cannot be positioned off-line in real time in an indoor environment without network. According to the real-time off-line indoor positioning method provided by the embodiments of the present application, stable and accurate pedestrian gait identification is achieved through a double sliding time window average algorithm, accurate and real-time step size estimation of a pedestrian is completed through a trained pedestrian step size model, and a real-time heading angle of the pedestrian is determined through a direction sensor of a smart phone. In this way, under the condition that the initial position of the pedestrian is known, real-time off-line positioning and moving trajectory drawing of the pedestrian indoors are achieved, and the existing problem that accurate positioning of the pedestrian is difficult to achieve in an indoor environment without network can be solved, meanwhile, the mechanism does not depend on a satellite, a network or an external device and merely depends on an acceleration sensor and a direction sensor which are built in a widely popularized smart phone to implement off-line real-time positioning, with low cost, high positioning accuracy, good practicability and strong popularization. [0074] FIG. 3 shows a flowchart of a real-time off-line indoor positioning method when the present embodiment is applied to a smart terminal for indoor positioning. Steps of the present embodiment will be described below in detail with reference to FIG. 3. [0075] In step S1, an indoor map of a place where a pedestrian is located is displayed on a smart terminal in a certain proportion, and a position of the pedestrian at a previous instant is acquired. If the positioning is just started, coordinates of the initial position of the pedestrian are acquired through manual setting of the pedestrian or other positioning methods. [0076] Before the step S1, the pedestrian step size model Lk = b + cf(k) + dv + δ needs to be established, where Lk is a real-time step size of the pedestrian at the k-th instant, f(k) is the step frequency of this step of the pedestrian, v is a variance of acceleration data generated in this step by the pedestrian, b is a constant, c and d are parameter values of corresponding variables, b, c and d can be obtained from training different pedestrians, and δ is model system noise and generally be set as δ = 0. [0077] After a pedestrian step size model is established, off-line training is performed on a pedestrian to obtain parameter values in the model. When off-line training is performed to collect data, the distance traveled by the user is specified, the user is required to walk at different walking speeds, the number of training walks is n, the walking time and the number of steps of the user is measured each time, the interval and variance of each step are recorded, distance divided by the number of steps is the average step size, the average value of frequencies and the average value of variances in all steps of the pedestrian in all measurements are respectively taken as the frequency and the variance of the measurement, so that n step size samples {L 1,L 2,..., Ln }, n frequency samples {f 1,f 2,...,fn } and n variance samples {v 1,v 2,...,vn } are obtained. The step size models of different pedestrians can be obtained by performing linear fitting on the n samples, so that the step sizes of different pedestrians at different instants can be estimated in real time during positioning. The step frequency f(k) for each step of the pedestrian is calculated by the formula f k = 1 t k − t k − 1 , [0078] In step S2, the overall acceleration is calculated, for example, a mobile phone three-dimensional coordinate system is established for a built-in three-axis sensor of the mobile phone, where the measured sensor data is relative to the three-dimensional coordinate system; the collected three-axis acceleration data in the smart phone is processed, and the overall acceleration is calculated, that is, a =a x 2 + a y 2 + a z 2, [0079] In step S3, in a walking process of the pedestrian, each walking step corresponds to an instant, and sliding time window average filtering is performed on acceleration data generated by the pedestrian during walking so as to remove some noise. The step S3 may include step S31 and step S32. [0080] In step S31, for the acceleration data collected in the walking process of the pedestrian, the acceleration each time collected by the accelerometer corresponds to an instant, and the acceleration value of the instant i is defined as ai = (a i-1 + ai )/2, where ai is the acceleration value of the instant i. [0081] In step S32, the sliding time average filtering is performed, according to the formulaa i = ∑ a j ∗ l j ∑ l j, [0082] In step S4, according to acceleration data generated from walking of the pedestrian, the double sliding time window average algorithm is used to determine whether the pedestrian has taken a possible step. The step S4 may include step S41 and step S42. [0083] In step S41, a long time window and a short time window are separately used for performing sliding average on the original acceleration data, i.e., the preset duration of one sliding time window is a little longer, the preset duration of the other sliding time window is a little shorter, and the sizes (e.g., durations) of the two windows are determined through training, so that the real peak of the original acceleration data is located at the intersection point of the acceleration data subjected to average of the two sliding windows. Through repeated experimental comparison, the long window size may be set as Is, the short window size may be set as 0.2s, and the above requirements can be satisfied. [0084] In step S42, after the acceleration data is averaged according to the preset two sliding time windows, two monitoring quantities oldSwapState and newSwapState are set for separately representing a state of the magnitude comparison of the acceleration value al obtained by averaging two adjacent detected acceleration data through the long sliding time window with the acceleration value a s obtained by averaging two adjacent detected acceleration data through the short sliding time window. If as > al at the current time point when the acceleration data is detected, newSwapState is set to true, otherwise, newSwapState is set to false. oldSwapState is a state of the magnitude comparison of al with a s at the previous time point when acceleration data was detected. If values of oldSwapState and newSwapState are different and newSwapState is true, it indicates that a possible step is detected, and the process goes to step E; otherwise, the value of newSwapState is assigned to oldSwapState to update oldSwapState, and al and a s at the next instant continue to be compared in magnitude. [0085] In step S5, on the basis of determining that the pedestrian has taken a possible step in step S4, it is determined whether the pedestrian actually takes one step according to the walking period of the pedestrian and whether the maximum acceleration value and the minimum acceleration value in one step meet a preset threshold value range. If the pedestrian is detected to have taken one step, step 6 is performed. The step S5 specifically includes steps S51, S52, S53 and S54. [0086] In step S51, it is detected whether the time interval between the previous detection of a possible step and the current detection of a possible step is greater than 0.2s. If the time interval between the previous detection of a possible step and the current detection of a possible step is greater than 0.2s, the process goes to step S52; if the time interval between the previous detection of a possible step and the current detection of a possible step is not greater than 0.2s, it indicates that the pedestrian has not taken one step, and the process goes to the wait for a start of step 5 again. [0087] In step S52, it is detected whether the time interval between the previous detection of a possible step and the current detection of a possible step is smaller than 0.2s. If the time interval between the previous detection of a possible step and the current detection of a possible step is smaller than 0.2s, the process goes to step S53; if the time interval between the previous detection of a possible step and the current detection of a possible step is not smaller than 0.2s, it indicates that the pedestrian has not taken one step, and the process goes to the wait for a start of step 5 again. [0088] In step S53, it is detected whether the difference between the maximum acceleration value among the original acceleration data in the possible step and the current acceleration value subjected to average through the long sliding time window is greater than 0.2. If the difference is greater than 0.2, the process goes to step S54; if the difference is not greater than 0.2, it indicates that the pedestrian has not taken one step, and the process goes to the wait for a start of step 5 again. [0089] In step S54, it is detected whether the difference between the current acceleration value subjected to average through the long sliding time window and the minimum acceleration value among the original acceleration data in the possible step is greater than 0.7. If the difference is greater than 0.7, it indicates that the pedestrian actually takes one step, and the time of the one step is recorded; if the difference is not greater than 0.7, it indicates that the pedestrian has not taken one step, and the process goes to the wait for a start of step 5 again. [0090] In step S6, on the basis of identifying one step of the pedestrian in step S5, the step size of each step of the pedestrian is estimated according to the off-line trained pedestrian step size model. [0091] In step S7, a real-time deviation angle is obtained through calculation, and the step may include the following steps: at the ending time when it is determined through pedestrian gait identification that the pedestrian really walks one step, the included angle between the horizontal projection in the y-axial direction of the mobile phone on the horizontal plane in the geographic coordinate system and the magnetic north direction is obtained through a direction sensor, i.e., an electronic compass, of the smart phone, so that the direction angle of the pedestrian relative to the magnetic north is obtained; the local magnetic declination of the pedestrian is obtained through table lookup in advance; a real-time heading angle of the pedestrian is obtained according to calculation of θk = βk - ε, where θk is the heading angle of the step of the pedestrian, βk is the direction angle of the pedestrian relative to the magnetic north measured by the direction sensor of the mobile phone at the ending time of the step, and ε is a local magnetic declination. [0092] In step S8, real-time positioning is performed, and the position (xk, yk ) of the pedestrian at any instant k is calculated as x 0 + ∑ i = 1 kL i sin θ i , y 0 + ∑ i = 1 kL i cos θ i [0093] In step S9, the positions of the pedestrian at all instants are connected and drew on the smart terminal map in proportion, thus completing the drawing of the moving trajectory of the pedestrian. [0094] In step S10, whether the positioning is ended is determined. If the positioning is not ended, the process returns to step S1; if the positioning is ended, step S11 is performed. [0095] In S11, the positioning is ended. [0096] According to the embodiments of the present application, a double sliding time window average algorithm is adopted for pedestrian gait identification, so that the accuracy is high, and the stability of pedestrian gait identification is improved; meanwhile, the present solution does not depend on a satellite, a network and an external device and merely depends on an acceleration sensor and a direction sensor built in a widely popularized smart mobile phone to achieve off-line real-time positioning. Accurate positioning service for the pedestrian in an indoor environment without network can be provided only with one smart mobile phone, which is low in cost, high in positioning accuracy, good in practicability and strong in popularization, and can be fused with other positioning solutions to achieve positioning with higher accuracy. [0097] The present embodiment provides a mobile terminal which may be any mobile electronic device such as a portable device convenient for a user to carry or an electronic device such as a vehicle-mounted device. [0098] The mobile terminal includes a memory and a processor. [0099] The memory is configured to store information. [0100] The processor is connected to the memory and configured to execute executable codes in the memory; the executable codes, when executed, are capable of implementing any of the positioning methods provided by the embodiments of the present application, e.g., the methods shown in FIG. 1 and/or FIG. 3. [0101] The processor may be connected to the memory via a structure such as a bus within the mobile terminal. The processor may be a central processing unit, a microprocessor, a digital signal processor, a programmable array, an application specific integrated circuit, or the like. [0102] An embodiment of the present application further provides a computer storage medium for storing computer-executable instructions, where after being executed, the computer-executable instructions are capable of implementing any of the positioning methods provided by the foregoing embodiments, e.g., the methods shown in FIG. 1 and/or FIG. 3. [0103] The computer storage medium may be any type of storage medium such as an optical disk, a hard disk, a flash disk, a CD, a USB flash disk, or the like, optionally a non-transitory storage medium. [0104] From the description of the above-mentioned embodiments, it will be apparent to those skilled in the art that the methods in the embodiments described above may be implemented by software plus a necessary general-purpose hardware platform, or may of course be implemented by hardware. However, in many cases, the former is a preferred implementation manner. Based on this understanding, the solutions provided by the present application substantially, or the part contributing to the existing art, may be embodied in the form of a software product. The computer software product is stored on a storage medium (such as a read-only memory (ROM)/random access memory (RAM), a magnetic disk or an optical disk) and includes several instructions for enabling a terminal device (which may be a mobile phone, a computer, a server, a network device or the like) to execute the methods according to the embodiments of the present application. [0105] Although the preferred embodiments of the present application have been disclosed for illustrative purposes, those skilled in the art will appreciate that various modifications, additions and substitutions are possible, and accordingly, the scope of the present application should not be limited to the embodiments set forth above.
权利要求:
Claims (12) [0001] A positioning method, comprising: determining an actual distance between a first position and a second position of a target object according to a preset model, and determining a heading angle corresponding to the actual distance according to a direction sensor, wherein the first position is a position corresponding to a current sampling instant, and the second position is a position at a previous sampling instant adjacent to the current sampling instant; and determining the first position of the target object according to a predetermined reference position, the actual distance and the heading angle, and taking the first position as a predetermined reference position for calculating a third position, wherein the third position is a position at a next sampling instant adjacent to the current sampling instant. [0002] The positioning method of claim 1, wherein after determining the actual distance between the first position and the second position of the target object according to the preset model, and determining the heading angle corresponding to the actual distance according to the direction sensor, the method further comprises:connecting the predetermined reference position and the first position to draw a motion route. [0003] The positioning method of claim 1, wherein before determining the actual distance between the first position and the second position of the target object according to the preset model, the method further comprises: acquiring accelerations in three direction axes of a three-dimensional coordinate system through a three-axis sensor, and determining an average acceleration of the current sampling instant according to the accelerations in the three direction axes; obtaining first acceleration data by passing average acceleration data of the current sampling instant through a first preset sliding time window, and obtaining second acceleration data by passing the average acceleration data of the current sampling instant through a second preset sliding time window, wherein processing time of the second preset sliding time window is shorter than processing time of the first preset sliding time window; and in response to the second acceleration data being larger than the first acceleration data, detecting whether the average acceleration data of the current sampling instant is same as average acceleration data of the previous sampling instant, and in a case where the average acceleration data of the current sampling instant is not the same as the average acceleration data of the previous sampling instant, determining that the target object actually displaced; and/or in response to the second acceleration data being smaller than or equal to the first acceleration data, updating the average acceleration data of the previous sampling instant with the average acceleration data of the current sampling instant. [0004] The positioning method of claim 3, wherein in the case where the average acceleration data of the current sampling instant is not the same as the average acceleration data of the previous sampling instant, determining that the target object actually displaced comprises: in the case where the average acceleration data of the current sampling instant is not the same as the average acceleration data of the previous sampling instant, detecting whether a time interval between an acquisition time of the average acceleration data of the previous sampling instant and an acquisition time of the average acceleration data of the current sampling instant is within a preset time range; and in a case where the time interval between the acquisition time of the average acceleration data of the previous sampling instant and the acquisition time of the average acceleration data of the current sampling instant is within the preset time range, detecting whether a first difference value between a maximum acceleration among original acceleration data corresponding to the average acceleration data of the current sampling instant and the average acceleration data of the current sampling instant is larger than a first preset difference value, and detecting whether a second difference value between the average acceleration data of the current sampling instant and a minimum acceleration among the original acceleration data corresponding to the average acceleration data of the current sampling instant is smaller than a second preset difference value; and in a case where the first difference value is larger than the first preset difference value and the second difference value is smaller than the second preset difference value, determining that the target object actually displaced. [0005] The positioning method of any one of claims 1 to 4, wherein a preset pedestrian step size model is:Lk = b + cf(k) + dv + δ, wherein Lk is an actual distance of the target object at a k-th t, f(k) is a moving frequency of the target object at the current sampling instant, v is a variance of acceleration data generated by the target object at the current sampling instant, b is a constant, c and d are parameter values of corresponding variables respectively, b, c and d are obtained from training the target object, and δ is model system noise. [0006] A positioning apparatus, comprising: a first determination module, which is configured to determine an actual distance between a first position and a second position of a target object according to a preset model, and determine a heading angle corresponding to the actual distance according to a direction sensor, wherein the first position is a position corresponding to a current sampling instant, and the second position is a position at a previous sampling instant adjacent to the current sampling instant; and a positioning module, which is configured to determine the first position of the target object according to a predetermined reference position, the actual distance and the heading angle, and take the first position as a predetermined reference position for calculating a third position, wherein the third position is a position at a next sampling instant adjacent to the current sampling instant. [0007] The positioning apparatus of claim 6, further comprising: a drawing module, which is configured to connect the predetermined reference position and the first position to draw a motion route. [0008] The positioning apparatus of claim 6, further comprising: a second determination module, which is configured to acquire accelerations in three direction axes of a three-dimensional coordinate system through a three-axis sensor, and determine an average acceleration of the current sampling instant according to the accelerations in the three direction axes; and obtain first acceleration data by passing average acceleration data of the current sampling instant through a first preset sliding time window, and obtain second acceleration data by passing the average acceleration data of the current sampling instant through a second preset sliding time window, wherein processing time of the second preset sliding time window is shorter than processing time of the first preset sliding time window; a detection module, which is configured to: in response to the second acceleration data being larger than the first acceleration data, detect whether the average acceleration data of the current sampling instant is same as average acceleration data of the previous sampling instant, and in a case where the average acceleration data of the current sampling instant is not the same as the average acceleration data of the previous sampling instant, determine that the target object actually displaced; and an update module, which is configured to: in response to the second acceleration data being smaller than or equal to the first acceleration data, update the average acceleration data of the previous sampling instant with the average acceleration data of the current sampling instant. [0009] The positioning apparatus of claim 8, wherein the detection module is configured to: in the case where the average acceleration data of the current sampling instant is not the same as the average acceleration data of the previous sampling instant, detect whether a time interval between an acquisition time of the average acceleration data of the previous sampling instant and an acquisition time of the average acceleration data of the current sampling instant is within a preset time range; in a case where the time interval between the acquisition time of the average acceleration data of the previous sampling instant and the acquisition time of the average acceleration data of the current sampling instant is within the preset time range, detect whether a first difference value between a maximum acceleration among original acceleration data corresponding to the average acceleration data of the current sampling instant and the average acceleration data of the current sampling instant is larger than a first preset difference value, and detect whether a second difference value between the average acceleration data of the current sampling instant and a minimum acceleration among the original acceleration data corresponding to the average acceleration data of the current sampling instant is smaller than a second preset difference value; and in a case where the first difference value is larger than the first preset difference value and the second difference value is smaller than the second preset difference value, determine that the target object actually displaced. [0010] The positioning apparatus of any one of claims 6 to 9, wherein a preset pedestrian step size model used by the first determination module is:Lk = b + cf(k) + dv + δ, wherein Lk is an actual distance of the target object at a k-th instant, f(k) is a moving frequency of the target object at the current sampling instant, v is a variance of acceleration data generated by the target object at the current sampling instant, b is a constant, c and d are parameter values of corresponding variables respectively, b, c and d are obtained from training the target object, and δ is model system noise. [0011] A mobile terminal, comprising the positioning apparatus of any one of claims 6 to 10. [0012] A computer storage medium storing computer-executable instructions, wherein upon being executed, the computer-executable instructions are capable of implementing the method of any one of claims 1 to 5.
类似技术:
公开号 | 公开日 | 专利标题 US9594150B2|2017-03-14|Determining device locations using movement, signal strength Kang et al.2014|SmartPDR: Smartphone-based pedestrian dead reckoning for indoor localization CN105547305B|2018-03-16|A kind of pose calculation method based on wireless location and laser map match CN106605123B|2021-01-12|System for determining an entrance position and a region of interest US8751151B2|2014-06-10|System and method for localizing a trackee at a location and mapping the location using inertial sensor information CN105702151B|2019-06-11|A kind of indoor map construction method and device US8965688B2|2015-02-24|System and method for locating, tracking, and/or monitoring the status of personnel and/or assets both indoors and outdoors CN104180805B|2017-02-01|Smart phone-based indoor pedestrian positioning and tracking method US9451496B2|2016-09-20|Method and apparatus of positioning mobile terminal based on geomagnetism EP3385672A1|2018-10-10|Scalable indoor navigation and positioning systems and methods EP3052957B1|2018-03-21|Enhanced indoor localization method Lan et al.2014|Using smart-phones and floor plans for indoor location tracking-Withdrawn CN107145578B|2020-04-10|Map construction method, device, equipment and system US9020752B2|2015-04-28|Method and device for indoor positioning using magnetic field properties Park et al.2017|Self-corrective knowledge-based hybrid tracking system using BIM and multimodal sensors US8633817B2|2014-01-21|Mapping wireless signals with motion sensors EP3213031B1|2020-11-25|Simultaneous localization and mapping by using earth's magnetic fields US20150354951A1|2015-12-10|Method and Apparatus for Determination of Misalignment Between Device and Pedestrian KR102022668B1|2019-09-18|Landmark positioning JP5963907B2|2016-08-03|Wireless access point information collection method and apparatus KR100926783B1|2009-11-13|Method for self-localization of a robot based on object recognition and environment information around the recognized object EP2616774B1|2017-04-26|Indoor positioning using pressure sensors TWI500003B|2015-09-11|Positioning and mapping based on virtual landmarks JP2016001875A|2016-01-07|Mobile object positioning apparatus Chen et al.2011|A smart phone based PDR solution for indoor navigation
同族专利:
公开号 | 公开日 CN109470238A|2019-03-15| EP3680616A4|2021-06-02| WO2019047637A1|2019-03-14|
引用文献:
公开号 | 申请日 | 公开日 | 申请人 | 专利标题
法律状态:
2019-03-16| STAA| Information on the status of an ep patent application or granted ep patent|Free format text: STATUS: THE INTERNATIONAL PUBLICATION HAS BEEN MADE | 2020-06-12| STAA| Information on the status of an ep patent application or granted ep patent|Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE | 2020-06-12| PUAI| Public reference made under article 153(3) epc to a published international application that has entered the european phase|Free format text: ORIGINAL CODE: 0009012 | 2020-07-15| AK| Designated contracting states|Kind code of ref document: A1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR | 2020-07-15| AX| Request for extension of the european patent|Extension state: BA ME | 2020-07-15| 17P| Request for examination filed|Effective date: 20200407 | 2020-12-16| DAV| Request for validation of the european patent (deleted)| 2020-12-16| DAX| Request for extension of the european patent (deleted)| 2021-06-02| A4| Supplementary search report drawn up and despatched|Effective date: 20210506 | 2021-06-02| RIC1| Information provided on ipc code assigned before grant|Ipc: G01C 21/16 20060101AFI20210429BHEP Ipc: G01C 21/20 20060101ALI20210429BHEP Ipc: G01C 22/00 20060101ALI20210429BHEP | 2021-12-03| STAA| Information on the status of an ep patent application or granted ep patent|Free format text: STATUS: EXAMINATION IS IN PROGRESS | 2022-01-05| 17Q| First examination report despatched|Effective date: 20211203 |
优先权:
[返回顶部]
申请号 | 申请日 | 专利标题 相关专利
Sulfonates, polymers, resist compositions and patterning process
Washing machine
Washing machine
Device for fixture finishing and tension adjusting of membrane
Structure for Equipping Band in a Plane Cathode Ray Tube
Process for preparation of 7 alpha-carboxyl 9, 11-epoxy steroids and intermediates useful therein an
国家/地区
|