Page 87 - Kaleidoscope Academic Conference Proceedings 2022
P. 87

Extended reality – How to boost quality of experience and interoperability




                                 Start                        The ARN and LR modules were implemented using three
                                                              activities namely the Map activity, AR activity and landmark
                                                              activity in Android Studio. Map activity displays the user’s
                        Obtain accelerometer readings         current location on a  map  and allows the user to select a
                                                              destination point to where they need AR directions. The AR
                                                              activity displays the AR turning  point  of  objects during
                       Perform magnitude thresholding         navigation. The landmark activity allows the user to point
                                                              their phone at an unknown landmark and displays its details
                          Perform temporal filtering          in AR.  The SDE module  was  implemented  using  the
                                                              accelerometer reading of the user to estimate their position
                                                              in the absence of GPS. The AR activity and the SDE activity
                             Obtain step count                are interconnected to implement the hybrid approach of user
                                                              localization. Some screenshots of the implemented Android
                                           Foot stride        application are  illustrated  in  figures 5 and 6.  Figure 5
                                                              displays the working of the ARN module and Figure 6
                                                              illustrates the LR module.
                          Determine distance traveled
                                                              4.1   ARN and LR modules

                                  End                         The two main parameters for testing AR-based navigational
                                                              applications are locational accuracy and latency of object
                  Figure 4 – Flowchart of SDE algorithm       generation and latency. The first parameter is mainly crucial
                                                              for the LR module. The testing for the landmark recognition
               Algorithm 2 - Landmark recognition algorithm   module involved validation  of  the  location  received from
                                                              Google Maps and the location calculated by the landmark
                                                              recognition algorithm. The findings of our experiments are
           Input  –  User’s  height, User’s current location, Phone’s   shown in Table 1.
           azimuth
           Output – Location of pointed landmark

           1. Initialize accelerometer, magnetometer sensors

           2. Obtain the landmark  view from the camera upon  user
           action

           3. Calculate the angle of inclination (tilt) of the phone using
           the accelerometer and magnetometer sensors [12]

           4. Find the distance of the base of the landmark from the user
           by applying the tangent function to the height and angle of
           inclination

           5. Using the distance, azimuth, and the current location, find
           the latitude and longitude of the landmark using the inverse
           Haversine formula [13].
                                                                       Figure 5 – ARN module screenshots
           6. Return the landmark location
                                                              The second parameter being the latency of object appearance
                  4.  IMPLEMENTATION AND RESULTS              on the user’s screen depends on the following three factors:
                                                              the framework used for rendering AR objects, complexity of
           The MAR-PNS was developed using Android Studio (Arctic   the algorithm used for localization of object position, and the
           Fox 2020.3.1) [15], Google Cloud Platform [16] and Beyond   device specific functionalities such as processor speed. The
           AR Framework [17].  Google  Directions  API  [18]  and   major contribution of latency lies in the framework chosen
           Google Places API [19] are used for getting directions and   for the application. Experimentally it is found  that the
           details of locations. Each module was tested on Android   latency of AR object appearance is in the range of 60ms to
           smartphones in an urban environment and the results were   80ms with an average of 74 ms.
           validated against the physical locations and distances in real
           time.








                                                          – 41 –
   82   83   84   85   86   87   88   89   90   91   92