James Alfei edited System Hardware.tex  about 10 years ago

Commit id: d10dd206c53271098d025fd19a0c8db1ea8e9e27

deletions | additions      

       

\subsection{System Walkthrough}  To fully understand the system that has been developed, this section will cover what happens from initial setup to usage from a user perspective.  Firstly, the phidget program must be executed. After this, the full hardware kit contains 3 USB powered devices. All of these must be plugged in in order to proceed with The  application usage. When all devices have been plugged in, a short delay allows the user to switch to Google Earth and allow the system to fully take control of the system.  When Google earth has been focused and the system has booted completely, the hardware begins on Swansea, United Kingdom. This  can be used immediately. Firstly, the user will likely want to zoom in to see a closer view of the planet easily changed  andsatellite imagery. To do this, the user would push the slider forward which would emulate a key press on the keyboard and would zoom the view in. To do the opposite, the user  can pull the slider back and similarly, the view would zoom out. To stop zooming, the user would place the slider bar somewhere in the middle of the range of motion. This is outside even take into account  the thresholds and causes no action, users's physical location (via GPS  andfurther allows  other phidgets access to the system.  The next action a user may want to perform is to navigate around the map in a similar way to using arrow keys. This has been implemented in our system using the circular phidget. The phidget is logically split in software into 8 distinct sections. These sections replicate up, down, left and right and any combination of diagonal movement using these keys. The circular phidget can be very senesitive, so before actually actioning any commands, we first check that a finger or physical contact has been made with the device. When location based services) however  this has been confirmed, the current position is checked against was discarded for  our rules and the actions performed. When the user releases the phidget, the virtual keys are released. project for performance purposes.  Now that the user can zoom and navigate around, to make use of the new 3D imagery and height mapping available in Google Earth, When  the joystick application  has been implemented to look around both vertically and rotate around the z axis. This movement is triggered when loaded,  the jostick is moved beyond a certain threshold map will start moving upwards  in the x or y axis. When this occurs, the camera can point up or down (by pushing direction  the joystick up or down respectively) as well as look left and right using user is physically facing with their device. To turn  the same physical pushes on map,  the joystick. This movement is fluid and feels natural user physically moves  in its use. the direction they would like to travel in, making them feel more involved with the application.  One of the more interesting features of the system Speed  is that of the RFID sensor. We have 3 tags available in the system. Each of these tags has a predefined location within the software upon first run. These work in a way such that when a tag is scanned, an event is fired within the software to detect the unique tag that was scanned. Upon detection, controlled via tilt sensors (accelerometer). Holding the phone at 90 degrees with  the software screen at head height facing the user's eyes, there  will look up the stored location (e.g. Swansea University). We then use a combination of virtual mouse click be no movement. Tilting the device forwards or backwards will initiate forwards  and keyboard presses backwards movement relative  toeffectively "Search" for  the location. amount of tilt.  This operation is hidden behind will peak when  the physical screen cover of device is parallel with  the system. ground at any orientation.  Although a great feature, fixed RFID tags can be quite limited in their scope. These may prove to be ideal for a museum style environment, however we decided that in order to develop a system that can cater for a range The aim  of environments, we would need to be able the game is  to reprogram these tags "on find  the fly". To demonstrate this, we used major city shown at  the pressure sensor bottom  of the phidget kits to effectively act as screen. It's latitude and longitude are also shown for educational purposes. A score is also shown which is incremented each time  a reprogramming button. If the button was depressed past user hits  a certain threshold, that would tell the software to put location. To successfully hit a location,  the RFID scanner into "reporgramming mode". From here, user must hover  the system would still operate completely as normal, apart from a visual cue on centre of their  screen alerting over  the user circle and/or marker  thatreprogramming mode  is active, and the RFID scanner waiting for a tag to be detected. Upon detected a tag, a piece of code would be launched displayed at  that will automatically action locations' latitude and longitude. If  a placemarker within google maps. The coordinates will then be copied to user successfully hits  the system clipboard. After this has been done, location,  the coordinates device  will be parsed correctly into the software and saved vibrate, giving physical feedback  to the unique RFID tag. Upon a subsequent detection of that RFID tag, the search would launch for a specific latitude and longitude coordinate saved from before. This produces user, as well as updating  the desired result of saving and retrieving a location using reporgrammable RFID tags. score.  We believe that the final system is intuitive and logical in its usage and implementation, and allows a full sense of feedback and interaction to the user.