Skip to content

Repo to store SAM applications both for server and client

License

Notifications You must be signed in to change notification settings

Devon-Wayman/SamProject

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

30 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Contributors Forks Stargazers Issues LinkedIn


SAM Project

Official project documentation for SAM (Smoke and Mirrors)

About The Project

Product Name Screen Shot

Operation Example

Built With

The following programs and libraries were used within this application's development:

Getting Started

This is an example of how you may give instructions on setting up your project locally. To get a local copy up and running follow these simple example steps.

Prerequisites

Hardware Needed:

  • iDevice (iPhone or iPad with face tracking capability)
  • OSX laptop or desktop (Windows platform support currently in the works)
  • A wireless router (5G band capable recommended). An adaptor to directly connect the iDevice and computer hosting the server can also be used
  • Soft LED light - Recommended to evenly illumiate performer's face to aid in facial tracking
  • XCode - Required to build the mobile app used for tracking performers face and sending data to the server

Installation

The instructions to setup the application are as follows

  1. Download the server build on an OSX machine (Windows support coming soon) as well as the XCode source project You will need an Apple developer account to build and deploy the mobile app required for face tracking at this time though more universal methods are being looked into.

  2. Launch the server application on the computer you intend to project from. You will be prompted to allow incoming connections; allow these. Failure to do so will prevent the mobile app on your LAN from reaching the server to update the face model.

  3. Build and deploy the mobile app to an iDevice via XCode. iOS 15 or greater is required

  4. Launch the mobile app. You will be asked to allow connections to devices on your local network. Again, allow this so that the mobile app can send the required data to the server.

  5. With the mobile and server app open, enter the iPv4 address of the server machine assigned by your router. If the address is correct and proper networking permissions have been granted, you should see the head model appear within the server application and follow the user's motions and expressions.

Usage and Tips

  • The application can be simply projected onto a wall though using polyester chiffon (recommend solid-platinum) is highly recommended as it is a great material to use for projecting images onto. With the proper lighting the fabric itself will be practically invisible, making the head appear to be floating in the air.

  • It is recommended that the performer have a soft LED light illuminating their face during use. While the main facial expression system is powered via the IR depth sensor of the iDevice, adaquet lighting will increase tracking accuracy and reduce stutter

  • To lower latency and decrease throttling on the user's local network, a LAN to Lightning Bolt/USB-C connector can be used to directly connect the mobile device and the laptop hosting the server directly. This will decrease traffic on the local wifi and greatly improve performance

Contact

Devon Wayman - devonwayman97@gmail.com

Project Link: Here

Acknowledgments

Special thanks to the following developers for their resources, inspiration and guidance

  • Dilmer Valecillos - For his treasure trove of AR development based educational resources.
  • Tom Weiland - For his open source networking framework, Riptide.

About

Repo to store SAM applications both for server and client

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published