π
π
π
π
Sensor Stream Pipe
Streaming a Dataset | Moetsi
Sensor Stream Tester | Moetsi
FrameStruct Frame Data Structure | Moetsi
Config File - Sensor Stream Pipe | Moetsi
Searchβ¦
π
π
π
π
Sensor Stream Pipe
What is Sensor Stream Pipe?
Getting Started
Installation Linux
Installation Windows
Installation Mac
Installation iOS (Unity Plugin)
Streaming a Video
Streaming with Azure Kinect DK
Streaming a Dataset
Streaming iOS ARKit RGB-D Data
Streaming an OAK-D using Xlink
Components Overview
Sensor Stream Server
Sensor Stream Client
Sensor Stream Tester
Sending Frames
IReaders
FrameStruct
Config File
IEncoders
Receiving Frames
Receiving Frames
How to Extend Sensor Stream Pipe
Add New Sensor Interface
Powered By
GitBook
Streaming iOS ARKit RGB-D Data
We have 2 different methods for streaming RGB-D on iOS:
1.
Streaming without running AR app (no Unity)
2.
Streaming while running AR app
(Unity Plugin)
Problems?! (shocker)
Reach out on
our discord
and we will get you going!
Streaming without running AR app
1.0 Build Sensor Stream Pipe for iOS (No Unity)
1.1 Follow the instructions:
Installation iOS (Unity Plugin)
β
2.0 Build Sensor Stream Pipe for Mac
2.1 Follow the instructions here:
Installation Mac
β
3.0 Run Sensor Stream Client on Mac
3.1 Navigate to
/Sensor-Stream-Pipe/build/bin
3.2 Run Sensor Stream Client with OpenCV visualization
1
./ssp_client_opencv 9999
Copied!
4.0 Run and Build Sensor Stream Server
4.1 Open ssp.xcodeproj in
Sensor-Stream-Pipe/build-ios
4.2 Sign, in Signing & Capabilities
4.3 Build and run ssp_server!
If running into issues update host to the ip address of the computer running ssp client opencv in
serve_ios_raw.yaml
:
Streaming while running AR app (Unity Plugin)
1.0 Build Sensor Stream Pipe for iOS (No Unity)
1.1 Follow the instructions:
Installation iOS (Unity Plugin)
β
2.0 Build Sensor Stream Pipe for Mac
2.1 Follow the instructions here:
Installation Mac
β
3.0 Setup Unity Project
3.1 git clone our sample project
1
git
clone
https
:
//
github
.
com
/
moetsi
/
SSP
-
Unity
-
Plugin
-
AR
-
App
.
git
Copied!
3.2 Update line 3 in Assets/StreamingAssets/serve_ios_raw.yaml to the correct destination IP address
4.0 Deploy the Unity Project
4.1 Go to Build Settings and hit "Build and Run"
4.2 Once the project has been built and XCode is open we will need to add 2 linker flags
-framework VideoToolbox
-lbz2
Go to "Build Settings" under the "Unity-iPhone" project and scroll down to "Other Linker Flags" and add the 2 linker flags
add script to stream data (needs ARSession reference)
Add plugin to /Assets/Plugins/iOS
Add linker flags once there
4.3 Run ssp_client_opencv on port 9999 on the device you are streaming frame data (check out
Streaming a Video
if need instruction on how to do this)
4.4 Deploy the project onto an RGB-D (LiDAR) iPhone and check out the RGB-D data streaming as the iPhone runs an Unity AR Application!
Previous
Streaming a Dataset
Next
Streaming an OAK-D using Xlink
Last modified
10mo ago
Copy link
Contents
Streaming without running AR app
1.0 Build Sensor Stream Pipe for iOS (No Unity)
2.0 Build Sensor Stream Pipe for Mac
3.0 Run Sensor Stream Client on Mac
4.0 Run and Build Sensor Stream Server
Streaming while running AR app (Unity Plugin)
1.0 Build Sensor Stream Pipe for iOS (No Unity)
2.0 Build Sensor Stream Pipe for Mac
3.0 Setup Unity Project
4.0 Deploy the Unity Project