We are working on a remote support project. Most of the features were completed with Wikitude succesfully.
Now, we have to use input plugin API to make "customer’s camera" as our "support technician source camera". We take a look at your documentation but we couldn’t success.
Our broadcasting video coming from RTMP server. We must make this as source camera for Augmentations.
Thank your for kind reponse. Most of all are ready.
Just i only need that, how can i use "YUVInputPlugin" to make a video streaming as a source of World?
I am using WebRTC to broadcast a video, i save the instant target objects. Everything is ok. But i need to know more details about YUTInputPlugin. Is it possible share a sample code block or a library?
Could you clarify what you refer to with the YUTInputPlugin? For our InputPlugin Feature, you can find a sample code in the JS or the Native SDK sample app. Additionally you can find details in the technical documentation (here the link to the Android JS SDK):
Here is an additional forum post dealing with WebRTC, which might help:
Thx and greetings
We are working on a remote support project and we need to use Input Plugin. As you know, it's necessary to broadcast a user screen to another user. But opposite side of the broadcast must use 1st user's screen as a camera source (World).
I saw that documentation but it is not so clear. Do we have to write a C++ plugin to get broadcasting to Wikitude source? I need more details. Thank you.
Yes you'd need to implement your own plugin using the plugins API feature. But as mentioned we don't have a specific sample or further code / documentation specifically for the remote AR use case (just the sample in the sample app that shows how the Input Plugins API can be used). If it helps, we can connect you to one of our Premium Partners working with Remote AR use cases that can assist you with the implementation of your app.
Thx and greetings
In respect to the item 4 Synchronization of rendering, I'm facing some trouble.
When streaming the video feed between two devices with different resolutions ans aspect ratios, the objects are placed in different positions. We tried to fix the camera and background textures to the same resolution (with Gökhan's help), but the problem remains.
Another problem is that each device uses it's own sensors data, changing the grid dimension when applying the Plane Detection sample.
Wikitude Unity SDK Professional Edition
After reading the other thread where Gökhan offered assistance, would it be possible for you to share the project with us in its current, with some instructions on how to build and run it properly? It seems like there might be slight differences between his version and yours and I just to make sure that we're focusing on the exact problem you are facing.
Regarding the sensor data issue, we have it on our roadmap and we'll let you know as soon as it becomes available.
I've replied to Gökhan on the other thread with all the information (code and screen shots) to replicate the issue.
We'll wait the data sensor issue development.