How to play audio in iOS in background? - codenameone

My goal is to make iOS play audio in a background.
When the audio is received by iPhone, it's not played unless I open the app in the foreground.
Basically, the iPhone is "waiting" when I click to open the app, and then it plays the audio.
Following Developer Guide, I added ios.background_modes=music to my .properties file.
However, nothing changed in behaviour.
I receive audio in real time via websocket class (onMessage(byte[] message) method).
What approach would you suggest in order to resolve this issue and make iOS play audio in a background?

This is the syntax for adding via the GUI not editing the properties file directly. You need to add the property with a codename1.arg. prefix.
Specifically: codename1.arg.ios.background_modes=music

Related

Video streaming including application overlay components

I want to make a react-native app having the capability of video streaming from a mobile app to a connected browser user. On top of that, I want to overlay some application components so connected users can see video streaming as well as some of the application UI.
For an example take a reference of the below-given image. Here, video streaming is running in the car showroom and there are a few app components shown as an overlay of the video like an app menu and a car image.
I want to achive same functionality and using VideoSDK platform for video streaming service.
So far I have created react-native app and able to stream video through camera to the connected browser user.
Next, I want to add my app menu on top of the video as per the image and therefore i am thinking screenshare with combination of video sharing is way to go.
The above image is the actual implementation using video SDK in the browser but as you can see screen share window is opening in a totally different context which is not the expected implementation.
Can someone suggest how can I achieve the functionality of video streaming having the capability of app overlay components?
I have reviewed your requirement and I am glad to inform you that we do have application with same requirements, for further discussion and demos can we connect over mail i.e. karan10010#gmail.com

React-Native Audio Waveform editor

I am planning to build an audio editor app with react-native. The functionalities include having a textbox where user can provide the URL for any audio file. Once the file is loaded on the UI, it will be played with a Waveform UI. User can select the start and endpoints of the audio by moving the slider on the waveform and once it's fixed, the app will get the start-time and end-time of the selected waveform, which will be then sent to the backend to cut the audio(probably using FFmpeg library).
I need but can't seem to find any react-native library that allows the user to interact with the waveform.
The UI can be somewhat similar to:
I don't believe that there is one that allows users to interact with the waveform out of the box.
You could use react-native-audiowaveform to show the waveform, and then capture the user's touches.

codeName one : full duplex conversation

I have a camera with a microphone and a speaker.
I want to make a full duplex audio communication between the browser of a terminal device and this camera.
Is it possible with CodeName One?
If so, can I have a tutorial?
We don't have a tutorial for this. If you want to use a browser within the app to do this you can use the BrowserComponent notice that this won't behave the same on the simulator and the device.

Is there any way to stream music in the background with Ionic framework?

I'm developing an Android app using Ionic for music streaming, like a personal Spotify. I can easily play an audio file using Streaming Media plugin, but it seems that this plugin was made for video purposes, because it always brings the player to the front in fullscreen, so you can't navigate through the app while the music is playing, you also can't lock your screen or the playback is stopped.
Is there any way to stream music in the background? Like when I tap the song I want to play, it just starts playing without bringing in another screen? (like you can easily do in AndroidStudio with MediaPlayer). I've search it anywhere for the past week and I just can't find any solution.
I realized Background-Audio in my ionic2/3 Project with
cordova-plugin-media 3.0.1 "Media"
cordova-plugin-music-controls 2.1.4 "MusicControls"
and to workaround some issues on ios I need to add this plugin also
nl.kingsquare.cordova.background-audio 1.0.1 "background-audio"

Xilium CefGlue in Vline can only receive calls

I have set up a vLine agent withing Xilium CefGlue and mostly it works. It can send text messages, receive voice and video calls that function as expected but when a voice or video call is made out of CefGlue to a Chrome browser or CefGlue the call appears to connect but no video or audio makes it from one client to the other. Local vidoe is displayed and it appears local audio is being captured but the other client can see nor hear anything.
[1202/163721:ERROR:webrtc_audio_renderer.cc(231)] NOT IMPLEMENTED shows up in CefGlue.log
Any ideas what can be done to make this work?
Run application (or add it programmatically) with --enable-media-stream switch.
http://apprtc.appspot.com/ should work.
If problems still persists, but works in chrome - then you probably should report issue into CEF issue tracker, or there is problem with vLine (what is it?).

Resources