Security News
The Unpaid Backbone of Open Source: Solo Maintainers Face Increasing Security Demands
Solo open source maintainers face burnout and security challenges, with 60% unpaid and 60% considering quitting.
@dtelecom/react-native
Advanced tools
npm install @dtelecom/react-native @dtelecom/react-native-webrtc
yarn add @dtelecom/react-native @dtelecom/react-native-webrtc
This library depends on @dtelecom/react-native-webrtc
, which has additional installation instructions found here:
Once the @dtelecom/react-native-webrtc
dependency is installed, one last step is needed to finish the installation:
In your MainApplication.java file:
import com.livekit.reactnative.LiveKitReactNative;
import com.livekit.reactnative.audio.AudioType;
public class MainApplication extends Application implements ReactApplication {
@Override
public void onCreate() {
// Place this above any other RN related initialization
// When AudioType is omitted, it'll default to CommunicationAudioType
// use MediaAudioType if user is only consuming audio, and not publishing
LiveKitReactNative.setup(this, new AudioType.CommunicationAudioType());
//...
}
}
In your AppDelegate.m file:
#import "LivekitReactNative.h"
@implementation AppDelegate
- (BOOL)application:(UIApplication *)application didFinishLaunchingWithOptions:(NSDictionary *)launchOptions
{
// Place this above any other RN related initialization
[LivekitReactNative setup];
//...
}
We've included an example app that you can try out.
In your index.js
file, setup the dTelecom SDK by calling registerGlobals()
.
This sets up the required WebRTC libraries for use in Javascript, and is needed for LiveKit to work.
import { registerGlobals } from '@dtelecom/react-native';
// ...
registerGlobals();
A Room object can then be created and connected to.
import { Participant, Room, Track } from '@dtelecom/livekit-client';
import { useRoom, AudioSession, VideoView } from '@dtelecom/react-native';
/*...*/
// Create a room state
const [room] = useState(() => new Room());
// Get the participants from the room
const { participants } = useRoom(room);
useEffect(() => {
let connect = async () => {
await AudioSession.startAudioSession();
await room.connect(url, token, {});
console.log('connected to ', url, ' ', token);
};
connect();
return () => {
room.disconnect();
AudioSession.stopAudioSession();
};
}, [url, token, room]);
const videoView = participants.length > 0 && (
<VideoView
style={{ flex: 1, width: '100%' }}
videoTrack={participants[0].getTrack(Track.Source.Camera)?.videoTrack}
/>
);
API documentation is located here.
Additional documentation for the dTelecom SDK can be found at https://docs.dtelecom.org/
As seen in the above example, we've introduced a new class AudioSession
that helps
to manage the audio session on native platforms. This class wraps either AudioManager on Android, or AVAudioSession on iOS.
You can customize the configuration of the audio session with configureAudio
.
By default, the audio session is set up for bidirectional communication. In this mode, the audio framework exhibits the following behaviors:
If you're leveraging dTelecom primarily for media playback, you have the option to reconfigure the audio session to better suit media playback. Here's how:
Note: iOS audio session customization is in development, and will be documented here when released.
useEffect(() => {
let connect = async () => {
// configure audio session prior to starting it.
await AudioSession.configureAudio({
android: {
// currently supports .media and .communication presets
audioTypeOptions: AndroidAudioTypePresets.media,
},
});
await AudioSession.startAudioSession();
await room.connect(url, token, {});
};
connect();
return () => {
room.disconnect();
AudioSession.stopAudioSession();
};
}, [url, token, room]);
Instead of using our presets, you can further customize the audio session to suit your specific needs.
await AudioSession.configureAudio({
android: {
preferredOutputList: ['earpiece'],
// See [AudioManager](https://developer.android.com/reference/android/media/AudioManager)
// for details on audio and focus modes.
audioTypeOptions: {
manageAudioFocus: true,
audioMode: 'normal',
audioFocusMode: 'gain',
audioStreamType: 'music',
audioAttributesUsageType: 'media',
audioAttributesContentType: 'unknown',
},
},
});
await AudioSession.startAudioSession();
Enabling screenshare requires extra installation steps:
Android screenshare requires a foreground service with type mediaProjection
to be present.
The example app uses @voximplant/react-native-foreground-service for this.
Ensure that the service is labelled a mediaProjection
service like so:
<service android:name="com.voximplant.foregroundservice.VIForegroundService"
android:foregroundServiceType="mediaProjection" />
Once setup, start the foreground service prior to using screenshare.
iOS screenshare requires adding a Broadcast Extension to your iOS project. Follow the integration instructions here:
https://jitsi.github.io/handbook/docs/dev-guide/dev-guide-ios-sdk/#screen-sharing-integration
It involves copying the files found in this sample project to your iOS project, and registering a Broadcast Extension in Xcode.
It's also recommended to use CallKeep,
to register a call with CallKit (as well as turning on the voip
background mode).
Due to background app processing limitations, screen recording may be interrupted if the app is restricted
in the background. Registering with CallKit allows the app to continue processing for the duration of the call.
Once setup, iOS screenshare can be initiated like so:
const screenCaptureRef = React.useRef(null);
const screenCapturePickerView = Platform.OS === 'ios' && (
<ScreenCapturePickerView ref={screenCaptureRef} />
);
const startBroadcast = async () => {
if (Platform.OS === 'ios') {
const reactTag = findNodeHandle(screenCaptureRef.current);
await NativeModules.ScreenCapturePickerViewManager.show(reactTag);
room.localParticipant.setScreenShareEnabled(true);
} else {
room.localParticipant.setScreenShareEnabled(true);
}
};
return (
<View style={styles.container}>
/*...*/ // Make sure the ScreenCapturePickerView exists in the view tree.
{screenCapturePickerView}
</View>
);
You will not be able to publish camera or microphone tracks on iOS Simulator.
This error could happen if you are using yarn and have incompatible versions of dependencies with dtelecom-client.
To fix this, you can either:
Apache License 2.0
FAQs
dTelecom sdk for React Native
The npm package @dtelecom/react-native receives a total of 0 weekly downloads. As such, @dtelecom/react-native popularity was classified as not popular.
We found that @dtelecom/react-native demonstrated a healthy version release cadence and project activity because the last version was released less than a year ago. It has 2 open source maintainers collaborating on the project.
Did you know?
Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.
Security News
Solo open source maintainers face burnout and security challenges, with 60% unpaid and 60% considering quitting.
Security News
License exceptions modify the terms of open source licenses, impacting how software can be used, modified, and distributed. Developers should be aware of the legal implications of these exceptions.
Security News
A developer is accusing Tencent of violating the GPL by modifying a Python utility and changing its license to BSD, highlighting the importance of copyleft compliance.