gpt4 book ai didi

ios - Capturer(iPhone 摄像头)未提供给 TVIVideoCapturer for TwilioVideo iOS SDK

转载 作者:塔克拉玛干 更新时间:2023-11-02 22:15:07 25 4
gpt4 key购买 nike

我收到一个错误,我的 TwilioVideo 模块需要一个捕获器(相机或麦克风),但没有接收到该输入。在我们切换到 Cocoapods 安装 SDK 和 PureLayout UI 库后,这个错误开始发生。之前我们手动将所有这些依赖项安装到 XCode 中。

我正在开发 React Native iOS 0.40.0 版本,react-native-cli 版本为 1.0.0。我正在使用 XCode 版本 8.2.1 (8C1002),iPhone 6 模拟器在 iOS 10.2 上运行。我正在使用 Cocoapods 1.2.0 版。我使用的是 TwilioVideo SDK 版本 1.0.0-beta5。还有一个 1.0.0-beta6 版本,我也试过了(结果相同)。恢复到版本 1.0.0-beta4 确实消除了错误,这对我来说表明我实现注册音频和视频轨道的方式存在问题。

这是我的 Pod 文件:

source 'https://github.com/CocoaPods/Specs'
source 'https://github.com/twilio/cocoapod-specs'

target 'MyApp' do
# Uncomment the next line if you're using Swift or would like to use dynamic frameworks
# use_frameworks!

# Pods for MyApp
pod 'TwilioVideo', '1.0.0-beta5'
pod 'PureLayout', '~> 3.0'
target 'MapleNativeProviderTests' do
inherit! :search_paths
# Pods for testing
end

end

我已经基于这个存储库在 XCode 中实现了一个 TwilioVideo 模块:react-native-twilio-video-webrtc .他最近更新了存储库以适用于 React Native 0.40.0,它更改了 XCode 的导入语法。我已经尝试使用旧的导入语法和新的导入语法,但当我尝试装载我的视频组件时,我继续收到以下错误:

enter image description here

这是 TwilioVideo SDK 的文档.这是 TVIVideoCapturer .

我对 react-native-twilio-video-webrtc 进行了修改,它本质上只是 TwilioVideo SDK 的一个薄包装器,使用 RCT_EXPORT_METHOD 来公开关键的 API 方法。该库在 init 方法中初始化音频和视频轨道,这会导致一些令人讨厌的行为,这些行为与应用程序启动时事件监听器不接收回调有关。因此,我将这些轨道移至名为 initialize 的自定义、公开的 RCT_EXPORT_METHOD。这是我从应用程序中的特定 View 调用的,它安装视频并初始化相机/麦克风输入。

我对 TWVideoModule.m 的实现是:

#import "TWVideoModule.h"

static NSString* roomDidConnect = @"roomDidConnect";
static NSString* roomDidDisconnect = @"roomDidDisconnect";
static NSString* roomDidFailToConnect = @"roomDidFailToConnect";
static NSString* roomParticipantDidConnect = @"roomParticipantDidConnect";
static NSString* roomParticipantDidDisconnect = @"roomParticipantDidDisconnect";

static NSString* participantAddedVideoTrack = @"participantAddedVideoTrack";
static NSString* participantRemovedVideoTrack = @"participantRemovedVideoTrack";
static NSString* participantAddedAudioTrack = @"participantAddedAudioTrack";
static NSString* participantRemovedAudioTrack = @"participantRemovedAudioTrack";
static NSString* participantEnabledTrack = @"participantEnabledTrack";
static NSString* participantDisabledTrack = @"participantDisabledTrack";

static NSString* cameraDidStart = @"cameraDidStart";
static NSString* cameraWasInterrupted = @"cameraWasInterrupted";
static NSString* cameraDidStopRunning = @"cameraDidStopRunning";

@interface TWVideoModule () <TVIParticipantDelegate, TVIRoomDelegate, TVIVideoTrackDelegate, TVICameraCapturerDelegate>

@end

@implementation TWVideoModule

@synthesize bridge = _bridge;

RCT_EXPORT_MODULE();

- (dispatch_queue_t)methodQueue
{
return dispatch_get_main_queue();
}

- (NSArray<NSString *> *)supportedEvents
{
return @[roomDidConnect,
roomDidDisconnect,
roomDidFailToConnect,
roomParticipantDidConnect,
roomParticipantDidDisconnect,
participantAddedVideoTrack,
participantRemovedVideoTrack,
participantAddedAudioTrack,
participantRemovedAudioTrack,
participantEnabledTrack,
participantDisabledTrack,
cameraDidStopRunning,
cameraDidStart,
cameraWasInterrupted];
}


- (instancetype)init
{
self = [super init];
if (self) {

UIView* remoteMediaView = [[UIView alloc] init];
//remoteMediaView.backgroundColor = [UIColor blueColor];

//remoteMediaView.translatesAutoresizingMaskIntoConstraints = NO;
self.remoteMediaView = remoteMediaView;


UIView* previewView = [[UIView alloc] init];
//previewView.backgroundColor = [UIColor yellowColor];

//previewView.translatesAutoresizingMaskIntoConstraints = NO;
self.previewView = previewView;

}
return self;
}

- (void)dealloc
{
[self.remoteMediaView removeFromSuperview];
self.remoteMediaView = nil;

[self.previewView removeFromSuperview];
self.previewView = nil;

self.participant = nil;
self.localMedia = nil;
self.camera = nil;
self.localVideoTrack = nil;
self.videoClient = nil;
self.room = nil;
}

RCT_EXPORT_METHOD(initialize) {
self.localMedia = [[TVILocalMedia alloc] init];
self.camera = [[TVICameraCapturer alloc] init];

NSLog(@"Camera %@", self.camera);

self.camera.delegate = self;

self.localVideoTrack = [self.localMedia addVideoTrack:YES
capturer:self.camera
constraints:[self videoConstraints]
error:nil];

self.localAudioTrack = [self.localMedia addAudioTrack:YES];

if (!self.localVideoTrack) {
NSLog(@"Failed to add video track");
} else {
// Attach view to video track for local preview
[self.localVideoTrack attach:self.previewView];
}

}

此文件的其余部分与添加和删除轨道以及加入/断开 Twilio channel 有关,因此我没有包含它。我还有 TWVideoPreviewManagerTWRemotePreviewManager,它们只是为本地和远程视频流的媒体对象提供 UIView。

我的 TwilioVideoComponent.js 组件是:

import React, { Component, PropTypes } from 'react'
import {
NativeModules,
NativeEventEmitter
} from 'react-native';

import {
View,
} from 'native-base';

const {TWVideoModule} = NativeModules;

class TwilioVideoComponent extends Component {

state = {};

static propTypes = {
onRoomDidConnect: PropTypes.func,
onRoomDidDisconnect: PropTypes.func,
onRoomDidFailToConnect: PropTypes.func,
onRoomParticipantDidConnect: PropTypes.func,
onRoomParticipantDidDisconnect: PropTypes.func,
onParticipantAddedVideoTrack: PropTypes.func,
onParticipantRemovedVideoTrack: PropTypes.func,
onParticipantAddedAudioTrack: PropTypes.func,
onParticipantRemovedAudioTrack: PropTypes.func,
onParticipantEnabledTrack: PropTypes.func,
onParticipantDisabledTrack: PropTypes.func,
onCameraDidStart: PropTypes.func,
onCameraWasInterrupted: PropTypes.func,
onCameraDidStopRunning: PropTypes.func,
...View.propTypes,
};

_subscriptions = [];

constructor(props) {
super(props);

this.flipCamera = this.flipCamera.bind(this);
this.startCall = this.startCall.bind(this);
this.endCall = this.endCall.bind(this);

this._eventEmitter = new NativeEventEmitter(TWVideoModule)
}

//
// Methods

/**
* Initializes camera and microphone tracks
*/
initializeVideo() {
TWVideoModule.initialize();
}

flipCamera() {
TWVideoModule.flipCamera();
}

startCall({roomName, accessToken}) {
TWVideoModule.startCallWithAccessToken(accessToken, roomName);
}

endCall() {
TWVideoModule.disconnect();
}

toggleVideo() {
TWVideoModule.toggleVideo();
}

toggleAudio() {
TWVideoModule.toggleAudio();
}

_unregisterEvents() {
this._subscriptions.forEach(e => e.remove());
this._subscriptions = []
}

_registerEvents() {

this._subscriptions = [

this._eventEmitter.addListener('roomDidConnect', (data) => {
if (this.props.onRoomDidConnect) {
this.props.onRoomDidConnect(data)
}
}),

this._eventEmitter.addListener('roomDidDisconnect', (data) => {
if (this.props.onRoomDidDisconnect) {
this.props.onRoomDidDisconnect(data)
}
}),

this._eventEmitter.addListener('roomDidFailToConnect', (data) => {
if (this.props.onRoomDidFailToConnect) {
this.props.onRoomDidFailToConnect(data)
}
}),

this._eventEmitter.addListener('roomParticipantDidConnect', (data) => {
if (this.props.onRoomParticipantDidConnect) {
this.props.onRoomParticipantDidConnect(data)
}
}),

this._eventEmitter.addListener('roomParticipantDidDisconnect', (data) => {
if (this.props.onRoomParticipantDidDisconnect) {
this.props.onRoomParticipantDidDisconnect(data)
}
}),

this._eventEmitter.addListener('participantAddedVideoTrack', (data) => {
if (this.props.onParticipantAddedVideoTrack) {
this.props.onParticipantAddedVideoTrack(data)
}
}),

this._eventEmitter.addListener('participantRemovedVideoTrack', (data) => {
if (this.props.onParticipantRemovedVideoTrack) {
this.props.onParticipantRemovedVideoTrack(data)
}
}),

this._eventEmitter.addListener('participantAddedAudioTrack', (data) => {
if (this.props.onParticipantAddedAudioTrack) {
this.props.onParticipantAddedAudioTrack(data)
}
}),

this._eventEmitter.addListener('participantRemovedAudioTrack', (data) => {
if (this.props.onParticipantRemovedAudioTrack) {
this.props.onParticipantRemovedAudioTrack(data)
}
}),

this._eventEmitter.addListener('participantEnabledTrack', (data) => {
if (this.props.onParticipantEnabledTrack) {
this.props.onParticipantEnabledTrack(data)
}
}),

this._eventEmitter.addListener('participantDisabledTrack', (data) => {
if (this.props.onParticipantDisabledTrack) {
this.props.onParticipantDisabledTrack(data)
}
}),

this._eventEmitter.addListener('cameraDidStart', (data) => {
if (this.props.onCameraDidStart) {
this.props.onCameraDidStart(data)
}
}),

this._eventEmitter.addListener('cameraWasInterrupted', (data) => {
if (this.props.onCameraWasInterrupted) {
this.props.onCameraWasInterrupted(data)
}
}),

this._eventEmitter.addListener('cameraDidStopRunning', (data) => {
if (this.props.onCameraDidStopRunning) {
this.props.onCameraDidStopRunning(data)
}
})

]

}

componentWillMount() {
this._eventEmitter.addListener('cameraDidStart', (data) => {
if (this.props.onCameraDidStart) {
this.props.onCameraDidStart(data)
}
});
this._registerEvents()
}

componentWillUnmount() {
this._unregisterEvents()
}

render() {
return this.props.children || null
}
}

export default TwilioVideoComponent;

我不确定如何修改 XCode 以兼容 TwilioVideo beta5 API。任何帮助将不胜感激。

最佳答案

在您的 podfile 中,查找 # use_frameworks! 并删除 #。

关于ios - Capturer(iPhone 摄像头)未提供给 TVIVideoCapturer for TwilioVideo iOS SDK,我们在Stack Overflow上找到一个类似的问题: https://stackoverflow.com/questions/42075289/

25 4 0
Copyright 2021 - 2024 cfsdn All Rights Reserved 蜀ICP备2022000587号
广告合作:1813099741@qq.com 6ren.com