mirror of
https://github.com/zakk4223/CocoaSplit.git
synced 2026-05-15 14:15:50 -06:00
Removed some spammy NSLogs
This commit is contained in:
parent
2e81011152
commit
cfb5fe4f6b
25 changed files with 0 additions and 154 deletions
|
|
@ -172,7 +172,6 @@
|
|||
|
||||
if (!self.currentDisplay)
|
||||
{
|
||||
NSLog(@"NO DISPLAY");
|
||||
return NO;
|
||||
}
|
||||
|
||||
|
|
@ -289,7 +288,6 @@
|
|||
|
||||
if (_displayStreamRef)
|
||||
{
|
||||
NSLog(@"STOP DISPLAY STREAM");
|
||||
CGDisplayStreamStop(_displayStreamRef);
|
||||
}
|
||||
|
||||
|
|
@ -387,7 +385,6 @@
|
|||
|
||||
-(void)dealloc
|
||||
{
|
||||
NSLog(@"DEALLOC DISPLAY STREAM");
|
||||
[self removeObserver:self forKeyPath:@"propertiesChanged"];
|
||||
[[NSNotificationCenter defaultCenter] removeObserver:self];
|
||||
}
|
||||
|
|
|
|||
|
|
@ -268,7 +268,6 @@
|
|||
for(sserv in servers)
|
||||
{
|
||||
|
||||
NSLog(@"Syphon UUID %@", [sserv objectForKey:SyphonServerDescriptionUUIDKey ]);
|
||||
NSString *sy_name = [NSString stringWithFormat:@"%@ - %@", [sserv objectForKey:SyphonServerDescriptionAppNameKey], [sserv objectForKey:SyphonServerDescriptionNameKey]];
|
||||
|
||||
CSAbstractCaptureDevice *newDev;
|
||||
|
|
|
|||
|
|
@ -21,7 +21,6 @@
|
|||
if (self = [super init])
|
||||
{
|
||||
|
||||
//NSLog(@"SUPER CLASS IS %@", super);
|
||||
|
||||
_nextCaptureTime = 0.0f;
|
||||
self.captureFPS = 30.0f;
|
||||
|
|
|
|||
|
|
@ -23,9 +23,4 @@
|
|||
return self;
|
||||
}
|
||||
|
||||
-(void)dealloc
|
||||
{
|
||||
NSLog(@"WINDOW CAPTURE VIEW DEALLOC");
|
||||
|
||||
}
|
||||
@end
|
||||
|
|
|
|||
|
|
@ -211,7 +211,6 @@
|
|||
_audio_capture_output = [[AVCaptureAudioDataOutput alloc] init];
|
||||
|
||||
|
||||
NSLog(@"AUDIO SAMPLERATE %d BITRATE %d", self.audioSamplerate, self.audioBitrate);
|
||||
if (self.useAudioEngine)
|
||||
{
|
||||
|
||||
|
|
|
|||
|
|
@ -110,7 +110,6 @@ void BufferCompletedPlaying(void *userData, ScheduledAudioSlice *bufferList);
|
|||
|
||||
|
||||
|
||||
//NSLog(@"SCHED SLICE %d", err);
|
||||
|
||||
return YES;
|
||||
}
|
||||
|
|
@ -119,8 +118,6 @@ void BufferCompletedPlaying(void *userData, ScheduledAudioSlice *bufferList);
|
|||
{
|
||||
|
||||
|
||||
//NSLog(@"THE BUFFER %@", sampleBuffer);
|
||||
|
||||
//credit to TheAmazingAudioEngine for an illustration of proper audiobufferlist allocation. Google leads to some really really bad allocation code...
|
||||
|
||||
AudioBufferList *sampleABL;
|
||||
|
|
@ -139,7 +136,6 @@ void BufferCompletedPlaying(void *userData, ScheduledAudioSlice *bufferList);
|
|||
|
||||
sampleABL->mNumberBuffers = bufferCnt;
|
||||
|
||||
//NSLog(@"BUFFERCNT %d CHANNELCNT %d, SAMPLES %ld, BYTES/FRAME %d", bufferCnt, channelCnt, numSamples, asbd->mBytesPerFrame);
|
||||
|
||||
for (int i=0; i<bufferCnt; i++)
|
||||
{
|
||||
|
|
@ -196,25 +192,12 @@ void BufferCompletedPlaying(void *userData, ScheduledAudioSlice *bufferList);
|
|||
ts.mFlags = kAudioTimeStampSampleTimeValid;
|
||||
ts.mSampleTime = -1;
|
||||
err = AudioUnitSetProperty(self.audioUnit, kAudioUnitProperty_ScheduleStartTimeStamp, kAudioUnitScope_Global, 0, &ts, sizeof(ts));
|
||||
|
||||
|
||||
|
||||
|
||||
//NSLog(@"PLAYER INPUT rate %f, %d channels, %d bits per channel", asbd.mSampleRate, asbd.mChannelsPerFrame, asbd.mBitsPerChannel);
|
||||
|
||||
|
||||
//NSLog(@"SCHEDULE START %d", err);
|
||||
|
||||
|
||||
}
|
||||
@end
|
||||
|
||||
|
||||
void BufferCompletedPlaying(void *userData, ScheduledAudioSlice *bufferList)
|
||||
{
|
||||
//NSLog(@"BUFFER COMPLETED PLAYING!!!! COMP %d BEGAN %d LATE %d", bufferList->mFlags & kScheduledAudioSliceFlag_Complete, bufferList->mFlags & kScheduledAudioSliceFlag_BeganToRender, bufferList->mFlags & kScheduledAudioSliceFlag_BeganToRenderLate);
|
||||
|
||||
|
||||
CAMultiAudioPCM *pcmObj = (__bridge CAMultiAudioPCM *)(userData);
|
||||
//maybe put this on a dedicated queue?
|
||||
//why a queue? don't want to do any sort of memory/managed object operations in an audio callback.
|
||||
|
|
|
|||
|
|
@ -27,8 +27,6 @@
|
|||
AudioComponentInstanceNew(comp, &_audioUnit);
|
||||
OSStatus err;
|
||||
err = AudioUnitInitialize(_audioUnit);
|
||||
NSLog(@"AU INIT %d", err);
|
||||
|
||||
}
|
||||
|
||||
|
||||
|
|
@ -51,10 +49,8 @@
|
|||
|
||||
AudioUnitGetProperty(_audioUnit, kAudioUnitProperty_StreamFormat, kAudioUnitScope_Output, 1, &asbd, &asbdSize);
|
||||
|
||||
NSLog(@"OUTPUT ASBD SAMPLERATE %f", asbd.mSampleRate);
|
||||
|
||||
err = AudioUnitSetProperty(toNode, kAudioUnitProperty_MakeConnection, kAudioUnitScope_Input, 1, &conn, sizeof(conn));
|
||||
NSLog(@"UNIT SET CONN %d", err);
|
||||
|
||||
err = AudioUnitSetProperty(toNode, kAudioUnitProperty_StreamFormat, kAudioUnitScope_Input, 1, &asbd, asbdSize);
|
||||
|
||||
|
|
|
|||
|
|
@ -35,7 +35,6 @@
|
|||
|
||||
if (bufferSize > self.preallocatedBuffersize)
|
||||
{
|
||||
NSLog(@"ALLOCATING FOR %d", bufferSize);
|
||||
_pcmData = malloc(bufferSize*2); //Assuming deinterleaved 2-ch, so allocate enough space for both channels
|
||||
self.preallocatedBuffersize = bufferSize;
|
||||
}
|
||||
|
|
@ -64,12 +63,6 @@
|
|||
|
||||
//for now assume Float32, 2 channel, non-interleaved. We have to interleave it outselves here.
|
||||
|
||||
|
||||
|
||||
|
||||
//NSLog(@"ENCODE BUFFER SIZE %u", (unsigned int)bufsize);
|
||||
|
||||
|
||||
AudioBuffer buffer0 = pcmBuffer->mBuffers[0];
|
||||
AudioBuffer buffer1 = pcmBuffer->mBuffers[1];
|
||||
Float32 *data0 = buffer0.mData;
|
||||
|
|
@ -298,7 +291,6 @@
|
|||
|
||||
|
||||
|
||||
// NSLog(@"CODEC INIT %d COOKIE SIZE %u MAX %u", err, magicCookie->mMagicCookieSize, cookiestructsize);
|
||||
outputSampleCount = 0;
|
||||
|
||||
|
||||
|
|
@ -310,10 +302,6 @@
|
|||
cookiestructsize = sizeof(CFDictionaryRef);
|
||||
|
||||
AudioCodecGetProperty(aacCodec, kAudioCodecPropertySettings, &cookiestructsize, &encoderState);
|
||||
//NSLog(@"ENCODER STATE %@", encoderState);
|
||||
|
||||
|
||||
|
||||
return;
|
||||
|
||||
}
|
||||
|
|
|
|||
|
|
@ -91,8 +91,6 @@
|
|||
CSAnimationRunnerObj *runner = [CaptureController sharedAnimationObj];
|
||||
|
||||
|
||||
NSLog(@"LOADING ANIMATIONS %@", runner);
|
||||
|
||||
NSDictionary *animations = [runner allAnimations];
|
||||
NSMutableArray *tmpList = [NSMutableArray array];
|
||||
|
||||
|
|
|
|||
|
|
@ -183,7 +183,6 @@
|
|||
|
||||
NSString *pName = [[NSString alloc] initWithBytes:propName length:strlen(propName) encoding:NSUTF8StringEncoding];
|
||||
id propertyValue = [self valueForKey:pName];
|
||||
NSLog(@"SETTING %@ TO %@",pName, propertyValue);
|
||||
|
||||
[newCopy setValue:propertyValue forKey:pName];
|
||||
}
|
||||
|
|
|
|||
|
|
@ -55,7 +55,6 @@
|
|||
-(void)dealloc
|
||||
{
|
||||
|
||||
NSLog(@"DEALLOC TEXT CAPTURE BASE!");
|
||||
|
||||
NSFontManager *fontManager = [NSFontManager sharedFontManager];
|
||||
|
||||
|
|
|
|||
|
|
@ -477,13 +477,6 @@
|
|||
NSError *compressError;
|
||||
|
||||
|
||||
if (![self.compressSettingsController commitEditing])
|
||||
{
|
||||
NSLog(@"FAILED TO COMMIT EDITING FOR COMPRESS EDIT");
|
||||
}
|
||||
|
||||
|
||||
|
||||
if (self.editingCompressor)
|
||||
{
|
||||
|
||||
|
|
@ -644,7 +637,6 @@
|
|||
|
||||
[self.streamServicePluginViewController.view removeFromSuperview];
|
||||
|
||||
NSLog(@"STREAM SERVICE CONFI WINDOW %@", self.streamServiceConfWindow);
|
||||
|
||||
[NSApp endSheet:self.streamServiceConfWindow];
|
||||
[self.streamServiceConfWindow close];
|
||||
|
|
@ -1052,11 +1044,8 @@ static CVReturn displayLinkRender(CVDisplayLinkRef displayLink, const CVTimeStam
|
|||
static dispatch_once_t onceToken;
|
||||
dispatch_once(&onceToken, ^{
|
||||
NSString *animationPluginPath = [[NSBundle mainBundle] pathForResource:@"CSAnimationRunner" ofType:@"plugin"];
|
||||
NSLog(@"ANIM PATH %@", animationPluginPath);
|
||||
NSBundle *animationBundle = [NSBundle bundleWithPath:animationPluginPath];
|
||||
NSLog(@"ANIM BUNDLE %@", animationBundle);
|
||||
Class animationClass = [animationBundle classNamed:@"CSAnimationRunnerObj"];
|
||||
NSLog(@"ANIM CLASS %@", animationClass);
|
||||
|
||||
sharedAnimationObj = [[animationClass alloc] init];
|
||||
|
||||
|
|
@ -1562,15 +1551,6 @@ static CVReturn displayLinkRender(CVDisplayLinkRef displayLink, const CVTimeStam
|
|||
}
|
||||
|
||||
|
||||
/*
|
||||
NSString *animationPluginPath = [[NSBundle mainBundle] pathForResource:@"CSAnimationRunner" ofType:@"plugin"];
|
||||
NSLog(@"ANIM PATH %@", animationPluginPath);
|
||||
NSBundle *animationBundle = [NSBundle bundleWithPath:animationPluginPath];
|
||||
NSLog(@"ANIM BUNDLE %@", animationBundle);
|
||||
Class animationClass = [animationBundle classNamed:@"CSAnimationRunnerObj"];
|
||||
NSLog(@"ANIM CLASS %@", animationClass);
|
||||
//self.animationRunner = [[animationClass alloc] init];
|
||||
*/
|
||||
self.extraPluginsSaveData = nil;
|
||||
|
||||
|
||||
|
|
@ -2153,28 +2133,15 @@ static CVReturn displayLinkRender(CVDisplayLinkRef displayLink, const CVTimeStam
|
|||
{
|
||||
|
||||
|
||||
//NSLog(@"AUDIO BUFFER %@", sampleBuffer);
|
||||
if (!self.captureRunning)
|
||||
{
|
||||
return;
|
||||
}
|
||||
|
||||
/*
|
||||
if (_firstFrameTime == 0)
|
||||
{
|
||||
//Don't start sending audio to the outputs until a video frame has arrived, with AVFoundation this can take 2+ seconds (!?)
|
||||
//Might need to prime the capture session first...
|
||||
return;
|
||||
}
|
||||
*/
|
||||
// NSLog(@"AUDIO SAMPLE BUFFER %@", sampleBuffer);
|
||||
|
||||
|
||||
CMTime orig_pts = CMSampleBufferGetPresentationTimeStamp(sampleBuffer);
|
||||
|
||||
|
||||
|
||||
//NSLog(@"AUDIO PTS %@", CMTimeCopyDescription(kCFAllocatorDefault, orig_pts));
|
||||
|
||||
if (CMTIME_COMPARE_INLINE(_firstAudioTime, ==, kCMTimeZero))
|
||||
{
|
||||
|
|
@ -2188,7 +2155,6 @@ static CVReturn displayLinkRender(CVDisplayLinkRef displayLink, const CVTimeStam
|
|||
CMTime pts = CMTimeAdd(real_pts, adjust_pts);
|
||||
|
||||
|
||||
//NSLog(@"AUDIO PTS %@", CMTimeCopyDescription(kCFAllocatorDefault, pts));
|
||||
|
||||
CMSampleBufferSetOutputPresentationTimeStamp(sampleBuffer, pts);
|
||||
|
||||
|
|
@ -2339,7 +2305,6 @@ static CVReturn displayLinkRender(CVDisplayLinkRef displayLink, const CVTimeStam
|
|||
|
||||
if (![self sleepUntil:(startTime += _frame_interval)])
|
||||
{
|
||||
//NSLog(@"SLEEP FAILED");
|
||||
continue;
|
||||
}
|
||||
|
||||
|
|
@ -2499,7 +2464,6 @@ static CVReturn displayLinkRender(CVDisplayLinkRef displayLink, const CVTimeStam
|
|||
|
||||
CFAbsoluteTime ptsTime = _frame_time - _firstFrameTime;
|
||||
|
||||
//NSLog(@"PTS TIME IS %f", ptsTime);
|
||||
|
||||
|
||||
_frameCount++;
|
||||
|
|
@ -2507,7 +2471,6 @@ static CVReturn displayLinkRender(CVDisplayLinkRef displayLink, const CVTimeStam
|
|||
|
||||
|
||||
pts = CMTimeMake(ptsTime*1000000, 1000000);
|
||||
//NSLog(@"PTS TIME IS %@", CMTimeCopyDescription(kCFAllocatorDefault, pts));
|
||||
|
||||
duration = CMTimeMake(1000, self.captureFPS*1000);
|
||||
|
||||
|
|
|
|||
|
|
@ -240,7 +240,6 @@
|
|||
CVImageBufferRef imageBuffer = frameData.videoFrame;
|
||||
|
||||
|
||||
//NSLog(@"WIDTH INPUT %zd HEIGHT %zd", CVPixelBufferGetWidth(imageBuffer), CVPixelBufferGetHeight(imageBuffer));
|
||||
|
||||
src_height = CVPixelBufferGetHeight(imageBuffer);
|
||||
src_width = CVPixelBufferGetWidth(imageBuffer);
|
||||
|
|
|
|||
|
|
@ -243,16 +243,13 @@ void getAudioExtradata(char *cookie, char **buffer, size_t *size)
|
|||
pkt.destruct = NULL;
|
||||
|
||||
|
||||
//NSLog(@"FFMPEG AUDIO PTS %lld/%d", pts.value, pts.timescale);
|
||||
|
||||
pkt.pts = av_rescale_q(pts.value, (AVRational) {1.0, pts.timescale}, _av_audio_stream->time_base);
|
||||
|
||||
|
||||
|
||||
//pkt.dts = pkt.pts;
|
||||
|
||||
|
||||
// pkt.pts = pts.value;
|
||||
if (av_interleaved_write_frame(_av_fmt_ctx, &pkt) < 0)
|
||||
{
|
||||
NSLog(@"AV WRITE AUDIO failed for %@", self.stream_output);
|
||||
|
|
@ -396,7 +393,6 @@ void getAudioExtradata(char *cookie, char **buffer, size_t *size)
|
|||
if (!(av_out_fmt->flags & AVFMT_NOFILE))
|
||||
{
|
||||
int av_err;
|
||||
NSLog(@"Doing AVIO_OPEN");
|
||||
if ((av_err = avio_open(&_av_fmt_ctx->pb, [_stream_output UTF8String], AVIO_FLAG_WRITE)) < 0)
|
||||
{
|
||||
NSString *open_err = [self av_error_nsstring:av_err ];
|
||||
|
|
@ -497,7 +493,6 @@ void getAudioExtradata(char *cookie, char **buffer, size_t *size)
|
|||
|
||||
if ([self resetOutputIfNeeded])
|
||||
{
|
||||
NSLog(@"OUTPUT RESET");
|
||||
return;
|
||||
}
|
||||
|
||||
|
|
@ -516,7 +511,6 @@ void getAudioExtradata(char *cookie, char **buffer, size_t *size)
|
|||
|
||||
if (!self.active)
|
||||
{
|
||||
NSLog(@"NOT ACTIVE");
|
||||
return;
|
||||
}
|
||||
|
||||
|
|
|
|||
|
|
@ -285,20 +285,9 @@ static NSArray *_sourceTypes = nil;
|
|||
[self updateRotationTransform];
|
||||
|
||||
|
||||
//[self positionOrigin:x_pos y:y_pos];
|
||||
|
||||
//self.layoutPosition = self.layer.frame;
|
||||
//if (width && height)
|
||||
// {
|
||||
// [self updateSize:width height:height];
|
||||
// }
|
||||
self.layoutPosition = self.layer.frame;
|
||||
|
||||
|
||||
//self.layoutPosition = self.layer.frame;
|
||||
|
||||
//NSLog(@"INIT %f %f %f %f", x_pos, y_pos, width, height);
|
||||
|
||||
_selectedVideoType = [aDecoder decodeObjectForKey:@"selectedVideoType"];
|
||||
|
||||
|
||||
|
|
@ -918,7 +907,6 @@ static NSArray *_sourceTypes = nil;
|
|||
-(void)dealloc
|
||||
{
|
||||
|
||||
NSLog(@"DEALLOC SOURCE INPUT");
|
||||
[self deregisterVideoInput:self.videoInput];
|
||||
for(id vInput in self.videoSources)
|
||||
{
|
||||
|
|
@ -1504,7 +1492,6 @@ static NSArray *_sourceTypes = nil;
|
|||
|
||||
if (isnan(x))
|
||||
{
|
||||
NSLog(@"HOLY CRAP NaN");
|
||||
return;
|
||||
}
|
||||
|
||||
|
|
|
|||
|
|
@ -131,7 +131,6 @@
|
|||
-(void)setDestination:(NSString *)destination
|
||||
{
|
||||
|
||||
NSLog(@"Destination set to %@", destination);
|
||||
if ([destination hasPrefix:@"rtmp://"] || [destination hasPrefix:@"udp:"])
|
||||
{
|
||||
self.output_format = @"FLV";
|
||||
|
|
@ -210,7 +209,6 @@
|
|||
return;
|
||||
}
|
||||
|
||||
NSLog(@"ATTACHING OUTPUT");
|
||||
if (!self.ffmpeg_out)
|
||||
{
|
||||
newout = [[FFMpegTask alloc] init];
|
||||
|
|
@ -234,7 +232,6 @@
|
|||
newout.samplerate = self.settingsController.audioSamplerate;
|
||||
newout.audio_bitrate = self.settingsController.audioBitrate;
|
||||
|
||||
NSLog(@"SETTINGS %@, FPS %f", self.settingsController, self.settingsController.captureFPS);
|
||||
|
||||
|
||||
self.ffmpeg_out = newout;
|
||||
|
|
|
|||
|
|
@ -1244,7 +1244,6 @@
|
|||
{
|
||||
//Without the autorelease NSColor leaks objects
|
||||
|
||||
NSLog(@"Preview: Creating Pixel Buffer Pool %f x %f", size.width, size.height);
|
||||
|
||||
NSMutableDictionary *attributes = [NSMutableDictionary dictionary];
|
||||
[attributes setValue:[NSNumber numberWithInt:size.width] forKey:(NSString *)kCVPixelBufferWidthKey];
|
||||
|
|
|
|||
|
|
@ -79,7 +79,6 @@
|
|||
-(BOOL)application:(NSApplication *)sender delegateHandlesKey:(NSString *)key
|
||||
{
|
||||
|
||||
NSLog(@"HANDLES KEY? %@", key);
|
||||
|
||||
NSArray *keys = @[@"layouts", @"width", @"height", @"fps", @"activelayout"];
|
||||
|
||||
|
|
|
|||
|
|
@ -57,12 +57,10 @@
|
|||
|
||||
NSString *sourceKey = [NSString stringWithFormat:@"%@:%@", ofType, uniqueID];
|
||||
|
||||
NSLog(@"SOURCE KEY IS %@", sourceKey);
|
||||
|
||||
|
||||
id cachedSource = [self.cacheMap objectForKey:sourceKey];
|
||||
|
||||
NSLog(@"CACHED SOURCE %@", cachedSource);
|
||||
|
||||
if (!cachedSource)
|
||||
{
|
||||
|
|
|
|||
|
|
@ -104,7 +104,6 @@
|
|||
|
||||
-(void)doAnimation:(NSDictionary *)threadDict
|
||||
{
|
||||
//[CATransaction flush];
|
||||
|
||||
CSAnimationRunnerObj *runner = [CaptureController sharedAnimationObj];
|
||||
|
||||
|
|
@ -113,24 +112,6 @@
|
|||
CALayer *rootLayer = threadDict[@"rootLayer"];
|
||||
|
||||
|
||||
/*
|
||||
InputSource *src = inpMap[@"spinnah"];
|
||||
CALayer *layer = src.layer;
|
||||
|
||||
|
||||
NSLog(@"LAYER IS %@", layer);
|
||||
|
||||
[CATransaction begin];
|
||||
[CATransaction setAnimationDuration:3.5];
|
||||
layer.zPosition = -1;
|
||||
[CATransaction begin];
|
||||
[CATransaction setAnimationDuration:5.5];
|
||||
[layer setValue:@500.0f forKeyPath:@"bounds.size.width"];
|
||||
[CATransaction commit];
|
||||
[CATransaction commit];
|
||||
|
||||
*/
|
||||
|
||||
@try {
|
||||
[runner runAnimation:modName forInput:inpMap withSuperlayer:rootLayer];
|
||||
}
|
||||
|
|
|
|||
|
|
@ -18,7 +18,6 @@ int main(int argc, const char * argv[])
|
|||
@autoreleasepool {
|
||||
|
||||
// insert code here...
|
||||
NSLog(@"Hello, World!");
|
||||
|
||||
|
||||
|
||||
|
|
|
|||
|
|
@ -48,7 +48,6 @@
|
|||
- (void)doInject:(NSRunningApplication *)toInject
|
||||
{
|
||||
|
||||
NSLog(@"WILL INJECT INTO APPLICATION %s", [toInject.localizedName UTF8String]);
|
||||
|
||||
pid_t pid = toInject.processIdentifier;
|
||||
|
||||
|
|
|
|||
|
|
@ -43,7 +43,6 @@ XPCListenerDelegate *captureDelegate;
|
|||
- (void) dealloc
|
||||
{
|
||||
|
||||
NSLog(@"HOLY SHIT DEALLOC");
|
||||
if (self.xpcProxy)
|
||||
{
|
||||
[self.xpcProxy release];
|
||||
|
|
@ -58,7 +57,6 @@ XPCListenerDelegate *captureDelegate;
|
|||
{
|
||||
|
||||
|
||||
NSLog(@"SHOULD ACCEPT NEW CONNECTION\n");
|
||||
NSXPCInterface *helperInterface = [[NSXPCInterface interfaceWithProtocol:@protocol(QTHelperProtocol)] retain];
|
||||
NSXPCInterface *masterInterface = [[NSXPCInterface interfaceWithProtocol:@protocol(CapturedFrameProtocol)] retain];
|
||||
|
||||
|
|
@ -82,7 +80,6 @@ XPCListenerDelegate *captureDelegate;
|
|||
{
|
||||
|
||||
NSArray *devices = [QTCaptureDevice inputDevicesWithMediaType:QTMediaTypeVideo];
|
||||
NSLog(@"DEVICES IN HELPER %@", devices);
|
||||
|
||||
NSMutableArray *retArray = [[NSMutableArray alloc] init];
|
||||
QTCaptureDevice *devinstance;
|
||||
|
|
@ -279,7 +276,6 @@ void qt_xpc_peer_event_handler(xpc_connection_t conn, xpc_object_t event)
|
|||
|
||||
} else {
|
||||
const char *message = xpc_dictionary_get_string(event, "message");
|
||||
NSLog(@"RECEIVED XPC MESSAGE %s", message);
|
||||
|
||||
if (!strcmp(message, "list_devices"))
|
||||
{
|
||||
|
|
|
|||
|
|
@ -16,21 +16,5 @@ int main(int argc, const char *argv[])
|
|||
xpc_main(qt_xpc_handle_connection);
|
||||
|
||||
|
||||
/*
|
||||
XPCListenerDelegate *listenDelegate = [[XPCListenerDelegate new] retain];
|
||||
NSLog(@"LISTEN DELEGATE %@", listenDelegate);
|
||||
|
||||
NSXPCListener *listener = [[NSXPCListener serviceListener] retain];
|
||||
|
||||
//NSString *bundleId = [[NSBundle mainBundle] bundleIdentifier];
|
||||
//NSXPCListener *listener = [[NSXPCListener alloc] initWithMachServiceName:bundleId];
|
||||
|
||||
NSLog(@"LISTENER %@", listener);
|
||||
listener.delegate = listenDelegate;
|
||||
|
||||
[listener resume];
|
||||
NSLog(@"SETUP XPC LISTENER");
|
||||
exit(EXIT_FAILURE);
|
||||
*/
|
||||
|
||||
}
|
||||
|
|
|
|||
|
|
@ -40,7 +40,6 @@
|
|||
{
|
||||
|
||||
|
||||
NSLog(@"SELECTED SERVER %@ STREAMKEY %@", self.selectedServer, self.streamKey);
|
||||
|
||||
if (self.selectedServer)
|
||||
{
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue