Removed some spammy NSLogs

This commit is contained in:
Zakk 2015-05-03 13:09:25 -04:00
parent 2e81011152
commit cfb5fe4f6b
25 changed files with 0 additions and 154 deletions

View file

@ -172,7 +172,6 @@
if (!self.currentDisplay)
{
NSLog(@"NO DISPLAY");
return NO;
}
@ -289,7 +288,6 @@
if (_displayStreamRef)
{
NSLog(@"STOP DISPLAY STREAM");
CGDisplayStreamStop(_displayStreamRef);
}
@ -387,7 +385,6 @@
-(void)dealloc
{
NSLog(@"DEALLOC DISPLAY STREAM");
[self removeObserver:self forKeyPath:@"propertiesChanged"];
[[NSNotificationCenter defaultCenter] removeObserver:self];
}

View file

@ -268,7 +268,6 @@
for(sserv in servers)
{
NSLog(@"Syphon UUID %@", [sserv objectForKey:SyphonServerDescriptionUUIDKey ]);
NSString *sy_name = [NSString stringWithFormat:@"%@ - %@", [sserv objectForKey:SyphonServerDescriptionAppNameKey], [sserv objectForKey:SyphonServerDescriptionNameKey]];
CSAbstractCaptureDevice *newDev;

View file

@ -21,7 +21,6 @@
if (self = [super init])
{
//NSLog(@"SUPER CLASS IS %@", super);
_nextCaptureTime = 0.0f;
self.captureFPS = 30.0f;

View file

@ -23,9 +23,4 @@
return self;
}
-(void)dealloc
{
NSLog(@"WINDOW CAPTURE VIEW DEALLOC");
}
@end

View file

@ -211,7 +211,6 @@
_audio_capture_output = [[AVCaptureAudioDataOutput alloc] init];
NSLog(@"AUDIO SAMPLERATE %d BITRATE %d", self.audioSamplerate, self.audioBitrate);
if (self.useAudioEngine)
{

View file

@ -110,7 +110,6 @@ void BufferCompletedPlaying(void *userData, ScheduledAudioSlice *bufferList);
//NSLog(@"SCHED SLICE %d", err);
return YES;
}
@ -119,8 +118,6 @@ void BufferCompletedPlaying(void *userData, ScheduledAudioSlice *bufferList);
{
//NSLog(@"THE BUFFER %@", sampleBuffer);
//credit to TheAmazingAudioEngine for an illustration of proper audiobufferlist allocation. Google leads to some really really bad allocation code...
AudioBufferList *sampleABL;
@ -139,7 +136,6 @@ void BufferCompletedPlaying(void *userData, ScheduledAudioSlice *bufferList);
sampleABL->mNumberBuffers = bufferCnt;
//NSLog(@"BUFFERCNT %d CHANNELCNT %d, SAMPLES %ld, BYTES/FRAME %d", bufferCnt, channelCnt, numSamples, asbd->mBytesPerFrame);
for (int i=0; i<bufferCnt; i++)
{
@ -196,25 +192,12 @@ void BufferCompletedPlaying(void *userData, ScheduledAudioSlice *bufferList);
ts.mFlags = kAudioTimeStampSampleTimeValid;
ts.mSampleTime = -1;
err = AudioUnitSetProperty(self.audioUnit, kAudioUnitProperty_ScheduleStartTimeStamp, kAudioUnitScope_Global, 0, &ts, sizeof(ts));
//NSLog(@"PLAYER INPUT rate %f, %d channels, %d bits per channel", asbd.mSampleRate, asbd.mChannelsPerFrame, asbd.mBitsPerChannel);
//NSLog(@"SCHEDULE START %d", err);
}
@end
void BufferCompletedPlaying(void *userData, ScheduledAudioSlice *bufferList)
{
//NSLog(@"BUFFER COMPLETED PLAYING!!!! COMP %d BEGAN %d LATE %d", bufferList->mFlags & kScheduledAudioSliceFlag_Complete, bufferList->mFlags & kScheduledAudioSliceFlag_BeganToRender, bufferList->mFlags & kScheduledAudioSliceFlag_BeganToRenderLate);
CAMultiAudioPCM *pcmObj = (__bridge CAMultiAudioPCM *)(userData);
//maybe put this on a dedicated queue?
//why a queue? don't want to do any sort of memory/managed object operations in an audio callback.

View file

@ -27,8 +27,6 @@
AudioComponentInstanceNew(comp, &_audioUnit);
OSStatus err;
err = AudioUnitInitialize(_audioUnit);
NSLog(@"AU INIT %d", err);
}
@ -51,10 +49,8 @@
AudioUnitGetProperty(_audioUnit, kAudioUnitProperty_StreamFormat, kAudioUnitScope_Output, 1, &asbd, &asbdSize);
NSLog(@"OUTPUT ASBD SAMPLERATE %f", asbd.mSampleRate);
err = AudioUnitSetProperty(toNode, kAudioUnitProperty_MakeConnection, kAudioUnitScope_Input, 1, &conn, sizeof(conn));
NSLog(@"UNIT SET CONN %d", err);
err = AudioUnitSetProperty(toNode, kAudioUnitProperty_StreamFormat, kAudioUnitScope_Input, 1, &asbd, asbdSize);

View file

@ -35,7 +35,6 @@
if (bufferSize > self.preallocatedBuffersize)
{
NSLog(@"ALLOCATING FOR %d", bufferSize);
_pcmData = malloc(bufferSize*2); //Assuming deinterleaved 2-ch, so allocate enough space for both channels
self.preallocatedBuffersize = bufferSize;
}
@ -64,12 +63,6 @@
//for now assume Float32, 2 channel, non-interleaved. We have to interleave it outselves here.
//NSLog(@"ENCODE BUFFER SIZE %u", (unsigned int)bufsize);
AudioBuffer buffer0 = pcmBuffer->mBuffers[0];
AudioBuffer buffer1 = pcmBuffer->mBuffers[1];
Float32 *data0 = buffer0.mData;
@ -298,7 +291,6 @@
// NSLog(@"CODEC INIT %d COOKIE SIZE %u MAX %u", err, magicCookie->mMagicCookieSize, cookiestructsize);
outputSampleCount = 0;
@ -310,10 +302,6 @@
cookiestructsize = sizeof(CFDictionaryRef);
AudioCodecGetProperty(aacCodec, kAudioCodecPropertySettings, &cookiestructsize, &encoderState);
//NSLog(@"ENCODER STATE %@", encoderState);
return;
}

View file

@ -91,8 +91,6 @@
CSAnimationRunnerObj *runner = [CaptureController sharedAnimationObj];
NSLog(@"LOADING ANIMATIONS %@", runner);
NSDictionary *animations = [runner allAnimations];
NSMutableArray *tmpList = [NSMutableArray array];

View file

@ -183,7 +183,6 @@
NSString *pName = [[NSString alloc] initWithBytes:propName length:strlen(propName) encoding:NSUTF8StringEncoding];
id propertyValue = [self valueForKey:pName];
NSLog(@"SETTING %@ TO %@",pName, propertyValue);
[newCopy setValue:propertyValue forKey:pName];
}

View file

@ -55,7 +55,6 @@
-(void)dealloc
{
NSLog(@"DEALLOC TEXT CAPTURE BASE!");
NSFontManager *fontManager = [NSFontManager sharedFontManager];

View file

@ -477,13 +477,6 @@
NSError *compressError;
if (![self.compressSettingsController commitEditing])
{
NSLog(@"FAILED TO COMMIT EDITING FOR COMPRESS EDIT");
}
if (self.editingCompressor)
{
@ -644,7 +637,6 @@
[self.streamServicePluginViewController.view removeFromSuperview];
NSLog(@"STREAM SERVICE CONFI WINDOW %@", self.streamServiceConfWindow);
[NSApp endSheet:self.streamServiceConfWindow];
[self.streamServiceConfWindow close];
@ -1052,11 +1044,8 @@ static CVReturn displayLinkRender(CVDisplayLinkRef displayLink, const CVTimeStam
static dispatch_once_t onceToken;
dispatch_once(&onceToken, ^{
NSString *animationPluginPath = [[NSBundle mainBundle] pathForResource:@"CSAnimationRunner" ofType:@"plugin"];
NSLog(@"ANIM PATH %@", animationPluginPath);
NSBundle *animationBundle = [NSBundle bundleWithPath:animationPluginPath];
NSLog(@"ANIM BUNDLE %@", animationBundle);
Class animationClass = [animationBundle classNamed:@"CSAnimationRunnerObj"];
NSLog(@"ANIM CLASS %@", animationClass);
sharedAnimationObj = [[animationClass alloc] init];
@ -1562,15 +1551,6 @@ static CVReturn displayLinkRender(CVDisplayLinkRef displayLink, const CVTimeStam
}
/*
NSString *animationPluginPath = [[NSBundle mainBundle] pathForResource:@"CSAnimationRunner" ofType:@"plugin"];
NSLog(@"ANIM PATH %@", animationPluginPath);
NSBundle *animationBundle = [NSBundle bundleWithPath:animationPluginPath];
NSLog(@"ANIM BUNDLE %@", animationBundle);
Class animationClass = [animationBundle classNamed:@"CSAnimationRunnerObj"];
NSLog(@"ANIM CLASS %@", animationClass);
//self.animationRunner = [[animationClass alloc] init];
*/
self.extraPluginsSaveData = nil;
@ -2153,28 +2133,15 @@ static CVReturn displayLinkRender(CVDisplayLinkRef displayLink, const CVTimeStam
{
//NSLog(@"AUDIO BUFFER %@", sampleBuffer);
if (!self.captureRunning)
{
return;
}
/*
if (_firstFrameTime == 0)
{
//Don't start sending audio to the outputs until a video frame has arrived, with AVFoundation this can take 2+ seconds (!?)
//Might need to prime the capture session first...
return;
}
*/
// NSLog(@"AUDIO SAMPLE BUFFER %@", sampleBuffer);
CMTime orig_pts = CMSampleBufferGetPresentationTimeStamp(sampleBuffer);
//NSLog(@"AUDIO PTS %@", CMTimeCopyDescription(kCFAllocatorDefault, orig_pts));
if (CMTIME_COMPARE_INLINE(_firstAudioTime, ==, kCMTimeZero))
{
@ -2188,7 +2155,6 @@ static CVReturn displayLinkRender(CVDisplayLinkRef displayLink, const CVTimeStam
CMTime pts = CMTimeAdd(real_pts, adjust_pts);
//NSLog(@"AUDIO PTS %@", CMTimeCopyDescription(kCFAllocatorDefault, pts));
CMSampleBufferSetOutputPresentationTimeStamp(sampleBuffer, pts);
@ -2339,7 +2305,6 @@ static CVReturn displayLinkRender(CVDisplayLinkRef displayLink, const CVTimeStam
if (![self sleepUntil:(startTime += _frame_interval)])
{
//NSLog(@"SLEEP FAILED");
continue;
}
@ -2499,7 +2464,6 @@ static CVReturn displayLinkRender(CVDisplayLinkRef displayLink, const CVTimeStam
CFAbsoluteTime ptsTime = _frame_time - _firstFrameTime;
//NSLog(@"PTS TIME IS %f", ptsTime);
_frameCount++;
@ -2507,7 +2471,6 @@ static CVReturn displayLinkRender(CVDisplayLinkRef displayLink, const CVTimeStam
pts = CMTimeMake(ptsTime*1000000, 1000000);
//NSLog(@"PTS TIME IS %@", CMTimeCopyDescription(kCFAllocatorDefault, pts));
duration = CMTimeMake(1000, self.captureFPS*1000);

View file

@ -240,7 +240,6 @@
CVImageBufferRef imageBuffer = frameData.videoFrame;
//NSLog(@"WIDTH INPUT %zd HEIGHT %zd", CVPixelBufferGetWidth(imageBuffer), CVPixelBufferGetHeight(imageBuffer));
src_height = CVPixelBufferGetHeight(imageBuffer);
src_width = CVPixelBufferGetWidth(imageBuffer);

View file

@ -243,16 +243,13 @@ void getAudioExtradata(char *cookie, char **buffer, size_t *size)
pkt.destruct = NULL;
//NSLog(@"FFMPEG AUDIO PTS %lld/%d", pts.value, pts.timescale);
pkt.pts = av_rescale_q(pts.value, (AVRational) {1.0, pts.timescale}, _av_audio_stream->time_base);
//pkt.dts = pkt.pts;
// pkt.pts = pts.value;
if (av_interleaved_write_frame(_av_fmt_ctx, &pkt) < 0)
{
NSLog(@"AV WRITE AUDIO failed for %@", self.stream_output);
@ -396,7 +393,6 @@ void getAudioExtradata(char *cookie, char **buffer, size_t *size)
if (!(av_out_fmt->flags & AVFMT_NOFILE))
{
int av_err;
NSLog(@"Doing AVIO_OPEN");
if ((av_err = avio_open(&_av_fmt_ctx->pb, [_stream_output UTF8String], AVIO_FLAG_WRITE)) < 0)
{
NSString *open_err = [self av_error_nsstring:av_err ];
@ -497,7 +493,6 @@ void getAudioExtradata(char *cookie, char **buffer, size_t *size)
if ([self resetOutputIfNeeded])
{
NSLog(@"OUTPUT RESET");
return;
}
@ -516,7 +511,6 @@ void getAudioExtradata(char *cookie, char **buffer, size_t *size)
if (!self.active)
{
NSLog(@"NOT ACTIVE");
return;
}

View file

@ -285,20 +285,9 @@ static NSArray *_sourceTypes = nil;
[self updateRotationTransform];
//[self positionOrigin:x_pos y:y_pos];
//self.layoutPosition = self.layer.frame;
//if (width && height)
// {
// [self updateSize:width height:height];
// }
self.layoutPosition = self.layer.frame;
//self.layoutPosition = self.layer.frame;
//NSLog(@"INIT %f %f %f %f", x_pos, y_pos, width, height);
_selectedVideoType = [aDecoder decodeObjectForKey:@"selectedVideoType"];
@ -918,7 +907,6 @@ static NSArray *_sourceTypes = nil;
-(void)dealloc
{
NSLog(@"DEALLOC SOURCE INPUT");
[self deregisterVideoInput:self.videoInput];
for(id vInput in self.videoSources)
{
@ -1504,7 +1492,6 @@ static NSArray *_sourceTypes = nil;
if (isnan(x))
{
NSLog(@"HOLY CRAP NaN");
return;
}

View file

@ -131,7 +131,6 @@
-(void)setDestination:(NSString *)destination
{
NSLog(@"Destination set to %@", destination);
if ([destination hasPrefix:@"rtmp://"] || [destination hasPrefix:@"udp:"])
{
self.output_format = @"FLV";
@ -210,7 +209,6 @@
return;
}
NSLog(@"ATTACHING OUTPUT");
if (!self.ffmpeg_out)
{
newout = [[FFMpegTask alloc] init];
@ -234,7 +232,6 @@
newout.samplerate = self.settingsController.audioSamplerate;
newout.audio_bitrate = self.settingsController.audioBitrate;
NSLog(@"SETTINGS %@, FPS %f", self.settingsController, self.settingsController.captureFPS);
self.ffmpeg_out = newout;

View file

@ -1244,7 +1244,6 @@
{
//Without the autorelease NSColor leaks objects
NSLog(@"Preview: Creating Pixel Buffer Pool %f x %f", size.width, size.height);
NSMutableDictionary *attributes = [NSMutableDictionary dictionary];
[attributes setValue:[NSNumber numberWithInt:size.width] forKey:(NSString *)kCVPixelBufferWidthKey];

View file

@ -79,7 +79,6 @@
-(BOOL)application:(NSApplication *)sender delegateHandlesKey:(NSString *)key
{
NSLog(@"HANDLES KEY? %@", key);
NSArray *keys = @[@"layouts", @"width", @"height", @"fps", @"activelayout"];

View file

@ -57,12 +57,10 @@
NSString *sourceKey = [NSString stringWithFormat:@"%@:%@", ofType, uniqueID];
NSLog(@"SOURCE KEY IS %@", sourceKey);
id cachedSource = [self.cacheMap objectForKey:sourceKey];
NSLog(@"CACHED SOURCE %@", cachedSource);
if (!cachedSource)
{

View file

@ -104,7 +104,6 @@
-(void)doAnimation:(NSDictionary *)threadDict
{
//[CATransaction flush];
CSAnimationRunnerObj *runner = [CaptureController sharedAnimationObj];
@ -113,24 +112,6 @@
CALayer *rootLayer = threadDict[@"rootLayer"];
/*
InputSource *src = inpMap[@"spinnah"];
CALayer *layer = src.layer;
NSLog(@"LAYER IS %@", layer);
[CATransaction begin];
[CATransaction setAnimationDuration:3.5];
layer.zPosition = -1;
[CATransaction begin];
[CATransaction setAnimationDuration:5.5];
[layer setValue:@500.0f forKeyPath:@"bounds.size.width"];
[CATransaction commit];
[CATransaction commit];
*/
@try {
[runner runAnimation:modName forInput:inpMap withSuperlayer:rootLayer];
}

View file

@ -18,7 +18,6 @@ int main(int argc, const char * argv[])
@autoreleasepool {
// insert code here...
NSLog(@"Hello, World!");

View file

@ -48,7 +48,6 @@
- (void)doInject:(NSRunningApplication *)toInject
{
NSLog(@"WILL INJECT INTO APPLICATION %s", [toInject.localizedName UTF8String]);
pid_t pid = toInject.processIdentifier;

View file

@ -43,7 +43,6 @@ XPCListenerDelegate *captureDelegate;
- (void) dealloc
{
NSLog(@"HOLY SHIT DEALLOC");
if (self.xpcProxy)
{
[self.xpcProxy release];
@ -58,7 +57,6 @@ XPCListenerDelegate *captureDelegate;
{
NSLog(@"SHOULD ACCEPT NEW CONNECTION\n");
NSXPCInterface *helperInterface = [[NSXPCInterface interfaceWithProtocol:@protocol(QTHelperProtocol)] retain];
NSXPCInterface *masterInterface = [[NSXPCInterface interfaceWithProtocol:@protocol(CapturedFrameProtocol)] retain];
@ -82,7 +80,6 @@ XPCListenerDelegate *captureDelegate;
{
NSArray *devices = [QTCaptureDevice inputDevicesWithMediaType:QTMediaTypeVideo];
NSLog(@"DEVICES IN HELPER %@", devices);
NSMutableArray *retArray = [[NSMutableArray alloc] init];
QTCaptureDevice *devinstance;
@ -279,7 +276,6 @@ void qt_xpc_peer_event_handler(xpc_connection_t conn, xpc_object_t event)
} else {
const char *message = xpc_dictionary_get_string(event, "message");
NSLog(@"RECEIVED XPC MESSAGE %s", message);
if (!strcmp(message, "list_devices"))
{

View file

@ -16,21 +16,5 @@ int main(int argc, const char *argv[])
xpc_main(qt_xpc_handle_connection);
/*
XPCListenerDelegate *listenDelegate = [[XPCListenerDelegate new] retain];
NSLog(@"LISTEN DELEGATE %@", listenDelegate);
NSXPCListener *listener = [[NSXPCListener serviceListener] retain];
//NSString *bundleId = [[NSBundle mainBundle] bundleIdentifier];
//NSXPCListener *listener = [[NSXPCListener alloc] initWithMachServiceName:bundleId];
NSLog(@"LISTENER %@", listener);
listener.delegate = listenDelegate;
[listener resume];
NSLog(@"SETUP XPC LISTENER");
exit(EXIT_FAILURE);
*/
}

View file

@ -40,7 +40,6 @@
{
NSLog(@"SELECTED SERVER %@ STREAMKEY %@", self.selectedServer, self.streamKey);
if (self.selectedServer)
{