Hi there. I've been using your code to learn Objective-C and so far so good. Wonderful work. The thing is I'm also using the GPUImage framework and each time I take a picture with the Amatorka filter on live stream I get a white image. The preview is fine, I mean, I take the picture and it shows up great on the App, I click done to save to camera roll and I get a white image. When I pick an image from the camera roll and apply the filter it works great, but not on live stream. All the other filters work great on live stream except the LTU filters (LookupFilter). Did I do something wrong, or is it just not possible to apply LTU filters to live stream. Thanks in advance.
Your question concerns GPUImage, not this project.
I don't think this is a good example of where you should learn Objective-C.
I suggest you read the Big Nerd Ranch Guide to iOS and do the problems http://www.bignerdranch.com/book/ios_programming_the_big_nerd_ranch_guide_rd_edition_
Thanks dmitric. I was able to solve the problem. Everything is now fantastic. And thanks for the tip. I'll check it out. I have C++ background, so I'm just exploring to see the main differences between the two languages.
@dpestana ah okay thats fair. @alexperri7 asked for help regarding this so maybe you could hook him up too?
Sure, no problem. I'll be glad to help. It's really quite simple, just create a filter group and add the Lookup filter to an RGB filter. It should work. If you get memory issues, just make your photocapturebutton your FirstResponder under viewDidAppear.
I forgot to say, if you choose not to use a filter group, all your images will be white after capture on live stream.