Sunday, 27 March 2011

The end

With the project nearly done, I was given the job of putting together the end credits.  Initially, some members of the group suggested that I just do a "Created by..." credit and leave it at that, but I thought it would be good if I listed each of our jobs/contributions within the project.  I started by typing out the credits in TextEdit, while confirming with the rest of my group who did which job.
Im Adobe Premier, I started copying the credits into title stills, which we decided would appear in the Helvetica font.
When we were filming the green screen sequence, we did quite a lot of filming of us just mucking around, so we decided that this would look good alongside the credits.  I went through this footage in Premier and selected a few clips to go against the credits and then exported that to another sequence.
Next, I imported this video into the credit sequence and I then scaled the video down and placed it right next to the credits.  I made the length of the credits slightly longer, so that the video would fit around them and keyed it so the video fades towards the end of the credits.  This is an example of very basic compositing to get our credits done.  Here is the outcome.

Saturday, 26 March 2011

Compositing the creature: Shot 19

The second shot that I was required to composite was the 19th shot of the piece; this largely involved applying shadows to the scene, so that the creature would blend in with the various shades.  Here is the original live action footage.
In order to integrate this creature into this footage, I opened the 3D blockout that was lighted by Matt and had the creature animated by Will.
The idea was to render the creature's running separately to the shadows.  How I did this was by selecting the lambert textured surfaces that made up the wall and floor, and then assigning them to a separate rendering layer.
This layer was called layer1.  I turned this layer off and rendered only the masterLayer that included the creature.  I rendered the creature's run in mental ray as we had agreed as a group, here is a still from the batch render.
Next, I turned off the masterLayer and turned on Layer 1.  I assigned a material to the surfaces called 'Use Background', which meant that they would not be visible, but that they would still pick up the shadows of the creature and the wall.  I rendered this using Maya Software, because mental ray was not capable of rendering the shadows out properly.  Here is a shot of the rendered shadows.
The next stage was to import these two into After Effects against the live action footage.
I scaled down the rendered sequences of both the creature and shadows so that the shadow is aligned with the building to the right and so the creature can run through the scene.  I made sure both the shadow layer and creature layer were positioned and scaled identically.  

The rendered images were initially too small within the live action, meaning that the creature just randomly appeared on screen rather than creating the illusion of the creature running on screen.  In order to fix this, I just scaled up the live action footage and CG footage so that the creature seemed to run on screen and into the shade of the council estate.

Next, I needed to amend the shadows; they were initially too solid to be considered shadows, so I applied a Gaussian Blur to make them more natural.  Finally, I discovered that the shadow of the creature was coming out white, so to amend this, I applied a Set Channel and changed all the settings to Alpha so that it was a darker grey colour.
Here is the final, fully composited shot of the chase.

Compositing the creature: Shot 6

In the post production process, we needed to composite the creature into each shot.  This also involved the use of lighting and shadow to blend the creature in with the rest of the scene.  We were each given some shots to composite, based on the blockouts of the scene and the lighting that had been set up by Matt.

The shots that I was given were, shot 6; which is where the creature emerges from the manhole, and shot 19; in which the creature chases Richard around a corner.

The first that I will talk through is shot 6.  Here is the original video.
And here is a screenshot of the animation in Maya that I had to integrate into this scene.
I rendered the Maya Animation out in mental ray and I then imported it into After Effects with the live action footage.
The thing about this shot is that I needed to make it look as if this creature was emerging from a hole, as opposed to off camera.  I needed to create a hole for this creature to emerge from.  In order to create this hole, I simply created a mask in the live action footage in the area of the manhole.
I then feathered the mask so the hole wasn't too rigid.  At the moment, the hole just looks black and doesn't have any depth.  In order to fix this, I looked for an image of inside the manhole to put under the masked layer, so that it looked like it led to a sewer or somewhere.  I came across this image.
I put this image under the live action footage and around the mask.  I then applied a colour correction in order for this image to fit in with the scene and here is how it looked.
Here is the final composited shot.

Thursday, 24 March 2011

Some action figure close-ups

Towards the end of the sequence, the parkour runner and creature morph into action figures and in a surprise twist, the scenario is just a product of our tutor, Jared's imagination.  It was my job to key out the green screen in the close-up shots of both action figures and apply a rendered CGI background.
Here are each of the clips, which have also had the brightness adjusted so they blend in with the environment.

This last shot had a lot of noise in it, so I had to apply a mask, which I then key framed in order to remove some of it.  Even then, there was still a bit of noise left in areas that were too close around the action figure to key out adequately.  I have found that noise tends to come up on green screen significantly more than blue screen.

The shaking manhole

Along with the green screen shots, I was given some shots to render and then composite.  The first being the close up shot of the manhole, which suddenly begins to shake.  Not only was I required to render this manhole and composite it with live action, but I was required to animate the shaking.

My team members stated that the manhole needed to be shake in a subtle but menacing movement.  Here is what I came up with.
Next, I rendered this in the mental ray format and I then masked out the lambert texture around the manhole.
The thing about this shot is that the dark shades on the manhole do not match the brightness of the ground surrounding.  I amended this by applying a Curves filter to the live action footage so that I could alter the brightness and make it darker.
In the outcome, the footage looked overall much darker and this suited the mood that we were trying to convey in this shot.

Thursday, 17 March 2011

Richard drops

Another shot that I was given to composite was a shot of Richard dropping from a massive jump down a flight of stairs.  Producing this composition involved layering a green screen shot of Richard jumping from a chair, with a CG shot of the staircase and floor that he eventually lands on.

Here is the green screen footage and a rendered close up of the staircase.

There is some very recognisable camera movement in the video, particularly when he lands.  I decided that I would motion track it in Match Mover.  Here is a shot of the sequence in Match Mover.
Match Mover seemed to track the camera movement fairly efficiently.  However, when I exported the motion tracking to a camera in Maya.  I found that the camera just remained static; the camera movements were not significant enough to be applied to the camera.  Instead, I just went straight into After Effects and composited the live action and CGI together.  I also applied masks around the tape on the green screen, to make them invisible.

Motion tracking Richard's run

The following shot that I will be talking about involves our runner, Richard running away before being cornered against a graffiti wall.  In order to make this shot believable, I used motion tracking, as well as masking and green screen.  Here is the original footage that we recorded in one of the studios at uni.
The green screen is where I will place the CG environment.  The pieces of sellotape that we have place on the green screen are what we will use to track the movement and speed of the camera when we key out the green screen.  They will also be used by Match Mover, a program that will track and reproduce the camera movements, so I can export them to a camera in Maya, meaning that the cameras in both the live action and CGI will move in sync with each other.

The next stage was to load the sequence in Match Mover.  In this program, I applied Automatic Tracking, which tracks the camera movement in each shot.  After this tracking was applied, I used the clean assistant to reduce the number of tracks, so that the camera movement was not too erratic.  Here is how the scene looks within Match Mover after an Automatic Track.
The green lines in this shot display the movements that have been tracked.  Not only does the Automatic Track capture the camera movement, but also Richard's movements.  At first, I didn't think this was a problem, but when I exported the camera movements into Maya, I found that the camera moved really erratically and even moved out of the scene.  I could not figure out why the camera was moving like this.

After showing this problem to my tutor, he advised that I delete all of the trackers the are near or on Richard's figure, because the camera was tracing his movements, as well as the movements of the camera.  I went through the whole sequence within Match Mover and deleted each tracker that appeared on Richard.  Here is the same shot, after I had gone through the sequence.
I exported these camera movements to Maya.  I then applied them to the relevant scene, where I made a few manual arrangements, such as deleting a few unnecessary key frames and making the movement smoother in the Graph Editor.  Here is an unrendered Playblast of the camera movement in the chase scene.
I now needed to composite this with Richard's running sequence, and the desired outcome would be that Richard's run would move in sync with the camera movements in Maya.  In After Effects, I uploaded both the camera sequence in Maya, fully rendered, and the green screen footage of Richard running.  I overlaid the live action footage over the green screen footage and then I keyed out the green screen, with a Keylight filter.
The white of the walls around the green screen was still present.  I didn't key out the white, because keying out any colour, other than green or blue would have caused problems, as it would start keying out things like Richard's skin tone and clothing.  In order to get rid of the white walls, I place Masks around them, which I then keyframed as they moved around.
In order for Richard to look as though he is running towards the wall, I keyfamed the position and scale of the video footage, to create the illusion of Richard running away from the camera.  As a finishing touch, I applied a Brightness and Contrast filter, which I then keyframed to make Richard's figure become darker as he ran into the darkness, which would highlight how he is running into peril.