CM30075: Computer Graphics maspmh/ Peter Hall pmh@cs.bath.ac.uk

Preview:

Citation preview

CM30075: Computer Graphicswww.bath.ac.uk/~maspmh/

Peter Hall

pmh@cs.bath.ac.uk

!!

Warning!!

These slides do not replace text book reading

L01: about this course

Aim: to teach the elements of Computer GraphicsTraditional 3D photorealismModern use of images, and non-photorealism

Method: traditional formslecturespersonal readingpersonal practical work

Assumptions: knowledge of analytic mathematicsvectors and matricesintegration and differentiation

Lecture number / curriculum index Topic Learning objective

01 / 1 Introduction to CM30075And Compuer Graphics

Pointers to course materials given., Traditional Graphics outlined briefly.

Traditional Computer Graphics

02 / 1.1.1 Modelling – Brep The B-rep modelling scheme is introduced as just one way to build objects.

03 / 1.3.1 Rendering – cameras and projection Camera models introduced. Points are projected, so that objects models can be rendered as wire-frames.

04 / A.1 Rendering – ray-casting Lines intersected with planar polygons, result used as a basis for ray-casting. Now objects can be rendered to look solid.

05 / 1.2.1 Simple reflection models Point light sources are introduced; Lambertian and Specular reflection so objects can be shaded.

06 / 1.3.2 Rendering – ray tracing Cast shadows, reflection and refraction amongst many objects. Ray-tracing as a tree.

Milestone: material for advance practical work is now covered

07 / 1.2.2 Advance lighting and reflection The BRDF is introduced. Physical models of reflection and refraction.

08 / 1.2.3 The Lighting Equation;Radiosity

The full complexity of lighting is revealed; radiosity and ray-tracing are seen as specific solutions.

09/ 1.1.1 Modelling revisited – advanced B-rep Use of spline surfaces as B-rep models; how to define them and how to render them

10/ 1.1.2 Modelling revisited – CSG and Voxels Voxel models and Volume Rendering

11 / 1.1.3 Texture maps. Texture maps introduced as things to be rendered that relieve the burden of modelling. Aliasing issue is mentioned.

Milestone: static images now (mostly) covered

12 / 1.4.1 Animation basics The basics of animation are introduced from both technical and artistic points of view.

13 / 1.4.2 Articulated figures Modelling bodies and hierarchies of transforms; inverse kinematics/dynamics is discussed

14 / 1.4.3 Soft Objects and Fluids Elastic properties – use of physics – to model cloth, putty and such like, including fluids.

15 / 1.4.5 Fire and Fluids Use of dynamic textures and particles to model animate objects with no definitive boundary.

Milestone: Traditional Computer Graphics material covered

Modern Computer Graphics: Photograph editing and non-photorealistic rendering

16 / 2.1.1 Compositing How images layer to make a whole; simple methods for making panoramas from holiday snaps.

17 / 2.1.2 Panoramic Views How to make panoramas from holiday snaps – and from video too.

18 / 2.1.3 Intelligent scissors Cutting and pasting objects in pictures; so more on compositing

19 / 2.1.4 Texture Propagation Filling in holes in images, maybe left by cutting.

20 / 2.2.1 NPR from 3D models Why is photorealism the aim? People paint! How to paint, not photograph from models.

21 / 2.2.2 NPR from photos How can photographs be processed to look like a painting? Why is this difficult?

22 / 2.2.3 Generalised Cameras The eye of Art is often not a real camera!

Milestone: course content fully introduced

23 revision

24 revision

Reading

Lecture Slides• are NOT intended to replace text book reading.

Standard texts• Watt, 3D Computer Graphics, Addison Wesley.• Foley et al, 3D Computer Graphics, Addison Wesley.• Watt and Watt, Advanced Computer Graphics, Addison Wesley.

For the interested.• SIGGRAPH proceedings (published as journal special issue; Transactions on Graphics)

• Eurographics proceedings• IEEE Transactions on Visualization and Computer Graphics• Computer Graphics Forum

PracticalAim: You are to write a simple ray-casting / ray-tracing program.

The practical work is broken into three stages

• ray-caster with self-shading

• cast-shadows

• full ray-tracer.

You should spend no more than 15 hours on this component;

including time to prepare the documents needed for your assessment.

Assessment

Assessment will cover all material given in lectures, in assigned reading and in practical work.

75% Sat ExaminationQuestions usually comprise 4 parts:a) Basic knowledge (3rd class)b) Moderate knowledge, basic understanding (2.2nd class)c) Good knowledge / moderate understanding (2.2nd class)d) Good understanding – can solve new problems (1st class)

25% practical workAlready explained

Computer Graphics Basics

focus

window

objectcamera

point (pixel)

Traditional: What colour is this pixel?

Rendering Pipeline

Make 3D models

transform models into place

illuminate models

project models

clip invisible parts

raterization

display

The rendering pipeline shows

the flow of information uses

and the processes needed

to synthesise an image.

In fact, there are many rendering pipelines.

The order of processes can change depending, for example, on whether rendering time or rendering quality is more important.

These different pipelines differ only in details – the general flow of information from 3D model to 2D image is always the same.

Modern approaches

point (pixel)

target image source image

What colour is this pixel?

• The B-rep modelling scheme is introduced as just one way to build objects.

L02: B-rep basics

points points and lines points, lines, and polygons

B-rep basics

Brep = Boundary Representation

Start with a set of M points p = { (x, y, z)i : i = 1…M }

Make a set of N lines from points L = { (i, j)k : k = 1…N }

Make a set of m polygons from lines B = { (k1, k2, …, kn}l : l = 1…m }

A model is three-dimensional (3D),

if the points are 3D (as above).

Different informationsupports different rendering

• points onlydots (used in Chemistry, also in modern Point based Rendering)

• linesWire-frame rendering, good for quick tests in animation, say

• polygonsshaded surfaces

points points and lines points, lines, and polygons

x1 y1 z1

x2 y2 z2

xi yi zi

xM yM zM

2 i

2 k

1 3 4 6 8

2 1 7 * *

point table line table polygon table

!! ALWAYS INDEX POINTS !!

Practical Issues

• avoids repeated points, so more efficient

• avoids numerical error when animating

Different tables can be used

x1 y1 z1

x2 y2 z2

xi yi zi

xM yM zM

1 3 4

3 4 7

point table triangle table

Triangles are the most common polygon, because triangles are always flat.

But triangles are expensive – many of them; so other polygons used for modelling are often decomposed into triangles for rendering.

A Minor Complication!! POINT ORDER MATTERS !!

anti-clockwiseclockwise

The ordering gives the polygon a “front” and “back”.

eh! which side is which?

The normal of a triangular polygon

The normal of a polygon is used in lighting calculations (and in other calculations too).

Suppose a triangle has points, p, q, r, each point in 3D.

The normal direction is

n = (p-q) x (q-r)

where x is the vector cross product.

Exercise: Show that reversing the order of points also reverses the direction of the normal.

build a table of trianglesfrom these points

1

2

3

4 5

6

7

8

Exercise: The points are randomly numbered.

From the view given,point 4 is at the back of the cube,point 5 is the nearest corner.

Build a table of trianglesin which vertices are consistentlyordered clockwise,when each face of the cube is viewedfrom the outside.

B-rep basics: summary

• B-rep = boundary representation

• Models objects with points, lines, polygons

• Points are ordered around a polygon

• Best to index into tables

• The form of tables dictates rendering algorithms

Exercise: Build a 3D cube from well ordered triangles, render it as a wire-frame. See L03 for projection methods.

• Camera models introduced.

• Points are projected, so that objects models can be rendered as wire-frames.

L03: Cameras and Projection

The Linear CameraIn a linear camera

• rays of light travel in straight lines from a object

• the camera captures all rays passing through a single focus

• intersect a planar window to make the image

• the normal from the plane to the focus is the optical axis

focus

windowobject

image

ray

optical axis

The Linear Camera

focuswindow

object

image

Two variants exist:

• the “physical” model – shown in the previous slidehas the focus between the object and the window,in which the image is inverted

• the “mathematical model” – which we use, is shown belowhas the window between the object and focusleaving the image the right way up

Basic Perspective Projection

h

W

wHh

w

Uses similar triangles to compute the height of the image

object

image

focus

W

H

3D is almost as easy as 2D

),(),( yxz

fvu

v xu

y

zf

),,( zyx

),,( fvu

In the canonical camera, focal length (f) is taken as 1.

Homogeneous Points

Make projection easy and convenient

3D point (x, y, z) written as (x,y,z,1)homogeneous point (x, y, z, ) maps to real point 1/ (x, y, z)

The homogeneous points p = (x,y,z,1) and q = (sx, sy, sz, s)differ only by a scale factor s; this makes them equivalent in homogenous space.

Notice (x,y,z) and (sx,sy,sz) are two points on the same straight line passing through the origin. This makes them equivalent – they represent the same line!

In homogeneous space, lines and points are dual concepts!

object point

image point

straight line (a ray of light)

focus:at the origin

a particular line – the optical axis – passesnormally through the window and through thefocus. This line has the image of the focus –a vanishing point.

the set of all rays projectnormally onto the windowto make a pattern of “spokes”

A little homogeneous geometry

window

a scaled version of the image

a scaled version of the object

Projection with a matrix

Using homogeneous coordinates, projection can be written as a matrix

we do need to divide by homogeneous depth, , after this

Compare this to with f = 1.),(),( yxz

fvu

The camera as a matrixUsing a matrix for projection is very convenient in many ways.

1. It means we can model the camera as a matrix, C, say.Now projection of a homogenous point p is just

q = pCand we know that the homogeneous image point q is just a scale factor away from being correct – and all we need do is scale it by its depth (last element).

2. It means we can move the camera about in space just by pre-multiplying by a matrix transform

q = pMC

3. It means we can change the internals of the camera (focal length, aspect ratio, etc) by post-multiplying by a matrix

q = pMCK

4. We can do all of this at once! just set A = MCK, now A is a linear camera q = pA

Some Examples

Setting

using

rotates then translates the camera before projection.

Notice this is equivalent to applying M-1 to the point.

Another exampleTo change focal length, set

now post-multiply

All at once

Define a new linear projective camera in a new place and with a new focal length

It is easy to use this camera…

Of course, you can set M and K as you please

– not just rotate and translate and new focal length!

Exercise: See notes for an exercise

Rendering with a Projection Matrix

1. Project all points using the projection matrix

2. Keep all points that lie within the window bounds

3. Connect points in the picture that are connected in 3D

If we had a model built just of points (no lines, planes etc) then we could make a simple images using this simple technique:

This produces a “wire frame” picture – easy, and fast!

If all pixels inside a projected triangle can be identified,then they too can be coloured.

This is the basis of scan-conversion.

• Lines are intersected with planar polygons.

• The result used as a basis for ray-casting.

• Now objects can be rendered to look solid.

L04: Ray-Casting

Ray-casting algorithm

For each pixelCast a ray from the focus through the pixelCompute all intersections with all polygonsFind the nearest polygonColour pixel with polygon colour

Actually, the polygon colour is modified to create the effect of shading, as on a sphere.

Ray Casting Basics

• One ray per pixel, cast into the scene• Look for nearest intersection• Colour pixel accordingly

Expensive part: computing the intersection of a ray with a polygon

focus

pixel

window

scene of objects

camera

Line / Polygon intersection

p r

x(s)

c

n

an infinitely wide plane

an infinitely long line

compute scale factor

Given the scale factor,

the intersection is easy to get

Point Inside Triangle?inside

outside

all three “turns” are anti-clockwise some “turns” are anti-clockwise,others are clockwise

direction of turn:

p1

p2

p3

x

parallel direction of turn, sij > 0:

A practical problem

The frame buffer is in the computer.

We imagine casting rays thru’ pixel centres

i

j

The window is in “space”;

where are pixel centres in space?

Solution: recall the camera as a matrix, C = MPK

y = xC = ((xM)P)K

• external parameters M locates the camera in spaceM is a 4x4 matrix, (eg rotation)notice M maps a model point in spaceuse the inverse of M to move camera!

• a projection P maps 3D points, to the windowP is rank degeneratecan be (4x3) but usually is (4x4)

• internal parameters, maps window points to frame-buffertransform confined to the planea matrix K (a 3x3 will do!)

K

x

y

i

j

The internal parameters transform the pixels centres from “frame buffer” coordinates to “window coordinates”.

There is no single “answer” – here I’ve mapped a rectangular frame buffer to a window which is square, so pixels get stretched into rectangles – but you may keep pixels square.

pixel is typically at (j,i) point is at (x,y)

It is very common (almost universal) to “flip” coordinates so pixel (j,i) maps to location (x,y);and notices row (j) increase going DOWN - take care when designing K!

K-1

The external parameters transforms the canonical camera

in “camera space” to the desired camera in “world space”.

M-1

M

Ray-casting takes a long time

Why does ray-casting take so long?

Because every ray is compared to every polygon.

Is this necessary?

What are bounding spheres?

Homework question: what is a BSP tree?

• Point light sources are introduced.

• Lambertian and Specular reflection, so objects can be shaded.

L05: Simple Reflection

Point light sources

• A point light source is, er, a point (x,y,z) that emits light.

• Actually the point light can be at infinity (think of the sun), in which case (x,y,z) is its direction.

Most point lights sources are often at infinity in computer graphics.This makes shading/reflection efficient to compute because the direction of the light is the same for every intersection point.

Basic Reflection Model

mirror direction

light direction

surface normal

Mirror reflection - specular reflection – Phong reflectionDiffuse refection – Lambertian reflection

Diffuse/Lambertian ReflectionA beam of light spreads over an area, and is reflected equally in all directions.So here we’ve not drawn any reflected direction

The energy in the light is spread out more over the wider area.

So, the area will appear less bright.

over anarrow area over a wide area

these beamsare of equal width

Functional analogy: the sun warms the earth more near the equator than the pole

w

h

h = w/cos()

w

The light energy in the beam of width w is spread over a length h = w/cos().

Energy density in beam: I/w

Energy density on ground: I/h = Icos()/w

Lambertian reflection in Graphics

Iout = Ilightcos()The cosine we need is the dot (inner) product of the unit normal andunit vector pointing to the light.

We know that light from the source strikes a surface, to be spread equallyin all directions. So, only the angle between the light direction and thesurface is important.

Iout = Ilight n.l

BUT the surface can absorb some light, so only a fraction c is reflected

Iout = cIlight n.l

The “diffuse reflection coefficient” c depends on the surface material.

Diffuse reflection in colour

Rout = credRlight n.l

To handle colour is easy: just do the same calculation for each colour channel

Bout = cblueBlight n.l

Gout = cgreenGlight n.l

For white light, Rlight = Glight = Blight.

The coefficients c are often thought of as the colour of the material.

Specular Reflection

mirror direction, r

light direction, l

surface normal, n

eye direction, e

How much light is reflected in a particular direction?

Intuitively, the angle between the mirror and eye directions is important.Homework: The eye direction is a given, what is the mirror direction?

Phong Reflection

Phong’s specular reflection model

Iout = kIlight(r.e)n

Notice how the cosine (r.e) is used.

k is the “specular reflection coeffiecient” for the surface.

Phong reflection works for colourin just the same sort of way as Lambertian reflection

A complete but simplereflection model

Iout = aIamb+ cIlight n.l + kIlight(r.e)n

A complete model (for a single colour channel) is

We see the diffuse and specular terms are added up.

There is an additional term for “ambient” light - this is light that comes equally from every direction- ambient light allows us see the very darkest regions of a picture

Specular highlights often look white, (just look around at things to see this) so k is often given the same value in every colour channel

Ray-tracing is a global lighting method.

• Cast shadows

• Reflection and refraction amongst many objects.

• Ray-tracing as a tree.

L06: Ray-tracing

Cast Shadows

Is the box resting on the table ?

Is a point in shadow?

point light

lightpoint

A point is in shadow if the line between it and a light is blocked by an object

How can we tell?

The point to light ray is a line

Look for intersections of the this line with any object

This is almost an exact repetition of the ray-casting,

except intersections need not be ordered,

just finding one is enough

Reversibility of light

Ray tracing relies on a physical principle:a ray of light can be traced in either direction

So, if a ray of light splits into two parts then its energy is divided.

But we can “run this backwards” and add up the energy in the divided rays to get the energy in the original.

The paths of the light rays are identical, backwards or forwards.

Ray tracing extends ray casting

eye window

rayobject

reflected ray

intersection

refracted ray

Ray tracing tree

A ray/object intersection generate a new (reflected, refracted) pair of rays.These new rays also intersect to generate new rays, hence a tree

original ray from eye thru’ pixel

intersection

reflected refracted

intersection

reflected refracted

intersection

reflected refracted

Stop when a ray “leaves” the scene, or a set depth

intersection

a leaf

the root

a “parent” raywhen goingUP the tree

Using the tree in a simple wayThe ray-tracing tree is used “bottom up”, from the leaves to the root.

There are no reflected or refracted rays at a leaf.You can work out the light to the parent using the simple lighting model.But now the direction to the eye is in fact the direction of the parent ray.

The leaf contributes some energy to its parent,this parent receives light contributions from both its children.

Now work out the simple light model for the parent, as if it was a leaf,Add this to the contributions from its children.

And so on.

A detail of the tree

surface

normalto light

mirrordirection

to parent

leaf intersectionlocal light only

surface

normal

to light mirrordirection

to parent

leaf intersectionlocal light only

surface

to parent

non-leaf intersectionlocal light + child rays

Whitted ray-tracingTurner Whitted produced the first ray-racer (1980).It was the first global illumination model. It includes

• Hidden surface removal• Self shading (as in diffuse/specular reflection)• Cast shadows• Reflection and refraction (making the model “global”)

The full coursework is to write a Whitted ray-tracer.

A restriction is to not include reflection and refraction, which is “advanced” ray casting.

A further restriction include self-shading only, which is “simple” ray-casting, and the minimum required to make a picture of some kind.

• The BRDF is introduced.

• Physical models of reflection and refraction are considered.

L07: Advanced Reflectionand Refraction

Overall energy transfer

incident light Ein

reflected light, Erefl

refracted light, Erefr

schematically

Ein = Erefl + Erefr + Eabsconservation of energy

Recall the simple lighting model

c = kaIa + Ikd(n.l) + Iks(r.e)

This takes into account• ambient light• diffuse reflection• specular reflection

n

l re

The energy in the colour depends on where the viewer is, e,compared to the reflection r – and hence the light direction l.

A visualisation

Suppose we set Ia = 0, I = 1, kd = 1, ks=1, and n.l = 1Furher, fix r. Now the lighting depends only on e and .In fact

c = 1 + e.r

to get a picture of how the energy depends on point of view

A more general idea

The light energy depends on

• the direction of the incident light

• the direction of view

Diffuse/Specular lighting

diffuse to diffuse diffuse to specular

specular to diffuse specular to specular

The BDRF

BDRF = bi-directional reflectance function

f(

f the fraction of light energy transferredfrom an incident light ray at (1,2)to viewing direction (1, 2)

In the simple model this is (e.r) in which r depends on input direction l.(Recall computing r from l is a homework!)

A full BDRF is even more complicated – wavelength, polarization…!

The importance of the BDRF

The BDRF controls the kind of material an object appears to be made from, for example

• plastic

• metal

• ceramic

The simple model tends to make things look plastic.

Where do we get a BRDF from?

BRDF can be measured from the real thing; not easy!

In graphics, “micro-surfaces” can be used to estimate BRDF.

Look up Torrance Sparrow model.

see, eg www.cs.princeton.edu/~smr/cs348c-97/surveypaper.html

flat surface …under a microscope

a perfect mirroreach micro-facet a perfect mirrorbut overall surface is not!

• The full complexity of global lighting

• radiosity in brief

• radiosity and ray-tracing as specific solutions

L08: The rendering equation

See: Watt & Watt, Advanced Animation and Rendering, subsec 12.2

Global LightingEvery object reflects light

(otherwise, it would look like a black hole)

So, every object can be reflected in every other !

This makes global lighting very complicated.

what colour is the light coming out?

light in

light out

Scattering, Shadows and refraction complicate further still !

light in

light out

and real cases are still more complicated !

Kajya’s rendering equation

The light transported from point y to point x

g(x,y) is the “visibility” function, 0 if x is in shadow wrt y, 1/|y-x|2 otherwise.

(x,y) is the transfer directly from y to x

(x,y,z) is BDRF (scattered light) toward x by y given light source at z

S is the set of all points in the scene

Important: I(.,.) appears both sides.

Use of Rendering Equation

Different lighting models are special-case solutions:

The local-model

Ray-tracing

Radiosity

rewrite

as with R an integral operator

Now

hence

Solving the rendering equation

consider

first term – direct (local) lighting: x is the eye, y a point

subsequent terms account for scattering from other points

Ray-tracing

• Forward ray-tracing (from the eye to lights)

– specular to specular : bounces, Phong term– specular to diffuse: Lambertian– diffuse to diffuse (but badly!): Ambient

• Backward ray-tracing (to the eye from lights)

– diffuse to specular

Radiosity

• Diffuse to Diffuse

This is the (badly modelled) “ambient” term in “local” models

light falls onto a patch from all others

light radiates to all other patches

The light energy per unit area is called radiosity

total energy at a patch is its radiosity x its area

and energy is conserved, so at a patch

radiosity x area = emitted energy + reflected energy

In a closed environment the energy transfer betweenpatches will reach an equilibrium.

If we use a discrete environment (ie a finite model),then we can use a discrete form of the radiosity equation

The form factors are correlated:

so that, on division by Ai we get the basic equation used:

the radiosity equation in matrix form

the hard part is computing the form factors.

(See a standard text for how to do this)

Once they are at hand, solve the system for the B.

then render with a scan converter.

• Spline surfaces as B-rep models

• How to define them

• How to render them

L09: Brep with curves

Many real world objects are curved.

But so far, we have used flat modelling primitives.

Here, we learn how to use curved primitives.

We first look at curve lines in space,

and then at curved surfaces in space.

Both curves and surfaces are collections of points in space.

The points are related to one another by some function or other.

In principle the curves and surfaces contain an infinite number of points,but in practice we are forced to use finitely many points.

And we can think of surfaces as a collection of curves

There are many ways to define curves and surfaces

Implicit: “balances” a points coordinates

Unit Circle… x2 + y2 = 1

Unit sphere: x2 + y2 + z2 = 1

We will use implicit forms in CSG modelling.

Here we will use parametric forms.

Parametric forms allow you to compute points directly.

This is much better for Brep models.

Parametric forms allow you to compute points directly.

This is much better for Brep models.

Parametric curves require one parameter, u.

x(u)

is a point in 3D, on the curve.

Parametric surfaces require two parametersx(u,v)

is a point in 3D, on the surface.

Formally

x(u) is a mapping from the real line, R, to a subset of R3.

It’s as if the real line (x axis) is bent into the shape of the curve.

x(u,v) is a mapping from the real plane, R2, to a subset of R3.It’s as if the real plane (xy-plane) is bent into the shape of the curve.

u

u

v

x(u)

x(u,v)

Curves:

An easy way to specify a curve is with a polynomial;most modellers use cubics:

x(u) = a + bu + cu2 + du3

= [u3 u2 u 1][d c

b a]

The coefficients a, b, c, d specify the curve.

But this is not the most convenient way,because it’s hard to control.So the cubic is usually specified in some other way.

This is an abuse of notation!

Bezier Curves and Surfaces

are just one of the many alternatives

x(u) = UMP

U = [u3 u2 u 1]

M = [ -1 3 -3 1 3 -6 3 0 -3 3 0 0 1 0 0 0 ]

P = [p0

p1

p2

p3]

the “Bezier” matrix

parameter vector

4 control points

Examples of Bezier curves

control points

Bezier curve

Bezier surfaces

analogous issues over control of surfaces argue in favour of,

say, Bezier surfaces; again one of may controllable forms.

x(u,v) = UMPMtVt

The M is the Bezier matrix,

U and V parameter vectors,

P is now a 4x4 matrix of control points.

A Bezier surface example

Rendering curved surfaces

Can compute intersection of ray/patchbut this is difficult and expensive.

More often, the surface is broken into many small triangles;the approximation error is carefully controlled.

Homework: read and summarise (for yourself) one method for decomposing a surface into triangles.

Computing normals of surfaces

Could use normal from a triangle.Better to get the normal at a point directly.

compute partials(only one is shown)

get normal direction

Issues not considered here:

Stitching patches together to make a bigger surface.Continuity issues constrain the points around the joins.

The differential geometry of curves and surfacesFrenet framesGaussian curvature

• Construct Solid Geometry basics

• Voxel models

L10: CSG and Voxels

CSG basics

Brep is not the only way to represent objects.CSG is another common way.

CSG is models define sets of points via inequalities.

Example

A straight line, y = mx+c, divides the plane:

set f(x,y) = mx + c – y

then f(x,y) > 0 are points above the linef(x,y) = 0 are points on the linef(x,y) < 0 are point below the line

More generally

if f(x,y,z) is any scalar function of 3 spatial variablesthen it can be used to partition space points,using the sign of f(x,y,z).

More examples:

unit sphere: f(x,y,z) = x^2 + y^2 + z^2 – 1

unit cylinder: f(x,y,z) = x^2 + y^2 – 1

Let’s write f(x) in place of f(x,y,z);x has become a vector, x = (x,y,z,1).

Now suppose f(x) = x12 + x2

2 -1, a sphere

And suppose M is a transform,

then f(xM) is a mapped version of the sphere;

is an ellipse – maybe with a new centre.

Creating new shapes

Combining shapes

CSG primitives are just sets of points.

So they are easy to combine into more complex shapes.

plane(.)plane(.) plane(.) plane(.)

tetrahedron(.)

AND

sphere(.) cylinder(.)

DIF

doughnut(.)

Any set operation can be used to combine primitivesAny new shapes can be combined too.

Rendering CSG

Ray tracing could be suitable –

the “CSG” plane is already used by us!and bounding spheres are CSG models!

We just have to be clever about following the model treewhen deciding if a ray strikes an object (eg dougnut example).

And (partial) differentiating gives normals;

Homework: show the normal at a point x on the unit sphere is 2x.

CSG vs Brep

CSG B-rep

points implicit point explicit

sets easy to handle sets hard to handle

goodish to ray trace

hard for radiosity

hardish to ray-trace

good for radiosity

easy to combine hard to combine

hard to get from life easier to get from life

NB: This list is my personal view based on vast non-experience.

Voxel Models

A pixel is a picture element – an area of colour

A voxel is a volume element – a volume of colour

Voxels are often (partially) transparent.

Voxels (typcally) come from medical data:CAT - bonesMRI – soft tissue

a single voxel

many voxels

voxels are arranged into a brick;

each brick can contain hundreds – thousands – of voxels

in each direction

Typical rendering: CAT scans

CAT scan data produces voxels with a “density”,

so is a scalar field f(x,y,z).

1. Use the scalar field to assign colourand opacity

2. Compute a normal at each voxelUse finite differences for this

3. Shoot a ray through a pixel into the volume,compute lighting at each voxelintegrate colour and opacity over the ray

f

colo

ur

Example – taken from Wikipedia

• Texture maps

• Variations

• Two problems

L11: Texture maps

Texture maps: what and why

A texture map is a picture:

a carpeta label on a can of foodthe earth…many other examples

Texture maps are used where the level detail neededmake standard modelling inefficient / impossible.

The idea is to “wrap” or “warp” a flat pictureonto a 3D surface

In this example,a texture of the earth ismapped onto a sphere

+

=

Variations on a theme

• Bump mapping

• Environment mapping

• Procedural mapping

Two problems

1. Flat pictures will not map onto many (most) surfaces:toroidsbunniesetc

2. Aliasing – the texture map is made of pixelsthe pixels can show upand the texture is squashed / stretchedto fit the surfacemaking the problem worse

x

(sort of) solving problem 1

Use “intermediate” shapes, typicallyspherecylindercube

x is a point on the object

c is the object centre

y is the project of x on the im-shape

y

c

Basic algorithm has two steps:

1. Map the 2D texture (u,v) to the 3D surface (x,y,z):x = x(u,v),y = y(u,v),z = z(u,v)T(u,v) -> T(x,y,z)

This is called the S-mapping

2. Map the object to the same 3D surfacex = x’(xo,yo,zo)y = y’(xo,yo,zo)z = z’(xo,yo,zo)T(x,y,z) -> T(xo,yo,zo)

This is called the O-mapping

Other O-mappings from the object to im-surface exist

x x

x x

centroid object normal

surface normalreflected ray

this one makes reflect the environment

simple bump opacity

www.sli.unimelb.edu.au/envis/texture.html

environment

3D textures(procedural)

(sort of) solving problem 2

Recall: Aliasing of textures

solution : mip-mapping

mip = multim im parvo "many things in a small space“

The texture is scaled

and filtered

BEFORE use

1.The texture is scaled

and filtered

2.In use, the program selected the

correct section of the mipmap, according

to distance

• Animation as an art

• Animation of solid bodies

• Camera motion

L12: Animation Basics

Animation as an art

• technically, animation is motion

• artistically it means “bring to life”

Traditional animatorsuse many tricks to bringtheir characters to life –to animate them.

All tricks break Physics

streak-lines

ghosting

squash-and-stretch

rubber inertia

Animation of solid bodies

This is easy – the same transform is applied to all points

The transform is usually a matrix, but does not have to be.

The transform can be differential

x(t+dt) = f[ x(t) ]

but this often leads to numerical error;

so the transform is more often absolute

x(t) = f[ x(0) ]

The set of point {x(0)} is – conveniently – the “canonical” object

Examples

RiTiTiRi

Camera motion

Cameras often follow a path through a scene.

The path is often defined by a cubic curve.

But this just says where the camera is, we need to know:

* where it points

* and control the camera speed.

We must return to the differential geometry of a curve...

To control speed and acceleration along the curve

we need

Arc-Length Parameterisation

Recall the Brep curves; x(u) = U*P is a point on the curve

u is a “natural parameter”

parameterisation is arbitrary

define c(u) as the length of the arc

To control viewing direction along the curve

we need

Frenet Frames

• Model Specifications

• Hierarchies of transforms

• Inverse kinematics/dynamics

L13: Articulated Figures

Model specifications

Articulated models comprise

• a set of “limbs”

• joints that link the limbs

• methods to move the limbs

plus appearance information (colour etc)

Here the limbs are just rectangles;

limbs that are rigid bodies are much easier to handle

in which case rectangles are OK to use.

So we can think about the joints.

These are specified in many different ways.

Here a joint is just a point

the limbs move around the point

Often within a constrained angle

(In 3D, within a constrained solid angle)

rotation angle

parent frame

The way angles are measured

must be carefully specified.

A frame is needed.

Here the “x” axis runs between joints.

X-axis rotation is, then , angle moved.

Hierarchies of TransformsSimple articulated models have a tree-like structure

head

right upper arm

right lower armleft lower arm

left upper arm left upper leg

left lower leg

left foot

right upper leg

right lower leg

right foot

body

neck

This tree is used to locate the whole body first, then the next layer of limbs,

and so on. In this tree the feet are placed last of all.

Placing a transform at each node in the tree

is a very common method

to make the object move.

We already know how to orient the limb inside its frame,

so all we do is move the frame!

Which way around is correct?

y = xMbodyMupperMlowerMfoot

or

y = xMfootMlowerMupperMbody

or will either do?does it matter?

Problem:

in the modified video,

why does the foot not goes though the floor?

Inverse kinematics/dynamics

• Kinematics – the location of the limbs at any given time.

• Dynamics – the motion of the limbs at any given time.

IK method

Put the “puppet” into a poses at various times.

The computer works out how to move from one pose to the next,

in the time available.

The solution has to be subject to some constraints,

typically minimisation of some energy measure.

The method is hard, and tedious – too much so for these note.

See Watt & Watt, section 16.4 for details.

• Non-physical models

• Physical models; elasticity etc for cloth, putty

• Water and fluids

L14: Soft Objects and Fluids

Soft Objects and Physics

Animation of rigid bodies is easy to achieve;

Animation of articulated figures benefits from“inverse kinematics” and “inverse dynamics”.

Animation of soft objects all but requires the use of physics.

Non-physics approaches

• Parametric transforms• Spline surfaces• Free-form deformations

Parametric transforms, eg angle of turn about z depend s on z:

R( theta(z) )

Or could animate splines; eg.

Bezier surfaces are easy to move – just move the control points !

• Free form deformations

FFD’s “move the space”.

FFD’s are tricubic Bezier patches:

Q(u,v,w) = sum sum sum p(i,j,k) Bi(u) Bj(v) Bk(w)

Any point on any object and (u,v,w) maps to the point Q.

Physics models

Simple use of physics;

i) Use Newton’s second law

F = m d2x/dt2

write as a set of first order ODES

F = m du/dtu = dx/dt

ii) Integrate the differential equations over time;setting boundary conditions at time = 0;

• Elasticity is a well used model

Soft objects are stretchy !!

Force = -k * stretch

Thus given a form for F, motion is determined.

Different variants can be – and are - used

Bending forces (twisting forces)

Water and fluids

The “atom and force” model is good for “soft solids”

But liquids as best modelled in other ways

Again, though, physics is the basis of all the models

• Not given in 2008

L15: Fire and Smoke

• Why is photorealism the aim? People paint!

• What is NPR?

• NPR issues

L16: NonPhotorealistic Rendering

Why photorealism? people paint!

Why photorealism ?

• Physics provides starting point for models

• Photographs provide a foil to test against

But people paint !

• Artwork shows off important regions better

• Photography is just one depictive style amongst many

Which is the simplest to understand?

What is NPR?

• NPR is a sub-branch of Computer Graphics that studies making images that do not look photographic.

• NPR is said to have started around 1990.

• Several sub-divisions exist– paint-boxes– rendering 3D models– rendering from photographs and/or video

Paint boxes

from the simple… …to the complex

Rendering 3D models

Breslav et al, SIGGRAPH 2007

Rendering from photographs and/or video

NPR issues

• NPR depends on perception and interpretation– “drawing implies seeing”– models are much more complex– interaction is common

• Mark making– What kind of mark?– Where to mark?

• No single foil to test against– there is no “correct” drawing

NPR depends on Perception and Interpretation

No single foil to test against

Mark making: what and where?

What: Emulate real media

pencil, oil paint, chalk, …..

physical models possible

can test for quality of a given mark

What: Produce new media

Can the computer be used to make marks not made before?

Where: Should marks be made locally, globally, or both?

How to decide where to place marks?

Where: How can marks be made stable in animations?

• The problems

• Some solutions

L17: NPR from models

The problems

Consider a pencil drawing, marks are used to depict:

• object boundaries and other contours

• shadows and shading

• texture

Apparent ridges

Judd et al, SIGGRAPH 2007

For apparent ridges we need:

• differential geometry of surfaces

• projection

The curvature at a surface point S(x,y) – how curved is the surface?

Use the Hessian, H, which is a 2x2 matrix of 2nd order partials of S.

The EVD of H = UKUT

principle directions, U = [u1 u2], principle curvature, K = [k1 k2]

Ridges and valleys where k1 is an extremum in direction u1.

Project the local surface geometry (tangents, normals ‘n all),

Look for extremum in this observed geometry.

Shadows and shading

Shading lines – cross hatches etc

Can be based in “image space”………or in “object space”

Again, curvature can be used, this time to direct pencil lines.

Other methods use solid texture maps, that adapt to local lighting and geometry

Marks make textures too…

With NPR, even the marks can be animated (look at the sea)

L18: NPR from images

L18: Smart cut-and-paste

L19: Texture Propagation

L21: NPR from images

L22: Generalised Cameras

Recommended