91精品国产91久久久久久_国产精品二区一区二区aⅴ污介绍_一本久久a久久精品vr综合_亚洲视频一区二区三区

合肥生活安徽新聞合肥交通合肥房產生活服務合肥教育合肥招聘合肥旅游文化藝術合肥美食合肥地圖合肥社保合肥醫院企業服務合肥法律

代寫CSE 167、輔導3D OpenGL Rendering

時間:2023-11-21  來源:合肥網hfw.cc  作者:hfw.cc 我要糾錯


UCSD CSE 167 Assignment 3:

3D OpenGL Rendering

Figure 1: We will develop an interactive interface for inspecting 3D models in this homework.

As you can probably tell from the previous homeworks, rendering requires computing interactions between

millions of pixels and billions of triangles. This leads to significant challenges in performance, especially when

we want to interact with the content in real-time. To make things really fast, pioneers in computer graphics

came up with the solution to use domain-specific hardware to speedup rendering. Instead of using a general

purpose computer to compute everything, we build chips that specialize at rendering. These processors are

called the Graphics Processing Units (GPUs). The idea of GPUs can be traced back to more than 40 years

ago: The first GPU, Geometry Engine was developed by Jim Clark and Marc Hannah in 1981. Jim Clark

formed the company Silicon Graphics Inc (SGI) in the same year and SGI was one of the most important

computer graphics companies in the history. Nowadays, GPUs are found to be general enough to compute

very wide-range of computation, including deep learning and many scientific computing tasks, and they are

indispensable to the human society. GPU is one of the most successful examples of domain-specific hardware.

In this homework, we will write code to render things using GPUs on your computer. To command your

GPUs, we need to send commands to it using some sort of “Application Programming Interface” (API).

These interfaces are collectively decided by the GPU companies and some other organizations, and each

hardware will come with some “drivers” that actually implement these interfaces using underlying hardware

instructions. The most popular APIs are: OpenGL, DirectX, Metal, Vulkan, and WebGPU. Among these,

DirectX is Windows only, Metal is MacOS only, WebGPU is only for browsers, and Vulkan is extremely

low-level and very verbose for providing fine-grained control (it takes literally a thousand lines to render a

single triangle in Vulkan). Therefore, we will use OpenGL in this homework: even though DirectX, Metal,

and Vulkan are more update to date (the lastest version of OpenGL is 6 years ago), OpenGL is still use

in practice and supported by all major GPUs and OSes, and it is significantly easier to learn compared to

other lower-level APIs. Just like programming languages, it’ll be a lot easier to learn other APIs once you’ve

learned OpenGL.

In this homework, we will mostly follow an online tutorial: learnopengl.com, because they likely write

significantly better tutorials than me. We will implement what we did in the previous homework in OpenGL

and hopefully see significant speedup. We will also create a Graphics User Interface (GUI) and enable

real-time interaction.

This homework is also more “open-ended” compared to the previous ones. We do not ask you to produce

the exact same output as we do. At this point, you should be familiar with the theory of rasterization. We’re

just wrangling with hardware interface, so allowing a bit of creativity seems reasonable.

1

1 Creating a window (10 pts)

Our first task, instead of rendering a single triangle, is to create a window! Read the chapters of OpenGL,

Creating a window, and Hello Window in learnopengl.com to see how to create a window with OpenGL

context using GLFW. Pick your favoriate background color. We have included GLFW and glad in balboa,

so you shouldn’t have to download them. We’re using OpenGL 3.3, but feel free to use the version you like.

Implement your code in hw_3_1 in hw3.cpp. Test it using

./balboa -hw 3_1

Once you are done, take a screenshot of the window you created and save it as outputs/hw_3_1.png.

2 Rendering a single 2D triangle (20 pts)

Yeah, it’s that time again! Read the Hello Triangle chapter and render a single triangle with constant color

(pick one that you like the most). Make sure you’ve become familiar with the ideas of shaders, VAO, VBO,

and EBO. Just to make things slightly different so that we are not just copy and pasting code, let the triangle

rotate in the image plane over time (it can be clockwise or counterclockwise, your choice). For the rotation,

you can do it whichever way you want, but I recommend you do it in the vertex shader. Read the Shaders

chapter and understand how to pass in a uniform variable, then you can use the uniform variable as the

rotation angle.

float vs. double By default, balboa uses double precision floats through the Real type. However, by

default, GLSL uses single precision floats. Be careful of this discrepancy. You can use Vector3f/Matrix3x3f

to switch to float in balboa. Also feel free to use the glm library which is used in the tutorial.

Implement your code in hw_3_2 in hw3.cpp. Test it using

./balboa -hw 3_2

This time, do a screen recording of your rotating triangle and save it as outputs/hw_3_2.mp4 (or whatever

encoding you are using).

3 Rendering 3D triangle meshes with transformations (35 pts)

Next, we’ll use OpenGL to render the type of scenes we handled in the previous homework. Read the

chapters Transformations, Coordinate systems, and cameras, and that should give you enough knowledge to

render the JSON scenes like the ones in the previous homeworks.

This part is a big jump from the previous parts. I would recommend you to do things incrementally. E.g.,

handle two 2D triangles first, add projection matrix, add view matrix, add model matrix, handle multiple

triangle meshes, and finally add camera interaction.

Below are some notes and tips:

Clip space. In Homework 2, our projection matrix convert from camera space directly to the screen space.

In OpenGL, the hardware expects the projection to convert from camera space to the clip space, which by

default ranges from −1 to 1 for x, y, and z axes. Everything outside of the clip space is clipped. Note that

the clipping happens at the far side of z as well – we use the z_far parameter in the camera in our JSON

scene to specify this. The difference in spaces means that we need to use a different projection matrix:

1

as

0 0 0

0

1

s

0 0

0 0 −

zfar

zfar−znear

zfarznear

zfar−znear

0 0 −1 0

?**7;

?**8;

?**8;

?**9;

, (1)

2

where s is the scaling/film size parameter as before, and a is the aspect ratio. The first row and the second

row scale the x and y clipping plane to [−1, 1] respectively. The third row compresses z values from −znear

to −zfar to [−1, 1]. The fourth row is the perspective projection using homogeneous coordinates.

Depth test. By default, OpenGL does not reject triangles when they are occluded. Remember to turn

on depth testing using glEnable(GL_DEPTH_TEST) and clear the Z buffer (e.g., glClear(GL_COLOR_BUFFER_BIT

| GL_DEPTH_BUFFER_BIT)).

Vertex colors. In contrast to the learnopengl tutorial, balboa stores the vertex color in a separate array.

Therefore it’s likely more convienent to create two VBOs:

unsigned int VBO_vertex;

glGenBuffers(1, &VBO_vertex);

glBindBuffer(GL_ARRAY_BUFFER, VBO_vertex);

glBufferData(GL_ARRAY_BUFFER, ...);

glVertexAttribPointer(0 /* layout index */,

3, GL_FLOAT, GL_FALSE, 3 * sizeof(float), (void*)0);

glEnableVertexAttribArray(0);

unsigned int VBO_color;

glGenBuffers(1, &VBO_color);

glBindBuffer(GL_ARRAY_BUFFER, VBO_color);

glBufferData(GL_ARRAY_BUFFER, ...);

glVertexAttribPointer(1 /* layout index */,

3, GL_FLOAT, GL_FALSE, 3 * sizeof(float), (void*)0);

You only need one VAO per mesh regardless.

Multiple meshes. To handle multiple meshes in a scene, create a VAO for each mesh.

Window resizing. We don’t require you to handle window resizing in this homework. It’s annoying

because you’ll need to regenerate the projection matrix every time the aspect ratio changes.

Gamma correction. When we save the image in balboa, we perform a gamma correction by taking a power

of 1

2.2

. OpenGL does not by default do this. To enable gamma correction, use glEnable(GL_FRAMEBUFFER_SRGB).

Read the gamma correction chapter in learnopengl.com to learn more.

Camera interaction. Like the tutorial, you should also implement a simple camera interaction scheme,

see the Camera chapter. A simple WSAD style translation suffices. To obtain the camera direction and

right vector, you can look at the columns of the cam_to_world matrix.

As a bonus (15 pts), add camera rotation based on mouse input like the tutorial. Note that the rotation

in the tutorial assumes a particular camera frame and would not work for our case. I recommend doing the

following: 1) store yaw and pitch angles and the original cam_to_world matrix from the scene. 2) update the

yaw and pitch based on the mouse movement offsets like in the tutorial. 3) form a rotation matrix R based

on yaw and pitch, then form a new cam_to_world matrix by multiplying the original cam_to_world matrix

with R. (Don’t overwrite the original cam_to_world matrix!)

For rotation, it might be tempting to keep only one cam_to_world matrix by keep multiplying it with

new rotation matrices. However, this is going to produce unintuitive behavior (try it!) since yaw and

pitch rotations are not commutative: applying yaw first then pitch will produce different result compared to

applying pitch first then yaw. As a result, when you chain together many pitches and yaws matrix rotations,

they will not represent the desired rotation. Yes, rotation is weird. This is why you should explicitly store

the yaw and pitch angles and modify those instead.

3

Passing parameters in callback functions. If you dislike global variables as much as me, you would

like the functions glfwSetWindowUserPointer and glfwGetWindowUserPointer. You will use it like this:

void mouse_callback(GLFWwindow* window, double xpos, double ypos) {

StructIWanttoPasstoCallback *data_ptr =

glfwGetWindowUserPointer(window);

}

GLFWwindow* window = glfwCreateWindow(width, height, "Balboa", NULL, NULL);

StructIWanttoPasstoCallback data = ...;

glfwSetWindowUserPointer(window, &data);

glfwSetCursorPosCallback(window, mouse_callback);

Debugging. Debugging OpenGL (and other graphics API) programs is painful: if you do one thing wrong,

you’ll likely get a black screen. The learnopengl tutorial provides useful tips for debugging. To debug shaders,

it’s particularly useful to use a debugger such as renderdoc. Unfortunately, none of the existing OpenGL

debuggers work on MacOS anymore (Apple makes it extremely hard to develop OpenGL on MacOS because

they want people to use Metal). For MacOS users, a potential debugging strategy is to emulate the shader

on CPU: write the same code on CPU and print out the values, and see if it does what you expect. It’s going

to be painful regardless, I’m sorry. On the other hand, this is a fruitful research area that awaits innovation

to make things better!

For the 3D transformation, copy your Homework 2 code to the parse_transformation function in hw3_scenes.cpp.

Implement the rest in hw_3_3 in hw3.cpp.

Test your OpenGL rendering using the following commands:

./balboa -hw 3_3 ../scenes/hw3/two_shapes.json

./balboa -hw 3_3 ../scenes/hw3/cube.json

./balboa -hw 3_3 ../scenes/hw3/spheres.json

./balboa -hw 3_3 ../scenes/hw3/teapot.json

./balboa -hw 3_3 ../scenes/hw3/bunny.json

./balboa -hw 3_3 ../scenes/hw3/buddha.json

For two_shapes and cube, they should render to the same images as the previous homework (before you

move the camera yourself). The rest are new scenes. (teapot.json is a higher-resolution version that has 10

times more triangles!) Record a video of you moving the camera for each scene and save them as:

outputs/hw_3_3_two_shapes.mp4

outputs/hw_3_3_cube.mp4

outputs/hw_3_3_spheres.mp4

outputs/hw_3_3_teapot.mp4

outputs/hw_3_3_bunny.mp4

outputs/hw_3_3_buddha.mp4

Acknowledgement. The bunny model was scanned by Greg Turk and Marc Levoy back in 1994 at

Stanford, so it is sometimes called the Stanford bunny. The texture of the bunny model was made by

KickAir_8p who posted the scene in blenderarists.org. The buddha texture was generated by Kun Zhou et

al. for their Texturemontage paper.

Bonus: textures (15 pts). Read the Textures chapter of learnopengl.com and implement textures for

the shapes above. We have provided the UV maps for the models except two_shapes and cube. I have also

included the original textures I used to produce the vertex colors for teapot, bunny, and buddha.

4

4 Lighting (25 pts)

For this part, read the chapters of Colors and Basic Lighting in the tutorial, and implement some basic

lighting in our viewer. Be careful about the transformation of the normals! Use the vertex colors or texture

colors as the objectColor equivalent in the tutorial. Let’s assume ambientStrength=0.1, specularStrength=0.5

and lightDir is at normalize(vec3(1, 1, 1)). Note that you can extract the camera position by looking at

the fourth column of cam_to_world.

The way the tutorial does the lighting requires defining vertex normals (an alternative is to use face

normals, but it often looks uglier). We have provided vertex normals for the following scenes:

./balboa -hw 3_4 ../scenes/hw3/spheres.json

./balboa -hw 3_4 ../scenes/hw3/teapot.json

./balboa -hw 3_4 ../scenes/hw3/bunny.json

./balboa -hw 3_4 ../scenes/hw3/buddha.json

Save your output as screenshots:

outputs/hw_3_4_spheres.png

outputs/hw_3_4_teapot.png

outputs/hw_3_4_bunny.png

outputs/hw_3_4_buddha.png

Bonus: lighting animation (10 pts). Add some animation to the light. Make it move the way you like,

and submit a video recording of the animation.

Bonus: different types of lights (10 pts). Our light currently is a directional light. Implement point

lights and spot lights (see the Light casters chapter) in your renderer, and support multiple lights.

Bonus: shadow mapping (20 pts). Implement a basic shadow map. See the Shadow Mapping chapter

in learnopengl. Support of directional lights is good enough.

5 Design your own scenes (10 pts)

We’re at the fun part again. Design your own scene and render it using your new renderer!

請加QQ:99515681 或郵箱:99515681@qq.com   WX:codehelp

 

掃一掃在手機打開當前頁
  • 上一篇:指標代寫 代寫同花順指標公式
  • 下一篇:指標代寫 代寫同花順指標公式
  • 無相關信息
    合肥生活資訊

    合肥圖文信息
    2025年10月份更新拼多多改銷助手小象助手多多出評軟件
    2025年10月份更新拼多多改銷助手小象助手多
    有限元分析 CAE仿真分析服務-企業/產品研發/客戶要求/設計優化
    有限元分析 CAE仿真分析服務-企業/產品研發
    急尋熱仿真分析?代做熱仿真服務+熱設計優化
    急尋熱仿真分析?代做熱仿真服務+熱設計優化
    出評 開團工具
    出評 開團工具
    挖掘機濾芯提升發動機性能
    挖掘機濾芯提升發動機性能
    海信羅馬假日洗衣機亮相AWE  復古美學與現代科技完美結合
    海信羅馬假日洗衣機亮相AWE 復古美學與現代
    合肥機場巴士4號線
    合肥機場巴士4號線
    合肥機場巴士3號線
    合肥機場巴士3號線
  • 短信驗證碼 目錄網 排行網

    關于我們 | 打賞支持 | 廣告服務 | 聯系我們 | 網站地圖 | 免責聲明 | 幫助中心 | 友情鏈接 |

    Copyright © 2025 hfw.cc Inc. All Rights Reserved. 合肥網 版權所有
    ICP備06013414號-3 公安備 42010502001045

    91精品国产91久久久久久_国产精品二区一区二区aⅴ污介绍_一本久久a久久精品vr综合_亚洲视频一区二区三区
    色欧美乱欧美15图片| 精品免费日韩av| 久久久久综合一区二区三区| av不卡在线看| 久久国产主播| 精品免费99久久| 国产精品亲子乱子伦xxxx裸| 中文字幕视频一区二区三区久| 亚洲男人的天堂av| 免费不卡在线视频| 成人免费视频一区二区| 欧美一区三区二区在线观看| 国产一区二区无遮挡| 国产精品乱码一区二区三区| 欧美性猛交xxxx黑人交| 久久精品人人爽人人爽| 一区二区三区日韩欧美| 国产在线精品不卡| 黄色亚洲在线| 678五月天丁香亚洲综合网| 中文字幕一区二区三区四区| 午夜欧美电影在线观看| 91免费小视频| 欧美日韩精品一区二区三区蜜桃| 久久久久国产精品麻豆| 美日韩一级片在线观看| 在线观看一区视频| 欧美成人激情免费网| 亚洲成人精品在线观看| 你懂的国产精品永久在线| 欧美在线看片a免费观看| 中文字幕一区二区三区蜜月 | 亚洲韩国精品一区| 欧美日韩精品免费观看| 精品欧美乱码久久久久久1区2区| 三级欧美韩日大片在线看| 99精品久久| 综合欧美亚洲日本| 欧美特黄一区| 国产精品视频一二三区| 91色婷婷久久久久合中文| 欧美顶级少妇做爰| 丝袜亚洲另类丝袜在线| 伊人久久久大香线蕉综合直播| 欧美二区在线观看| 日韩激情一区二区| 美女视频黄 久久| 午夜影院日韩| 水野朝阳av一区二区三区| 天天色综合天天| 亚洲永久网站| 亚洲精品乱码久久久久久日本蜜臀| 国产成人在线观看免费网站| 色婷婷综合激情| 天天色天天爱天天射综合| 永久域名在线精品| 一区二区三区中文字幕电影| 国内揄拍国内精品久久| 国产欧美一区二区三区另类精品 | 亚洲成av人影院| 99伊人成综合| 日本欧美一区二区| 色婷婷亚洲精品| 久久国产尿小便嘘嘘| 色噜噜狠狠色综合欧洲selulu| 午夜精品在线视频一区| 在线日韩国产精品| 国产一区二区按摩在线观看| 欧美久久久影院| 波多野结衣的一区二区三区| 久久老女人爱爱| 日韩一区二区久久| 亚洲第一电影网| 日韩亚洲欧美一区| 97久久精品人人爽人人爽蜜臀| 欧美激情一区二区| 国产日韩一区二区三区在线播放| 日韩avvvv在线播放| 欧美日本视频在线| 韩国欧美国产1区| 中文字幕免费在线观看视频一区| 91久久中文| 国产一区二区在线电影| 欧美韩国日本不卡| 欧美另类高清视频在线| 国产精品久久综合| 亚洲精一区二区三区| 激情小说亚洲一区| 日韩欧美自拍偷拍| 美女亚洲精品| 欧美福利电影在线观看| 天堂成人免费av电影一区| 久久精品夜夜夜夜久久| 久久激情一区| 在线精品在线| 国产伦精一区二区三区| 亚洲高清一区二区三区| 久久午夜老司机| 在线日韩av| 91亚洲国产成人精品一区二三| 日韩在线一区二区| 国产精品麻豆一区二区| 欧美一区中文字幕| 欧美性色综合网| 校园激情久久| 一区二区福利| 国产综合色一区二区三区| 成人av电影在线播放| 亚洲成人激情社区| 亚洲国产精品久久久久秋霞影院| 久久综合久久综合九色| 日韩午夜av一区| 欧美综合久久久| 国产日韩一区二区| 国产欧美一级| 国产精品色网| 性一交一乱一区二区洋洋av| 伊大人香蕉综合8在线视| 午夜国产一区| 国内自拍一区| 精品成人在线| 欧美日韩综合网| 黄色成人在线网站| 夜夜嗨一区二区| 国产一区自拍视频| 国产视频一区在线观看一区免费| 国产日本精品| 欧美日韩一本到| 久久久噜噜噜久噜久久综合| 欧美国产视频在线| 欧美韩日一区二区三区| 亚洲一区在线观看免费观看电影高清 | 国产精品入口66mio| 日本大香伊一区二区三区| 91精品福利在线| 日韩午夜av电影| 国产精品久线观看视频| 亚洲成人精品影院| 天堂va蜜桃一区二区三区漫画版| 国产精品99久久久久久似苏梦涵| 91视视频在线观看入口直接观看www | 色网综合在线观看| 免费中文字幕日韩欧美| 色av成人天堂桃色av| 欧美精品在欧美一区二区少妇| 国产精品综合| 精品伦理精品一区| 日本视频一区二区| 激情av一区| 精品国产乱码久久久久久图片 | 激情久久综合| 欧美精品一区二区在线播放| 亚洲一区av在线| 黄色欧美日韩| 国产农村妇女精品一区二区| 国产精品欧美综合在线| 国产成人免费视频一区| 影音先锋日韩资源| 欧美videossexotv100| 国产精品羞羞答答xxdd| 在线观看日韩高清av| 免费观看一级欧美片| 国产精品美女黄网| 有码一区二区三区| 亚洲视频一区| 综合分类小说区另类春色亚洲小说欧美 | 国产伦精品一区二区三区免费迷| 亚洲永久免费| 日本在线观看不卡视频| 欧美福利视频导航| 久久av老司机精品网站导航| 久久久综合网| av中文字幕一区| 国产日产精品一区| 好吊色欧美一区二区三区四区| 欧美另类videos死尸| 黄页视频在线91| 国产女主播在线一区二区| 国产精品xnxxcom| 亚洲第一在线综合网站| 狠狠色噜噜狠狠狠狠色吗综合| 偷偷要91色婷婷| 欧美日韩一区二区三区免费看| 久久99国产精品尤物| 国产精品高潮久久久久无| 99re66热这里只有精品4| 亚洲国产精品一区二区久久 | 欧美日本国产视频| 激情欧美日韩一区| 手机精品视频在线观看| 欧美一卡二卡三卡四卡| 不卡高清视频专区| 亚洲国产视频a| 精品91自产拍在线观看一区| 亚洲精品一二| 成人丝袜高跟foot| 激情综合色播激情啊| 国产精品久久久久久久午夜片| 欧洲激情一区二区| 一本久久a久久免费精品不卡|