<samp id="e4iaa"><tbody id="e4iaa"></tbody></samp>
<ul id="e4iaa"></ul>
<blockquote id="e4iaa"><tfoot id="e4iaa"></tfoot></blockquote>
    • <samp id="e4iaa"><tbody id="e4iaa"></tbody></samp>
      <ul id="e4iaa"></ul>
      <samp id="e4iaa"><tbody id="e4iaa"></tbody></samp><ul id="e4iaa"></ul>
      <ul id="e4iaa"></ul>
      <th id="e4iaa"><menu id="e4iaa"></menu></th>

      CEG5304代做、代寫Java/c++編程語言

      時間:2024-04-11  來源:  作者: 我要糾錯



      Project #2 for CEG5304: Generating Images through Prompting and Diffusion-based Models.
      Spring (Semester 2), AY 2023-2024
      In this exploratory project, you are to explore how to generate (realistic) images via diffusion-based models (such as DALLE and Stable Diffusion) through prompting, in particular hard prompting. To recall and recap the concepts of prompting, prompt engineering, LLVM (Large Language Vision Models), and LMM (Large Multi-modal Models), please refer to the slides on Week 5 (“Lect5-DL_prompt.pdf”).
      Before beginning this project, please read the following instructions carefully, failure to comply with the instructions may be penalized:
      1.This project does not involve compulsory coding, complete your project with this given Word document file by filling in the “TO FILL” spaces. Save the completed file as a PDF file for submission. Please do NOT modify anything (including this instruction) in your submission file.
      2.The marking of this project is based on how detailed the description and discussion are over the given questions. To score, please make sure your descriptions and discussions are readable, and adequate visualizations are provided.
      3.The marking of this project is NOT based on any evaluation criteria (e.g., PSNR) over the generated image. Generating a good image does NOT guarantee a high score.
      4.You may use ChatGPT/Claude or any online LLM services for polishing. However, purely using these services for question answering is prohibited (and is actually very obvious). If it is suspected that you generate your answers holistically with these online services, your assignment may be considered as committing plagiarism.
      5.Submit your completed PDF on Canvas before the deadline: 1759 SGT on 20 April 2024 (updated from the slides). Please note that the deadlines are strict and late submission will be deducted 10 points (out of 100) for every 24 hours.
      6.The report must be done individually. You may discuss with your peers, but NO plagiarism is allowed. The University, College, Department, and the teaching team take plagiarism very seriously. An originality report may be generated from iThenticate when necessary. A zero mark will be given to anyone found plagiarizing and a formal report will be handed to the Department/College for further investigation.

      Task 1: generating an image with Stable Diffusion (via Huggingface Spaces) and compare it with the objective real image. (60%)
      In this task, you are to generate an image with the Stable Diffusion model in Huggingface Spaces. The link is provided here: CLICK ME. You can play with the different prompts and negative prompts (prompts that instructs the model NOT to generate something). Your objective is to generate an image that looks like the following image:

      1a) First, select a rather coarse text prompt. A coarse text prompt may not include a lot of details but should be a good starting prompt to generate images towards our objective. An example could be “A Singaporean university campus with a courtyard.”. Display your generated image and its corresponding text prompt (as well as the negative prompt, if applicable) below: (10%)
      TO FILL
      TO FILL
      1b) Describe, in detail, how the generated image is compared to the objective image. You may include the discussion such as the components in the objective image that is missing from the generated image, or anything generated that does not make sense in the real world. (20%)
      TO FILL
      TO FILL
      Next, you are to improve the generated image with prompt engineering. Note that it is highly likely that you may still be unable to obtain the objective image. A good reference material for prompt engineering can be found here: PROMPT ENGINEERING. 
      1c) Describe in detail how you improve your generated image. The description should include display of the generated images and their corresponding prompts, and detailed reasoning over the change in prompts. If the final improved image is generated with several iterations of prompt improvement, you should show each step in detail. I.e., you should display the result of each iteration of prompt change and discuss the result of each prompt change. You should also compare your improved image with both the first image you generated above, as well as the objective image. (30%)
      TO FILL
      TO FILL
      TO FILL
      Task 2: generating images with another diffusion-based model, DALL-E (mini-DALL-E, via Huggingface Spaces). (40%)
      Stable Diffusion is not the only diffusion-based model that has the capability to generate good quality images. DALL-E is an alternative to Stable Diffusion. However, we are not to discuss the differences over these two models technically, but the differences over the generated images qualitatively (in a subjective manner). The link to generating with mini-DALL-E is provided here: MINI-DALL-E.
      2a) You should first use the same prompt as you used in Task 1a and generate the image with mini-DALL-E. Display the generated image and compare, in detail, the new generated image with that generated by Stable Diffusion. (10%)
      TO FILL
      TO FILL
      2b) Similar to what we performed for Stable Diffusion; you are to again improve the generated image with prompt engineering. Describe in detail how you improve your generated image. Similarly, if the final improved image is generated with several iterations of prompt improvement, you should show each step in detail. The description should include display of the generated images and their corresponding prompts, and detailed reasoning over the change in prompts. You should compare your improved image with both the first image you generated above, as well as the objective image.
      In addition, you should also describe how the improvement is similar to or different from the previous improvement process with Stable Diffusion. (10%)
      TO FILL
      TO FILL
      2c) From the generation process in Task 1 and Task 2, discuss the capabilities and limitations over image generation with off-the-shelf diffusion-based models and prompt engineering. You could further elaborate on possible alternatives or improvements that could generate images that are more realistic or similar to the 請加QQ:99515681  郵箱:99515681@qq.com   WX:codinghelp











       

       

       

       

       

      標簽:

      掃一掃在手機打開當前頁
    • 上一篇:越南簽證辦理托運流程多久(行李托運政策)
    • 下一篇:怎么申請菲律賓移民達沃?價格多少
    • 無相關信息
      昆明生活資訊

      昆明圖文信息
      蝴蝶泉(4A)-大理旅游
      蝴蝶泉(4A)-大理旅游
      油炸竹蟲
      油炸竹蟲
      酸筍煮魚(雞)
      酸筍煮魚(雞)
      竹筒飯
      竹筒飯
      香茅草烤魚
      香茅草烤魚
      檸檬烤魚
      檸檬烤魚
      昆明西山國家級風景名勝區
      昆明西山國家級風景名勝區
      昆明旅游索道攻略
      昆明旅游索道攻略
    • 福建中專招生網 NBA直播 短信驗證碼平臺 幣安官網下載 WPS下載

      關于我們 | 打賞支持 | 廣告服務 | 聯系我們 | 網站地圖 | 免責聲明 | 幫助中心 | 友情鏈接 |

      Copyright © 2025 kmw.cc Inc. All Rights Reserved. 昆明網 版權所有
      ICP備06013414號-3 公安備 42010502001045

      主站蜘蛛池模板: 亚洲日韩一区二区一无码| 蜜桃臀无码内射一区二区三区| 亚洲AV中文无码字幕色三| 国精品无码一区二区三区在线 | 亚洲精品无码高潮喷水在线| 亚洲AV无码国产精品色| 中文字字幕在线中文无码 | 在线播放无码后入内射少妇| 无码夫の前で人妻を侵犯 | 毛片无码免费无码播放| 亚洲Av无码国产情品久久| 亚洲中文字幕久久精品无码A| 国产成人无码18禁午夜福利p| 亚洲av无码兔费综合| 久久无码人妻一区二区三区午夜| 丰满少妇人妻无码| 韩国精品一区二区三区无码视频| 无码国产色欲XXXXX视频| 少妇无码太爽了在线播放| 国产成人无码网站| 国产成人无码精品久久久小说| 亚洲AV无码国产精品麻豆天美| 高潮潮喷奶水飞溅视频无码| 国产精品亚洲一区二区无码| 亚洲中文字幕久久无码| 永久免费av无码网站大全| 亚洲AⅤ永久无码精品AA | 无码国产乱人伦偷精品视频| 内射无码专区久久亚洲| 亚洲Av无码国产一区二区| 精品欧洲av无码一区二区14| 无码国内精品久久人妻蜜桃| 午夜亚洲AV日韩AV无码大全| 国产AV无码专区亚洲A∨毛片| 亚洲精品无码成人片久久| 亚洲V无码一区二区三区四区观看 亚洲爆乳精品无码一区二区三区 亚洲爆乳无码一区二区三区 | 亚洲精品无码专区2| 国产精品无码素人福利| 久久精品无码专区免费| 亚洲中文字幕无码永久在线| 亚洲国产精品无码成人片久久|