User Tools

Site Tools


cs501r_f2016:fp

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

cs501r_f2016:fp [2016/11/22 15:53]
wingated
cs501r_f2016:fp [2021/06/30 23:42]
Line 1: Line 1:
-====Objective:​==== 
  
-To creatively apply knowledge gained through the course of the semester to a substantial learning problem of your own choosing. 
- 
----- 
-====Deliverable:​==== 
- 
-For your final project, you should execute a substantial project of your own choosing. ​ Many different kinds of final projects are possible. ​ A few examples include: 
- 
-  * Learning how to render a scene based on examples of position and lighting 
-  * Learning which way is "​up"​ in a photo (useful for drone odometry) 
-  * Training an HTTP server to predict which web pages a user will likely visit next 
-  * Training an earthquake predictor 
-  * Using GANs to turn rendered faces into something more realistic (avoiding the "​uncanny valley"​) 
-  * Transforming Minecraft into a more realistic looking game with DNN post-processing 
-  * Using style transfer on a network trained for facial recognition (to identify and accentuate facial characteristics) 
-  * Using RGB+Depth datasets to improve geometric plausibility of GANs 
- 
-The project can involve any application area, but the core challenge must be tackled using some sort of deep learning. 
- 
-**Your project will be graded more on effort than results.** ​ As I have stated in class, I would rather have you swing for the fences and miss, than take on a simple, safe project. ​ **It is therefore very important that your final writeup clearly convey the scope of your efforts.** 
- 
-The best projects involve a new, substantive idea and novel dataset. ​ It may also be acceptable to use vanilla DNN techniques on a novel dataset, as long as you demonstrate significant effort in the "​science"​ of the project -- evaluating results, exploring topologies, thinking hard about how to train, and careful test/​training evaluation. ​ It may also be acceptable to simply implement a state-of-the-art method from the literature, but clear such projects with me first. 
- 
----- 
-====Grading standards:​==== 
- 
-Your entry will be graded on the following elements: 
- 
-  * 5% Clearly motivated problem 
-  * 10% Exploratory data analysis 
-  * 35% Description of technical approach 
-    * 5% How will you know if you succeed? ​ Are there quantitative metrics for success (such as a classification error rate), or will success be judged qualitatively (such as the image quality of GAN-generated images)? 
-    * 10% How did you prepare and analyze your data?  How did you establish baselines, and test/train splits? 
-    * 15% Describe how  DNNs fit into your solution method. ​ Discuss whether this is a supervised, unsupervised,​ or RL problem. 
-    * 5% Is there anything unique about your problem, or about the way you applied DNNs?  ​ 
-  * 45% Analysis of results 
-    * 25% Present your final results, including comparison to baselines, in whatever format is most appropriate to your problem 
-    * 20% Describe the process of getting to your final result. ​ What did you tweak? ​ Did you iterate on your topology? ​ How did you debug your model? ​ Include anything relevant to support your discussion, such as tensorboard screenshots,​ graphs of cost decreasing over time, charts comparing different topologies, etc. 
-  * 5% Tidy and legible final writeup 
- 
----- 
-====Notes:​==== 
- 
-You are welcome to use any publicly available code on the internet to help you. 
- 
-Here are some possible questions that you might consider answering as part of your report: 
- 
-  - **A discussion of the dataset** 
-    - Where did it come from?  Who published it? 
-    - Who cares about this data? 
-  - **A discussion of the problem to be solved** 
-    - Is this a classification problem? ​ A regression problem? 
-    - Is it supervised? ​ Unsupervised?​ 
-    - What sort of background knowledge do you have that you could bring to bear on this problem? 
-    - What other approaches have been tried? ​ How did they fare? 
-  - **A discussion of your exploration of the dataset**. 
-    - Before you start coding, you should look at the data.  What does it include? ​ What patterns do you see? 
-    - Any visualizations about the data you deem relevant 
-  - **A clear, technical description of your approach.** ​ This section should include: 
-    - Background on the approach 
-    - Description of the model you use 
-    - Description of the inference / training algorithm you use 
-    - Description of how you partitioned your data into a test/​training split 
-  - **An analysis of how your approach worked on the dataset** 
-    - What was your final RMSE on your private test/​training split? 
-    - Did you overfit? ​ How do you know? 
-    - Was your first algorithm the one you ultimately used for your submission? ​ Why did you (or didn't you) iterate your design? 
-    - Did you solve (or make any progress on) the problem you set out to solve? 
- 
----- 
-====Possible sources of interesting datasets==== 
- 
-Croudflower 
- 
-KDD cup 
- 
-UCI repository 
- 
-Kaggle (current and past) 
- 
-Data.gov 
- 
-AWS 
- 
-World bank 
- 
-BYU CS478 datasets 
- 
-data.utah.gov 
- 
-Google research 
- 
-BYU DSC competition 
cs501r_f2016/fp.txt · Last modified: 2021/06/30 23:42 (external edit)