爱情鸟第一论坛com高清免费_91免费精品国自产拍在线可以看_亚洲一区精品中文字幕_男人操心女人的视频

代做IEMS 5730、代寫 c++,Java 程序設計

時間:2024-03-11  來源:  作者: 我要糾錯



IEMS 5730 Spring 2024 Homework 2
Release date: Feb 23, 2024
Due date: Mar 11, 2024 (Monday) 11:59:00 pm
We will discuss the solution soon after the deadline. No late homework will be accepted!
Every Student MUST include the following statement, together with his/her signature in the submitted homework.
I declare that the assignment submitted on Elearning system is original except for source material explicitly acknowledged, and that the same or related material has not been previously submitted for another course. I also acknowledge that I am aware of University policy and regulations on honesty in academic work, and of the disciplinary guidelines and procedures applicable to breaches of such policy and regulations, as contained in the website http://www.cuhk.edu.hk/policy/academichonesty/.
Signed (Student_________________________) Date:______________________________ Name_________________________________ SID_______________________________
Submission notice:
● Submit your homework via the elearning system.
● All students are required to submit this assignment.
General homework policies:
A student may discuss the problems with others. However, the work a student turns in must be created COMPLETELY by oneself ALONE. A student may not share ANY written work or pictures, nor may one copy answers from any source other than one’s own brain.
Each student MUST LIST on the homework paper the name of every person he/she has discussed or worked with. If the answer includes content from any other source, the student MUST STATE THE SOURCE. Failure to do so is cheating and will result in sanctions. Copying answers from someone else is cheating even if one lists their name(s) on the homework.
If there is information you need to solve a problem, but the information is not stated in the problem, try to find the data somewhere. If you cannot find it, state what data you need, make a reasonable estimate of its value, and justify any assumptions you make. You will be graded not only on whether your answer is correct, but also on whether you have done an intelligent analysis.
Submit your output, explanation, and your commands/ scripts in one SINGLE pdf file.

 Q1 [20 marks + 5 Bonus marks]: Basic Operations of Pig
You are required to perform some simple analysis using Pig on the n-grams dataset of Google books. An ‘n-gram’ is a phrase with n words. The dataset lists all n-grams present in books from books.google.com along with some statistics.
In this question, you only use the Google books bigram (1-grams). Please go to Reference [1] and [2] to download the two datasets. Each line in these two files has the following format (TAB separated):
bigram year match_count
An example for 1-grams would be:
volume_count
circumvallate 1978 335 91 circumvallate 1979 261 95
This means that in 1978(1979), the word "circumvallate" occurred 335(261) times overall, from 91(95) distinct books.
(a) [Bonus 5 marks] Install Pig in your Hadoop cluster. You can reuse your Hadoop cluster in IEMS 5730 HW#0 and refer to the following link to install Pig 0.17.0 over the master node of your Hadoop cluster :
http://pig.apache.org/docs/r0.17.0/start.html#Pig+Setup
Submit the screenshot(s) of your installation process.
If you choose not to do the bonus question in (a), you can use any well-installed Hadoop cluster, e.g., the IE DIC, or the Hadoop cluster provided by the Google Cloud/AWS [5, 6, 7] to complete the following parts of the question:
(b) [5 marks] Upload these two files to HDFS and join them into one table.
(c) [5 marks] For each unique bigram, compute its average number of occurrences per year. In the above example, the result is:
circumvallate (335 + 261) / 2 = 298
Notes: The denominator is the number of years in which that word has appeared. Assume the data set contains all the 1-grams in the last 100 years, and the above records are the only records for the word ‘circumvallate’. Then the average value is:
 instead of
(335 + 261) / 2 = 298, (335 + 261) / 100 = 5.96
(d) [10 marks] Output the 20 bigrams with the highest average number of occurrences per year along with their corresponding average values sorted in descending order. If multiple bigrams have the same average value, write down anyone you like (that is,

 break ties as you wish).
You need to write a Pig script to perform this task and save the output into HDFS.
Hints:
● This problem is very similar to the word counting example shown in the lecture notes
of Pig. You can use the code there and just make some minor changes to perform this task.
Q2 [20 marks + 5 bonus marks]: Basic Operations of Hive
In this question, you are asked to repeat Q1 using Hive and then compare the performance between Hive and Pig.
(a) [Bonus 5 marks] Install Hive on top of your own Hadoop cluster. You can reuse your Hadoop cluster in IEMS 5730 HW#0 and refer to the following link to install Hive 2.3.8 over the master node of your Hadoop cluster.
https://cwiki.apache.org/confluence/display/Hive/GettingStarted
Submit the screenshot(s) of your installation process.
If you choose not to do the bonus question in (a), you can use any well-installed Hadoop cluster, e.g., the IE DIC, or the Hadoop cluster provided by the Google Cloud/AWS [5, 6, 7].
(b) [20 marks] Write a Hive script to perform exactly the same task as that of Q1 with the same datasets stored in the HDFS. Rerun the Pig script in this cluster and compare the performance between Pig and Hive in terms of overall run-time and explain your observation.
Hints:
● Hive will store its tables on HDFS and those locations needs to be bootstrapped:
$ hdfs dfs -mkdir /tmp
$ hdfs dfs -mkdir /user/hive/warehouse
$ hdfs dfs -chmod g+w /tmp
$ hdfs dfs -chmod g+w /user/hive/warehouse
● While working with the interactive shell (or otherwise), you should first test on a small subset of the data instead of the whole data set. Once your Hive commands/ scripts work as desired, you can then run them up on the complete data set.
 
 Q3 [30 marks + 10 Bonus marks]: Similar Users Detection in the MovieLens Dataset using Pig
Similar user detection has drawn lots of attention in the machine learning field which is aimed at grouping users with similar interests, behaviors, actions, or general patterns. In this homework, you will implement a similar-users-detection algorithm for the online movie rating system. Basically, users who rate similar scores for the same movies may have common tastes or interests and be grouped as similar users.
To detect similar users, we need to calculate the similarity between each user pair. In this homework, the similarity between a given pair of users (e.g. A and B) is measured as the total number of movies both A and B have watched divided by the total number of movies watched by either A or B. The following is the formal definition of similarity: Let M(A) be the set of all the movies user A has watched. Then the similarity between user A and user B is defined as:
𝑆𝑖𝑚𝑖𝑙𝑎𝑟𝑖𝑡𝑦(𝐴, 𝐵) = |𝑀(𝐴)∩𝑀(𝐵)| ...........(**) |𝑀(𝐴)∪𝑀(𝐵)|
where |S| means the cardinality of set S.
(Note: if |𝑀(𝐴)∪𝑀(𝐵)| = 0, we set the similarity to be 0.)
The following figure illustrates the idea:
Two datasets [3][4] with different sizes are provided by MovieLens. Each user is represented by its unique userID and each movie is represented by its unique movieID. The format of the data set is as follows:
<userID>, <movieID>
Write a program in Pig to detect the TOP K similar users for each user. You can use the
  
 cluster you built for Q1 and Q2 or you can use the IE DIC or one provided by the Google Cloud/AWS [5, 6, 7].
(a) [10 marks] For each pair of users in the dataset [3] and [4], output the number of movies they have both watched.
For your homework submission, you need to submit i) the Pig script and ii) the list of the 10 pairs of users having the largest number of movies watched by both users in the pair within the corresponding dataset. The format of your answer should be as follows:
<userID A>, <userID B>, <the number of movie both A and B have watched> //top 1 ...
<userID X>, <userID Y>, <the number of movie both X and Y have watched> //top 10
(b) [20 marks] By modifying/ extending part of your codes in part (a), find the Top-K (K=3) most similar users (as defined by Equation (**)) for every user in the datasets [3], [4]. If multiple users have the same similarity, you can just pick any three of them.
(c)
Hint:
1. In part (b), to facilitate the computation of the similarity measure as
defined in (**), you can use the inclusion-exclusion principle, i.e.
請加QQ:99515681  郵箱:99515681@qq.com   WX:codehelp 

標簽:

掃一掃在手機打開當前頁
  • 上一篇:&#160;ICT239 代做、代寫 java/c/c++程序
  • 下一篇:代寫COMP9334 Capacity Planning of Computer
  • 無相關信息
    昆明生活資訊

    昆明圖文信息
    蝴蝶泉(4A)-大理旅游
    蝴蝶泉(4A)-大理旅游
    油炸竹蟲
    油炸竹蟲
    酸筍煮魚(雞)
    酸筍煮魚(雞)
    竹筒飯
    竹筒飯
    香茅草烤魚
    香茅草烤魚
    檸檬烤魚
    檸檬烤魚
    昆明西山國家級風景名勝區
    昆明西山國家級風景名勝區
    昆明旅游索道攻略
    昆明旅游索道攻略
  • 短信驗證碼平臺 理財 WPS下載

    關于我們 | 打賞支持 | 廣告服務 | 聯系我們 | 網站地圖 | 免責聲明 | 幫助中心 | 友情鏈接 |

    Copyright © 2025 kmw.cc Inc. All Rights Reserved. 昆明網 版權所有
    ICP備06013414號-3 公安備 42010502001045

    爱情鸟第一论坛com高清免费_91免费精品国自产拍在线可以看_亚洲一区精品中文字幕_男人操心女人的视频
    <strike id="bfrlb"></strike><form id="bfrlb"><form id="bfrlb"><nobr id="bfrlb"></nobr></form></form>

        <sub id="bfrlb"><listing id="bfrlb"><menuitem id="bfrlb"></menuitem></listing></sub>

          <form id="bfrlb"></form>

            <form id="bfrlb"></form>

              <address id="bfrlb"></address>

              <address id="bfrlb"></address>
              国产精品日韩一区二区三区| 一区二区三区视频观看| 欧美好骚综合网| 一区二区欧美在线观看| 在线播放亚洲一区| 欧美日韩视频在线| 午夜国产欧美理论在线播放| 国外成人免费视频| 91久久精品国产| 精品成人一区二区三区四区| 一区二区激情视频| 国产有码一区二区| av成人免费在线观看| 欧美美女视频| 国产欧美精品日韩区二区麻豆天美| 久久久久99| 亚洲激情av在线| 国产精品久久网| 久久亚洲精品视频| 日韩一级二级三级| 国产综合18久久久久久| 久久久久久久久久看片| 久久在线免费视频| 亚洲午夜久久久久久尤物| 欧美人成免费网站| 欧美大片在线观看一区二区| 99国产精品视频免费观看| 一区二区av| 国产精品日本| 国产精品亚洲аv天堂网| 国产色视频一区| 国产精品久久久久一区二区| 久久综合狠狠综合久久综合88| 亚洲性xxxx| 日韩亚洲欧美成人一区| 国产模特精品视频久久久久| 亚洲视频二区| 久久久久久有精品国产| 香蕉成人啪国产精品视频综合网| 欧美天天在线| 欧美成人精品在线| 久久精品一区二区三区不卡牛牛| 亚洲欧洲精品成人久久奇米网| 久久精品一本| 欧美成人一品| 一本大道久久a久久精品综合| 亚洲经典在线| 影音先锋久久精品| 亚洲黄页视频免费观看| 欧美福利电影在线观看| 精品av久久707| 欧美精品一区二区三区一线天视频| 国产欧美精品va在线观看| 欧美大成色www永久网站婷| 久久综合久色欧美综合狠狠| 欧美与欧洲交xxxx免费观看| 欧美日韩国产综合一区二区| 欧美成人精品三级在线观看| 亚洲日本乱码在线观看| 欧美aⅴ一区二区三区视频| 亚洲最新合集| 香蕉久久a毛片| 亚洲永久网站| 国产精品综合不卡av| av成人黄色| 伊人久久久大香线蕉综合直播| 亚洲午夜高清视频| 亚洲狠狠丁香婷婷综合久久久| 在线观看福利一区| 国产精品视频免费在线观看| 国产精品高潮视频| 国产真实乱子伦精品视频| 激情懂色av一区av二区av| 欧美精品亚洲| 欧美激情一区二区三区高清视频| 亚洲欧美国产另类| 亚洲男女毛片无遮挡| 亚洲国产小视频在线观看| 欧美在线观看一区二区三区| 亚洲国产欧美精品| 欧美影院在线| 欧美日韩国产综合在线| 欧美极品一区二区三区| 国产精品亚发布| 国产精品人人爽人人做我的可爱| 一区二区高清视频在线观看| 亚洲欧美国产三级| 性色av一区二区三区红粉影视| 国产精品久久久久免费a∨| 亚洲人体一区| 中文精品99久久国产香蕉| 亚洲国产一区二区在线| 欧美一区二区在线免费播放| 欧美日韩大陆在线| 欧美小视频在线| 夜夜嗨一区二区| 日韩午夜在线观看视频| 国产精品日韩欧美一区二区| 亚洲天堂av电影| 国产一区二区成人久久免费影院| 巨胸喷奶水www久久久免费动漫| 欧美77777| 国产精品青草久久| 欧美日韩系列| 国产毛片精品视频| 欧美在线二区| 一区视频在线看| 午夜精品在线| 久久久夜精品| 美女任你摸久久| 亚洲网站视频福利| 亚洲理伦电影| 国产精品99久久久久久宅男| 欧美日韩国产欧| 一区二区精品在线| 亚洲专区一二三| 男女精品网站| 久久精品国产99精品国产亚洲性色| 午夜久久tv| 校园春色综合网| 国产精品久久久久秋霞鲁丝| 欧美日韩99| 欧美日韩国产区一| 亚洲一区尤物| 中日韩男男gay无套| 国产欧美高清| 一区视频在线看| 一本色道久久综合亚洲二区三区| 尤物九九久久国产精品的特点| 欧美美女福利视频| 欧美另类女人| 欧美视频免费| 在线播放国产一区中文字幕剧情欧美| 欧美欧美天天天天操| 欧美日韩理论| 欧美一区二区在线看| 久久精品国产亚洲一区二区三区| 亚洲男女自偷自拍图片另类| 欧美日韩免费高清| 欧美色图五月天| 欧美—级a级欧美特级ar全黄| 国产日韩在线一区二区三区| 久久天堂av综合合色| 久久福利视频导航| 国产欧美精品日韩区二区麻豆天美| 国产日韩在线看| 亚洲一区在线视频| 激情欧美日韩一区| 欧美电影在线免费观看网站| 中文在线资源观看网站视频免费不卡| 亚洲一区二区成人| 久久综合久久久| 欧美成人xxx| 亚洲视频免费| 在线精品视频一区二区三四| 欧美日韩裸体免费视频| 欧美精品入口| 国产精品美女在线观看| 国内综合精品午夜久久资源| 欧美久久在线| 亚洲国产精品小视频| 欧美日韩一区综合| 亚洲一品av免费观看| 欧美亚洲免费| 欧美成人精品高清在线播放| 亚洲成人影音| 亚洲毛片在线观看.| 一区二区三区欧美亚洲| 亚洲欧洲精品一区二区三区| 欧美一区二区黄| 亚洲视频在线观看视频| 国产精品一区二区三区久久久| 久久伊人精品天天| 正在播放亚洲一区| 国产一区二区中文字幕免费看| 国产精品白丝黑袜喷水久久久| 国内精品写真在线观看| 亚洲电影免费观看高清完整版在线| 亚洲国产婷婷| 国产欧美视频一区二区三区| 美女视频黄a大片欧美| 亚洲一区二区动漫| 亚洲欧美一区二区三区在线| 欧美在线免费观看| 国产精品xvideos88| 欧美电影免费观看| 久久精品国产69国产精品亚洲| 欧美成人r级一区二区三区| 老色鬼久久亚洲一区二区| 国产精品色一区二区三区| 一区在线免费观看| 欧美激情视频在线播放| 国产精品乱码久久久久久| 一本色道久久综合亚洲精品婷婷| 欧美jizz19性欧美| 噜噜噜在线观看免费视频日韩| 亚洲影音先锋| 欧美在线亚洲| 一区二区欧美在线观看| 欧美高清在线|