代写COMP3425、代做Data Mining、R代写、代做R程序设计 代写留学生 Statistics统计、回归、迭代|代做留学生Pro
COMP3425 Data Mining 2019
Assignment 2
Maximum marks 100
Weight 20% of the total marks for the course
Length Maximum of 10 pages, excluding cover sheet, bibliography and
appendices.
Layout A4 margin, at least 11 point type size, use of typeface, margins
and headings consistent with a professional style.
Submission deadline 9:00am, Monday, 20 May
Submission mode Electronic, via Wattle
Estimated time 15 hours
Penalty for lateness 100% after the deadline has passed
First posted: 1
st April, 9am
Last modified: 1
st April, 9am
Questions to: Wattle Discussion Forum
This assignment specification may be updated to reflect clarifications and modifications after it is
first issued.
It is strongly suggested that you start working on the assignment right away. You can submit as many
times as you like. Only the most recent submission at the due date will be assessed.
In this assignment, you are required to submit a single report in the form of a PDF file. You may also
attach supporting information (appendices) as one or more identified sections at the end of the
same PDF file. Appendices will not be marked but may be treated as supporting information to your
report. Please use a cover sheet at the front that identifies you as author of the work using your unumber
and name, and identifies this as your submission for COMP3425 Assignment 2. The cover
sheet and appendices do not contribute to the page limit. You are expected to write in a style
appropriate to a professional report. You may refer to http://www.anu.edu.au/students/learningdevelopment/writing-assessment/report-writing
for some useful stylistic advice. You are expected to
use the question and sub-question numbering in this assignment to identify the relevant answers in
your report.
No particular layout is specified, but you should follow a professional style and use no smaller than
11 point typeface and stay within the maximum specified page count. Page margins, heading sizes,
paragraph breaks and so forth are not specified but a professional style must be maintained. Text
beyond the page limit will be treated as non-existent. 2
This is a single-person assignment and should be completed on your own. Make certain you
carefully reference all the material that you use, although the nature of this assignment suggests few
references will be needed. It is unacceptable to cut and paste another author's work and pass it off
as your own. Anyone found doing this, from whatever source, will get a mark of zero for the
assignment and, in addition, CECS procedures for plagiarism will apply.
No particular referencing style is required. However, you are expected to reference conventionally,
conveniently, and consistently. References are not included in the page limit. Due to the context in
which this assignment is placed, you may refer to the course notes or course software where
appropriate (e.g. “For this experiment Rattle was used”), without formal reference to original
sources, unless you copy text which always requires a formal reference to the source.
An assessment rubric is provided. The rubric will be used to mark your assignment. You are advised
to use it to supplement your understanding of what is expected for the assignment and to direct
your effort towards the most rewarding parts of the work.
Your assignment submission will be treated confidentially. It will be available to ANU staff involved
in the course for the purposes of marking.
Task
You are to complete the following exercises. For simplicity, the exercises are expressed using the
assumption that you are using Rattle, however you are free to use R directly or any other data
mining platform you choose that can deliver the required functions. You are expected, in your own
words, to interpret selected tool output in the context of the learning task. Write just what is needed
to explain the results you see. Similarly, you should describe the methods used in terms of the
language of data mining, not in the terms of commands you typed or buttons you selected.
1. Platform
Briefly describe the platform for your experiments in terms of memory, CPU, operating system, and
software that you use for the exercises. If your platform is not consistent throughout, you must
describe it for each exercise. This is to ensure your results are reproducible.
2. Data
Look at the pairwise correlation amongst all the numeric variables using Pearson product-moment
correlation.
(a) Explain why you would expect DayofYear and WeekofYear to be highly correlated.
(b) Qualitatively describe the correlations amongst the variables High, Low, Open, Close and
Volume. Explain what you see in terms of the source of the data.3
3. Association mining: What factors affect volume of sales?
(a) Compute and give the 5-number summary for Volume. Qualitatively describe what it tells you
about Volume.
(b) For this exercise, bin Volume into 5 categories using Quantiles. Why is quantile binning
appropriate for association mining with Volume?
When you have completed this question 3, remove the extra variable you created so they do not
interfere with other exercises.
(c) Generate association rules, adjusting min support and min confidence parameters as you need.
What parameters do you use? Bearing in mind we are looking for insight into what factors affect
Volume, find 3 interesting rules, and explain both objectively and subjectively why they are
interesting.
(d) Comment on whether, in general, association mining could be a useful technique on this data.
4. Study a very simple classification task
Aim to build a model to classify Change. Use Change as the target class and set every other variable
as Input (independent). Using sensible defaults for model parameters is fine for this exercise where
we aim to compare methods rather than optimise them.
(a) This should be a very easy task for a learner. Why? Hint: Think how Change is defined.
(b) Train each of a Linear, Decision tree, SVM and Neural Net classifier, so you have 4 classifiers.
Hint: Because the dataset is large, begin with a small training set, 20%, and where run-time speeds
are acceptable, move up to a 70% training set. Evaluate each of these 4 classifiers, using a confusion
matrix and interpreting the results for the context of the learning task.
(c) Inspect the models themselves where that is possible to assist in your evaluation and to explain
the performance results. Which learner(s) performed best and why?
5. Predict a Numeric Variable
One investment strategy could rely on the previous day’s price to predict the opening price for a
stock, enabling you to place a buy or sell offer overnight ready for the next day. To predict the
opening price for a day, you cannot use any of the other prices or Volume for that same day as that
information is not available until the close of the day, when it is too late. So, using the variables you
have in the dataset, but ignoring all of High, Low, Close, Volume, Close-Open, Change, High-Low, and
HMLOL, aim to predict the opening price Open using the previous day’s closing price PriorClose. and
the date and stock-related variables. Use a regression tree or a neural net.
(a) Explain which you chose of a regression tree or neural net and justify your choice.
(b) Train your chosen model and tune by setting controllable parameters to achieve a reasonable
performance. Explain what parameters you varied and how, and the values you chose finally.
(c) Assess the performance of your best result using the subjective and objective evaluation
appropriate for the method you chose, and justify why you settled with that result. 4
6. More Complex Classification
An alternative investment strategy might be to predict where there will be a big proportional change
in the price over a day, once the day has opened, and so a good opportunity for a short term gain
(or loss) that day.
(a) Transform HMLOL to a categoric class variable by binning into 2 classes, using a k-means
clustering of the HMLOL. The skewed distribution of HMLOL is helpful here as the higher values we
need for investing are relatively rare. When you have completed this question 6, remove the extra
variable you created so it does not interfere with other exercises. Now, be sure to ignore most of
the current day price and volume variables, that is ignore all of High, Low, Close, Volume, CloseOpen,
Change, High-Low, and HMLOL. Explain why HMLOL should be ignored. This time, use the
Open price, PriorClose and all the date and company description variables for learning.
Initially, use a small training set, 20%, and where run-time speeds are acceptable, experiment with a
larger training set. Explain how you will partition the available dataset to train and validate
classification models below.
(b) Train a Decision Tree Classifier. You will need to adjust default parameters to obtain optimal
performance. State what parameters you varied and (briefly) their effect on your results. Evaluate
your optimal classifier using the error matrix, ROC, and any quality information specific to the
classifier method.
(c) Train an SVM Classifier. Then proceed as for (b) Decision Tree above.
(d) Train a Neural Net classifier. Then proceed as for (b) Decision Tree above.
7. Clustering
(a) Restore the dataset to its original distributed form, removing any new variables you have
constructed above. For clustering, use only the five raw variables, Date, Open, High, Low and Volume
and remove all of the others.
Experiment with clustering using the k-means algorithm. Rescale the variables to fall in the range 0-
1 prior to clustering. Use the full dataset for clustering (do not partition) by building cluster models
for each of k= 2, 5 and the recommended default (i.e. squareroot (n/2) for dataset of size n)
clusters. Choose your preferred k and it’s cluster model for k-means to answer the following.
(a) Justify your choice of k as your preferred (Hint: have look at parts b-d below for each cluster
model).
(b) Calculate the sum of the within-cluster-sum–of-squares for your chosen model. The withincluster-sum–of-squares
is the sum of the squares of the Euclidean distance of each object from its
cluster mean. Discuss why this is interesting.
(c) Look at the cluster centres for each variable. Using this information, discuss qualitatively how
each cluster differs from the others.
(d) Use a scatterplot to plot (a sample of) the objects projected on to each combination of 2
variables with objects mapped to each cluster by colour (Hint: The Data button on Rattle’s Cluster 5
tab can do this). Describe what you can see as the major influences on clustering. Include the
image in your answer.
8. Qualitative Summary of Findings (approx 1/2 page)
Would you use these results to advise your investment decisions?
Comparatively evaluate the techniques you have used and their suitability or not for mining this
data. This should be a qualitative opinion that draws on what you have found already doing the
exercises above. For example, what can you say about training and classification speeds, the size or
other aspects of the training data, or the predictive power of the models built? Finally, what else
would you propose to investigate to assist your investment decisions?6
Assessment Rubric
This rubric will be used to mark your assignment. You are advised to use it to supplement your understanding of what is expected for the assignment and to
direct your effort towards the most rewarding parts of the work. Your assignment will be marked out of 100, and marks will be scaled back to contribute to
the defined weighting for assessment of the course.
Review
Criteria
Max
Mark
Exemplary Excellent Good Acceptable Unsatisfactory
1. Platform & 2.
Data
10 9-10
1.Platform description
complete (memory, CPU,
operating system, software).
2. All required correlations
(Year/DayofYear & all pairs
of High, Low, Open, Close
and Volume) clearly
explained in terms of the
data domain, in the correct
directions and for correct
reasons demonstrating an
understanding of the data.
7-8
1. Platform description
complete.
2a partial or unclear
2b partial or unclear
2b partial explanation
5-6
1. Platform description
complete.
2a attempt but with
correlation direction wrong
2b Partial description of 4
variables or unclear
2b Partial explanation
0-4
1. Platform description
incomplete.
2a. correlation reason
missing or unrelated to
DayofYear and WeekofYear
2b. Description unrelated to
High, Low, Open, Close and
Volume
2b. Explanation unrelated to
data source
3. Association
mining
10 9-10
Answers demonstrate deep
understanding of association
mining, by the careful
selection of interesting and
differentiated rules and clear
rationale for interestingness.
Comment shows original and
insightful analysis of
association mining on the
problem.
7-8
a 5-number summary
complete
b 5-number explanation for
Volume clear
b Binning understood
c Support and confidence
clear
c 3 interesting rules given
c objective interestingness is
given for all 3
c subjective interestingness
attempted
d Comment makes sense
5-6
a 5number summary ok
b 5-number explanation for
Volume poor
b Binning misunderstood
c Support or confidence not
clear
c < 3 interesting rules given
c objective interestingness is
incomplete
c subjective interestingness
is incomplete
d Comment is cursory or offtrack
0-4
Required information not
provided and/or incorrect or
misleading, demonstrating
lack of engagement with the
problem7
Review
Criteria
Max
Mark
Exemplary Excellent Good Acceptable Unsatisfactory
4. Simple
classification
10 9-10
Deep understanding of the 4
models demonstrated
through analysis of
performance on the change
task.
7-8
a correctly explains why
definition of change makes it
seem easy
b 4 confusion matrixes given
b confusion matrixes
explained in terms of the
data and the method and
the model learnt.
c some evidence of
understanding what the
models are doing
c reasoning for comparative
performance demonstrating
understanding of the
methods behind them
5-6
a partially explains why
definition of change makes it
seem easy
b 4 confusion matrixes given
b confusion matrixes
explained at face value only
c weak understanding of
learnt models
c comparative performance
only cursorily presented
c reason for comparative
performance is shallow
0-4
a inadequate explanation
b confusion matrix missing
b confusion matrix
misunderstood
c Interpretation of confusion
matrix missing
c no apparent understanding
of what the models are doing
c missing or unexplained
comparative analysis
5. Prediction 20 17-20
Approach to problem
demonstrates effort to
produce good results and a
deep understanding of the
relative benefits of the 2
models in the context of the
problem domain.
Results are interpreted in
the context of the problem
domain.
14-16
a justification for choice
shows understanding of the
comparative benefits of
each and extensive
experiments with
performance.
b parameter variations
shows a combination of
experimentation and
understanding of the
parameters with justification
for stopping at selected
parameters.
c several subjective and
objective evaluation
measures used as
appropriate to method
chosen
c justification for stopping
demonstrates awareness of
appropriateness of best
12-13
a justification for choice
shows understanding of the
comparative benefits of
each and experiments with
performance.
b parameter variations
shows a combination of
experimentation and
understanding of the
parameters
c multiple subjective and
objective evaluation
measures used as
appropriate to method
chosen
c justification for stopping
demonstrates awareness of
appropriateness of best
result
10-11
a justification for choice
shows some understanding
of the comparative benefits
of each or experiments with
performance.
b parameter variation
demonstrates some
experimentation
c cursory evaluation given
c justification for stopping
perfunctory
0-9
a weak justification for
choice
b parameter variation
insufficient
c evaluation fails to
demonstrate effort or
understanding of evaluation
c justification for stopping
effectively absent8
Review
Criteria
Max
Mark
Exemplary Excellent Good Acceptable Unsatisfactory
result and scope of potential
for further improvement
6. Complex
Classification
30 26-30
Exemplary use of
classification models with
comprehensive and fit-forpurpose
performance
analysis on the problem
22-25
a explanation correct
b,c,d parameter variation
clear and extensive
demonstrating
understanding of effect in all
3 methods
b.c.d error matrix and ROC
correctly interpreted
in all 3 methods
b,c,d extensive use of
specific evaluation methods
used and significance clearly
explained in all 3 methods
18-21
a explanation correct
a satisfactory approach to
dataset partitioning
b parameter variation clear
and sufficient for good
results
b error matrix correctly
interpreted
b ROC correctly interpreted
b some specific evaluation
methods used
c parameter variation clear
and sufficient for good
results
c error matrix correctly
interpreted
c ROC correctly interpreted
c some specific evaluation
methods used
d parameter variation clear
and sufficient for good
results
d error matrix correctly
interpreted
d ROC correctly interpreted
d some specific evaluation
methods used
15-17
a explanation correct
a satisfactory approach to
dataset partitioning
b parameter variation
perfunctory
b error matrix given
b ROC given
b few specific evaluation
methods used
c parameter variation
perfunctory
c error matrix given
c ROC given
c few specific evaluation
methods used
d parameter variation
perfunctory
d error matrix given
d ROC given
d few specific evaluation
methods used
0-14
a explanation incorrect
a unsound use of
training/testing/validation
data
b no parameter variation
b no error matrix
b no or faulty ROC
b specific evaluation
methods missing
c no parameter variation
c no error matrix
c no or faulty ROC
c specific evaluation
methods missing
d no parameter variation
d no error matrix
d no or faulty ROC
d specific evaluation
methods missing9
Review
Criteria
Max
Mark
Exemplary Excellent Good Acceptable Unsatisfactory
7. Clustering 10 9-10
The application of k-means
algorithm to the dataset and
its evaluation demonstrates
exemplary understanding of
the algorithm, its evaluation,
and its limitations.
Suitable evaluation methods
or clustering experiments
beyond those required here
may be used.
7-8
a justification convincing
b measure calculated
correctly. Discussion
recognises value and
limitations
c discussion on centres
reflects numeric results and
emphasises the interesting
parts that relate to the
significance in domain terms
d correct image included and
description shows
understanding linked to data
domain
5-6
a justification offered but
not clear or unconvincing
b measure calculated
correctly
c discussion on centres
reflects numeric results
d correct image included
0-4
Clustering experimentation
and discussion inadequate
8. Qualitative
Summary
10 9-10
Many aspects of evaluation
are discussed and a clear
conclusion is drawn, with
direct reference to potential
goals of the domain of the
data.
Proposal for further
investigation demonstrates
creativity and thoughtful
engagement with the
problem, clearly building on
the work reported.
8
A clear conclusion is drawn
from the work reported and
a defended proposal for
further investigation is
proposed, with clear links to
both the work reported and
the domain of application.
7
A rounded, balanced
summary of the work is
presented with a justified
proposal given.
6
A summary of the work is
presented and a proposal
made.
0-4
Answer does not
demonstrate adequate
engagement with the
problem nor a qualitative
understanding of the work
reported.

热门主题

课程名

mktg2509 csci 2600 38170 lng302 csse3010 phas3226 77938 arch1162 engn4536/engn6536 acx5903 comp151101 phl245 cse12 comp9312 stat3016/6016 phas0038 comp2140 6qqmb312 xjco3011 rest0005 ematm0051 5qqmn219 lubs5062m eee8155 cege0100 eap033 artd1109 mat246 etc3430 ecmm462 mis102 inft6800 ddes9903 comp6521 comp9517 comp3331/9331 comp4337 comp6008 comp9414 bu.231.790.81 man00150m csb352h math1041 eengm4100 isys1002 08 6057cem mktg3504 mthm036 mtrx1701 mth3241 eeee3086 cmp-7038b cmp-7000a ints4010 econ2151 infs5710 fins5516 fin3309 fins5510 gsoe9340 math2007 math2036 soee5010 mark3088 infs3605 elec9714 comp2271 ma214 comp2211 infs3604 600426 sit254 acct3091 bbt405 msin0116 com107/com113 mark5826 sit120 comp9021 eco2101 eeen40700 cs253 ece3114 ecmm447 chns3000 math377 itd102 comp9444 comp(2041|9044) econ0060 econ7230 mgt001371 ecs-323 cs6250 mgdi60012 mdia2012 comm221001 comm5000 ma1008 engl642 econ241 com333 math367 mis201 nbs-7041x meek16104 econ2003 comm1190 mbas902 comp-1027 dpst1091 comp7315 eppd1033 m06 ee3025 msci231 bb113/bbs1063 fc709 comp3425 comp9417 econ42915 cb9101 math1102e chme0017 fc307 mkt60104 5522usst litr1-uc6201.200 ee1102 cosc2803 math39512 omp9727 int2067/int5051 bsb151 mgt253 fc021 babs2202 mis2002s phya21 18-213 cege0012 mdia1002 math38032 mech5125 07 cisc102 mgx3110 cs240 11175 fin3020s eco3420 ictten622 comp9727 cpt111 de114102d mgm320h5s bafi1019 math21112 efim20036 mn-3503 fins5568 110.807 bcpm000028 info6030 bma0092 bcpm0054 math20212 ce335 cs365 cenv6141 ftec5580 math2010 ec3450 comm1170 ecmt1010 csci-ua.0480-003 econ12-200 ib3960 ectb60h3f cs247—assignment tk3163 ics3u ib3j80 comp20008 comp9334 eppd1063 acct2343 cct109 isys1055/3412 math350-real math2014 eec180 stat141b econ2101 msinm014/msing014/msing014b fit2004 comp643 bu1002 cm2030
联系我们
EMail: 99515681@qq.com
QQ: 99515681
留学生作业帮-留学生的知心伴侣!
工作时间:08:00-21:00
python代写
微信客服:codinghelp
站长地图