{"id":22146,"date":"2024-11-06T07:43:44","date_gmt":"2024-11-06T02:13:44","guid":{"rendered":"http:\/\/www.sachdevajk.in\/?p=22146"},"modified":"2024-11-06T07:43:44","modified_gmt":"2024-11-06T02:13:44","slug":"cluster-analysis-and-anova-of-ipad","status":"publish","type":"post","link":"http:\/\/www.sachdevajk.in\/?p=22146","title":{"rendered":"Cluster Analysis and Anova of Ipad"},"content":{"rendered":"<p style=\"text-align: center\">Market Research Assignment-2<\/p>\n<p style=\"text-align: center\">Cluster Analysis and Anova of Ipad<\/p>\n<p style=\"text-align: center\">M2 Batch<br \/>Group Members:<\/p>\n<p style=\"text-align: center\">1. Samaya Rayaprolu<br \/>2. Shatakshi<br \/>3. Sneha Yadav<br \/>4. Paridhi Gangrade<\/p>\n<p>Case Processing Summary<br \/>Cases<br \/>Valid<br \/>N<br \/>Missing<br \/>N<br \/>Total<br \/>N<br \/>Percent Percent<br \/>Percent<br \/>52 98.1 1 1.9 53 100.0<br \/>a. Squared Euclidean Distance used<br \/>b. Average Linkage (BetweenGroups)<\/p>\n<p>Introduction and Objectives:<\/p>\n<p>This analysis employs cluster analysis and ANOVA to assess variations in characteristics (e.g., engine, comfort, storage) among cases in a dataset of 53 observations. The goal is to identify natural groupings and evaluate the degree of distinct on between these groups across various characteristics. The document uses Average Linkage (Between Groups) as the clustering method and Squared Euclidean Distance to calculate the distance between data points. These choices aim to build clusters based on similarities across multiple features and provide a detailed statistical view of the resulting clusters.<br \/>The process includes:<br \/>\u2022 Cluster Analysis: To examine how data points are grouped and understand the structural similarities within the dataset.<br \/>\u2022 ANOVA: To measure if significant differences exist between clusters on each characteris c,<br \/>providing an understanding of which features most differen ate the groups.<\/p>\n<p>Agglomera on Schedule<br \/>Cluster Combined<br \/>Stage Cluster First Appears<br \/>Cluster 1 Cluster 2 Cluster 1 Cluster 2<br \/>Stage<br \/>Coefficients<br \/>Next Stage<br \/>1 3 27 .000 0 0 29<br \/>2 36 44 3 20 43 4 10 29 5 17 53 6 22 26 7 4 23 8 4 36 9 46 51 10 2 48 11 24 38 12 6 21 13 12 17 14 37 50 15 22 28 16 1 12 17 41 46 18 14 39 19 9 16 20 41 47 21 34 37 22 20 24 23 1 5 3.000 3.000 3.000 4.000 4.000 4.000 4.500 5.000 5.000 5.000 5.000 5.000 6.000 6.000 6.000 6.500 7.000 7.000 7.667 8.000 8.000 8.000 0 0 8<br \/>0 0 22<br \/>0 0 27<br \/>0 0 13<br \/>0 0 15<br \/>0 0 8<br \/>7 2 25<br \/>0 0 17<br \/>0 0 33<br \/>0 0 22<br \/>0 0 25<br \/>0 5 16<br \/>0 0 21<br \/>6 0 32<br \/>0 13 23<br \/>0 9 20<br \/>0 0 38<br \/>0 0 24<br \/>17 0 31<br \/>0 14 32<br \/>3 11 31<br \/>16 0 29<br \/>24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 9 18 4 6 13 35 10 30 42 45 1 3 4 8 20 41 22 34 2 4 15 19 7 11 20 22 1 10 9 14 7 20 13 15 42 52 7 49 2 9 7 25 2 7 33 42 8.500 8.500 9.000 9.500 10.000 10.200 10.333 10.875 10.889 11.357 12.000 12.000 12.417 12.619 13.833 15.500 16.000 17.000 17.063 17.267 18.529 18.952 20.333 19 0 38<br \/>8 12 30<br \/>0 0 40<br \/>4 0 37<br \/>0 0 41<br \/>23 1 37<br \/>25 0 33<br \/>22 20 36<br \/>15 21 36<br \/>10 30 43<br \/>0 0 40<br \/>0 0 39<br \/>31 32 39<br \/>29 27 48<br \/>24 18 43<br \/>35 36 42<br \/>26 34 47<br \/>28 0 46<br \/>39 0 44<br \/>33 38 45<br \/>42 0 45<br \/>43 44 47<br \/>0 41 51<br \/>47 2 13 21.375 45 40 48<br \/>48 1 2 26.561 37 47 49<br \/>49 1 32 32.435 48 0 50<br \/>50 1 40 36.021 49 0 51<br \/>51 1 33 47.031 50 46 0<\/p>\n<p>Detailed Cluster Analysis<br \/>Case Processing Summary<br \/>The Case Processing Summary reveals that out of 53 cases, 52 are valid, and only 1 case is missing, ensuring that almost the enter dataset was considered. With a minimal number of missing cases, the analysis remains statistically robust, supportung reliable interpretations of the clustering and ANOVA results.<br \/>Agglomera on Schedule<br \/>The Agglomera on Schedule is central to understanding the clustering process, showing how clusters merge stepbystep based on increasing distance between them. Each row includes:<br \/>\u2022 Stages: Clusters are combined in successive stages. Lower stages involve more similar clusters,<br \/>while higher stages combine clusters with progressively greater dissimilari es.<br \/>\u2022 Coefficients: Coefficients reflect the Euclidean distances at which clusters merge, offering insights<br \/>into similarity.<\/p>\n<p>Interpretation of Key Stages:<br \/>1. Early Stages: In the ini al steps, clusters merge at low coefficients, reflecting high internal similarity. For instance:<br \/>o At Stage 1, clusters 3 and 27 merge with a coefficient of 0.000, meaning these clusters<br \/>contain data points that are either iden cal or nearly identical.<br \/>o By Stage 5, clusters 17 and 53 combine with a coefficient of 4.000, indica ng moderate<br \/>similarity but slightly more variance than earlier stages.<br \/>2. Middle Stages: By around Stage 25, the merging coefficient reaches 8.500, implying the clusters being combined are less homogeneous than those in earlier stages. This \u00a0indicates that the clustering process has begun grouping data points with larger differences, creating broader clusters that are less specific but still retain identifiable similarities.<br \/>3. Later Stages: Towards the end, such as Stage 50 with a coefficient of 36.021, the clusters merged are highly dissimilar, showing that most natural clusters have already formed. These later stages<br \/>reflect forced combina ons of groups that are more heterogeneous, typical in hierarchical<br \/>clustering as the process attempts to merge all data into fewer clusters.<\/p>\n<p>Initial and Final Cluster Centres:<\/p>\n<p>The Cluster Centres sec on provides insight into the values of each variable at the beginning and end of the clustering process, representing the average values for each characteristic in each cluster. This helps understand the underlying structure and distinctions among clusters.<br \/>\u2022 Initial Cluster Centres: Early cluster centres show the star ng values for each variable, establishing the initial configurations based on characteristics like &#8220;engine&#8221; and &#8220;storage.&#8221; For instance, one<br \/>cluster may ini ally priori ze &#8220;engine&#8221; at a higher level, indica ng a subgroup where engine<br \/>performance is a stronger focal point.<br \/>\u2022 Final Cluster Centres: As the clustering process progresses, the centres for each cluster become<br \/>more stable. By the final clusters, we see how each group has se led around specific average<br \/>values for each characteristic. This stabilization process implies that clusters are welldefined by<br \/>these final values, offering a clearer picture of the traits defining each cluster.<\/p>\n<p>Cluster Membership and Distribution:<\/p>\n<p>The Cluster Membership table assigns each case to a specific cluster and provides the distance of each case from its respec ve cluster centre.<\/p>\n<p>With 31 cases in Cluster 1 and 21 in Cluster 2, we observe a slightly uneven distribution of cases:<\/p>\n<p>\u2022 Cluster 1 contains a larger number of cases, suggesting it may capture a broader or more common.<br \/>set of characteristics.<br \/>\u2022 Cluster 2 is smaller, which could indicate a more specialized subset of cases with unique traits.<br \/>This membership structure allows us to see the grouping dynamics based on similarity in a ributes and helps visualize how the dataset divides into naturally occurring groups. The distances indicate the relative closeness of cases to their respective centres, providing insight into cluster cohesion.<\/p>\n<p>Distances Between Final Cluster Centres:<\/p>\n<p>The Distance between Final Cluster Centres measures the separation between clusters, with a value of 3.476, which is moderately large. This distance suggests:<br \/>\u2022 Moderate Distinction: A reasonable degree of differentiation on exists between the clusters, indicating.<br \/>that each cluster has a unique set of characteriscs that make it dis nct from the other.<br \/>\u2022 Interpretation: In hierarchical clustering, the distance between clusters can help assess the effectiveness of the clustering process. <br \/>Here, 3.476 implies that while some overlap may exist, each cluster retains a specific identity.<\/p>\n<p>ANOVA<br \/>Sum of Squares df Mean Square F Sig.<br \/>Display Between Groups 11.593 4 2.898 1.879 .130<br \/>Within Groups 72.484 47 1.542<br \/>Total 84.077 51<br \/>Operating System Between Groups 13.627 4 3.407 4.549 .003<br \/>Within Groups 35.200 47 .749<br \/>Total 48.827 51<br \/>Battery Life Between Groups 7.090 4 1.772 2.193 .084<br \/>Within Groups 37.987 47 .808<br \/>Total 45.077 51<br \/>Processor Between Groups 23.524 4 5.881 5.470 .001<br \/>Within Groups 50.534 47 1.075<br \/>Total 74.058 51<br \/>camera Between Groups 16.436 4 4.109 4.366 .004<br \/>Within Groups 44.237 47 .941<br \/>Total 60.673 51<br \/>storage Between Groups 11.679 4 2.920 3.225 .020<br \/>Within Groups 42.552 47 .905<br \/>Total 54.231 51<br \/>Connectivity Between Groups 6.483 4 1.621 1.366 Within Groups 55.748 47 1.186<br \/>Total 62.231 51<br \/>Apple Pencil Support Between Groups 13.221 4 3.305 3.574 Within Groups 43.471 47 .925<br \/>Total 56.692 51<br \/>Face ID Between Groups 34.262 4 8.566 7.634 &lt;.001<br \/>Within Groups 52.738 47 1.122<\/p>\n<p>Total 87.000 51<br \/>Interpretation of ANOVA Results<br \/>1. Display<br \/>F-Statistic: 1.879 with a significance level (Sig.) of 0.130, indicating no significant difference in<br \/>&#8220;Display&#8221; across clusters. This variable does not appear to contribute meaningfully to group distinctions.<br \/>2. Operating System<br \/>F-Statistic: 4.549 with a p-value of 0.003, indicating a statistically significant difference in &#8220;Operating System&#8221; across clusters. This suggests that the operating system attribute varies across clusters, potentially contributing to the differences between groups.<br \/>3. Battery Life<br \/>F-Statistic: 2.193 with Sig. = 0.084, which is not below the 0.05 threshold, but close. This hints that battery life might show some differentiation across clusters, though not strongly significant.<br \/>4. Processor<br \/>F-Statistic: 5.470 with Sig. = 0.001, indicating a significant difference in &#8220;Processor&#8221; across clusters. This attribute likely contributes meaningfully to the distinctions between clusters.<br \/>5. Camera<br \/>F-Statistic: 4.366 with Sig. = 0.004, which is significant, suggesting that &#8220;Camera&#8221; varies significantly across clusters. This differentiation likely plays a role in distinguishing between the clusters.<br \/>6. Storage<br \/>F-Statistic: 3.225 with Sig. = 0.020, showing a significant difference in &#8220;Storage&#8221; across clusters. This result indicates that storage attributes might contribute to group distinctions.<br \/>7. Connectivity<br \/>F-Statistic: 1.366 with Sig. = 0.260, indicating no significant difference in &#8220;Connectivity&#8221; across<br \/>clusters. Connectivity does not appear to contribute to cluster distinctions.<br \/>8. Apple Pencil Support<br \/>F-Statistic: 3.574 with Sig. = 0.013, suggesting a significant difference in &#8220;Apple Pencil Support&#8221; across clusters, indicating that this feature may play a role in cluster differentiation.<br \/>9. Face ID<br \/>F-Statistic: 7.634 with Sig. &lt; 0.001, showing a highly significant difference in &#8220;Face ID&#8221; across clusters.<br \/>This attribute strongly differentiates the groups.<\/p>\n<p>Summary of ANOVA Results:<\/p>\n<p>The ANOVA results indicate significant differences across clusters for &#8220;Operating System,&#8221; &#8220;Processor,&#8221; &#8220;Camera,&#8221; &#8220;Storage,&#8221; &#8220;Apple Pencil Support,&#8221; and &#8220;Face ID,&#8221; which are likely key contributors to cluster distinctions. Attributes such as &#8220;Display,&#8221; &#8220;Battery Life,&#8221; and &#8220;Connectivity&#8221; show little to no meaningful differentiation across groups. This suggests that specific features, particularly Face ID and Processor, may be primary factors in defining the clusters, while others play a lesser role.<\/p>\n<p>Display Operating System Battery Life Processor camera storage Connectivity 95% Confidence Interval<br \/>Point Estimate<br \/>Lower Upper<br \/>Eta-squared .138 .000 .266<br \/>Epsilon-squared .065 -.085 .203<br \/>Omega-squared Fixed-effect .063 -.083 .200<br \/>Omega-squared Random-effect .017 -.020 .059<br \/>Eta-squared .279 .041 .417<br \/>Epsilon-squared .218 -.040 .368<br \/>Omega-squared Fixed-effect .214 -.040 .363<br \/>Omega-squared Random-effect .064 -.010 .125<br \/>Eta-squared .157 .000 .289<br \/>Epsilon-squared .086 -.085 .228<br \/>Omega-squared Fixed-effect .084 -.083 .225<br \/>Omega-squared Random-effect .022 -.020 .068<br \/>Eta-squared .318 .069 .454<br \/>Epsilon-squared .260 -.010 .407<br \/>Omega-squared Fixed-effect .256 -.010 .402<br \/>Omega-squared Random-effect .079 -.002 .144<br \/>Eta-squared .271 .036 .409<br \/>Epsilon-squared .209 -.046 .359<br \/>Omega-squared Fixed-effect .206 -.045 .355<br \/>Omega-squared Random-effect .061 -.011 .121<br \/>Eta-squared .215 .004 .353<br \/>Epsilon-squared .149 -.081 .298<br \/>Omega-squared Fixed-effect .146 -.079 .294<br \/>Omega-squared Random-effect .041 -.019 .094<br \/>Eta-squared .104 .000 .221<br \/>Epsilon-squared .028 -.085 .155<br \/>Omega-squared Fixed-effect .027 -.083 .153<br \/>Omega-squared Random-effect .007 -.020 .043<br \/>Apple Pencil Support Eta-squared .233 .013 .372<br \/>Epsilon-squared .168 -.071 .318<br \/>Omega-squared Fixed-effect .165 -.069 .314<br \/>Omega-squared Random-effect .047 -.016 .103<br \/>Face ID Eta-squared .394 .135 .522<br \/>Epsilon-squared .342 .061 .481<br \/>Omega-squared Fixed-effect .338 .060 .476<br \/>Omega-squared Random-effect .113 .016 .185<\/p>\n<p>a. Eta-squared and Epsilon-squared are estimated based on the fixed-effect model.<br \/>b. Negative but less biased estimates are retained, not rounded to zero.<\/p>\n<p>Effect Size Interpretation:<\/p>\n<p>Effect sizes provide additional insights by quantifying the extent to which variability in each characteristic is due to clustering:<\/p>\n<p>&#8211; Eta-Squared: This measure ranges from low to moderate for most characteristics, indicating minimal to moderate cluster-based variance. For example, &#8220;Face ID&#8221; has an eta-squared of 0.394, suggesting it accounts for a substantial portion of cluster variance, whereas attributes like &#8220;Connectivity&#8221; have lower eta-squared values, indicating they contribute minimally to clustering differences.<\/p>\n<p>&#8211; Epsilon-Squared and Omega-Squared: These measures adjust eta-squared to provide less biased estimates. Negative values for some characteristics reflect that these variables do not meaningfully vary by cluster, suggesting limited or non-significant contributions to the clustering structure.<\/p>\n<p>Conclusions and Implications:<\/p>\n<p>This analysis offers a comprehensive examination of clusters within the dataset, with clustering insights and statistical evaluations through ANOVA. <br \/>Key takeaways include:<\/p>\n<p>&#8211; Cluster Similarity and Distinctiveness: Clusters initially show clear distinctions, as seen in the early<br \/>agglomeration stages. However, later stages combine increasingly dissimilar points, resulting in broader and more heterogeneous clusters.<\/p>\n<p>&#8211; ANOVA and Effect Sizes: Significant F-values and moderate to high effect sizes for variables like<br \/>&#8220;Operating System,&#8221; &#8220;Processor,&#8221; &#8220;Camera,&#8221; &#8220;Storage,&#8221; &#8220;Apple Pencil Support,&#8221; and &#8220;Face ID&#8221; suggest these attributes contribute notably to the cluster distinctions. In contrast, attributes with lower effect sizes, such as &#8220;Display&#8221; and &#8220;Connectivity,&#8221; show limited differentiation across clusters.<\/p>\n<p>&#8211; Implications for Further Research: Future analyses could consider alternative clustering methods, additional variables, or larger sample sizes to detect more nuanced subgroups, enhancing insights into the factors driving subgroup characteristics.<\/p>\n<p>In conclusion, while clustering helps organize the dataset and offers a structural overview, significant differences are primarily observed for a select set of attributes. This highlights the importance of these variables in defining clusters and suggests the potential benefit of further refinement in clustering approaches or variable selection to yield more detailed subgroup characteristics.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Market Research Assignment-2 Cluster Analysis and Anova of Ipad M2 BatchGroup Members: 1. Samaya Rayaprolu2. Shatakshi3. Sneha Yadav4. Paridhi Gangrade Case Processing SummaryCasesValidNMissingNTotalNPercent PercentPercent52 98.1 1 1.9 53 100.0a. Squared Euclidean Distance usedb. Average Linkage (BetweenGroups) Introduction and Objectives: This analysis employs cluster analysis and ANOVA to assess variations in characteristics (e.g., engine, comfort, storage)&hellip; <a class=\"more-link\" href=\"http:\/\/www.sachdevajk.in\/?p=22146\">Continue reading <span class=\"screen-reader-text\">Cluster Analysis and Anova of Ipad<\/span><\/a><\/p>\n","protected":false},"author":139647,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[39],"tags":[384,446],"class_list":["post-22146","post","type-post","status-publish","format-standard","hentry","category-marketing","tag-itm","tag-itm-kharghar","entry"],"_links":{"self":[{"href":"http:\/\/www.sachdevajk.in\/index.php?rest_route=\/wp\/v2\/posts\/22146","targetHints":{"allow":["GET"]}}],"collection":[{"href":"http:\/\/www.sachdevajk.in\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"http:\/\/www.sachdevajk.in\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"http:\/\/www.sachdevajk.in\/index.php?rest_route=\/wp\/v2\/users\/139647"}],"replies":[{"embeddable":true,"href":"http:\/\/www.sachdevajk.in\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=22146"}],"version-history":[{"count":1,"href":"http:\/\/www.sachdevajk.in\/index.php?rest_route=\/wp\/v2\/posts\/22146\/revisions"}],"predecessor-version":[{"id":22147,"href":"http:\/\/www.sachdevajk.in\/index.php?rest_route=\/wp\/v2\/posts\/22146\/revisions\/22147"}],"wp:attachment":[{"href":"http:\/\/www.sachdevajk.in\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=22146"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"http:\/\/www.sachdevajk.in\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=22146"},{"taxonomy":"post_tag","embeddable":true,"href":"http:\/\/www.sachdevajk.in\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=22146"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}