Computer Science Homework Help

University of the Cumberlands Week 11 Data Mining Discussion

 

I need support with this Computer Science question so I can learn better.

Accuracy is one of the characteristics of data. Information is supposed to be accurate and correct. If one wants to determine if the information, they are using is correct or not, one needs to ask themselves if that particular data reflects on the real-world position. Inaccurate data in any business or organization will cause problems that will hinder the organization from meeting its goals and objectives. Data should also be relevant. There must be a valid reason why the data is being collected. One should contemplate if they need that information. Data should also be up to date. This is a crucial characteristic because untimely details will lead to making wrong decisions (Hamel et al, 2017).

In the prototype-based cluster, observations carried out are assigned to the centroids and medoids. This is unlike density-based clustering, whereby the unsupervised learning methods recognize different groups in the information or data. On the other hand, graph-based clustering consists of unsupervised algorithms mainly designed to group the graph edges and vertices.

The scalable clustering algorithm is defined as identifying the same descriptions in different groups of data on a profile basis. Salable clustering algorithm involves distance metrics whereby the data points resemble various partition points (Vishwasrao & Sangaiah, 2017).

For one to choose the correct algorithm, one needs to consider the number of features. In addition to this, the accuracy of the output should be deemed when choosing the correct algorithm. One should also collect enough amount of information if they want to get reliable predictions. Linearity is also essential to consider when choosing a good algorithm. The number of parameters helps in selecting the right algorithm. Algorithms that have large numbers of parameters need many trials to get the right combination.

Computer Science Homework Help

Literature Review

 

Topic-IMPACTS AND ISSUES OF PHYSICAL SECURITY

Literature Review

  • Introduction
  • Subsections (based on a deductive approach)
  • Summary

APA format is must. Required pages 60-65 pages. References – 65.

Computer Science Homework Help

CS 330 Final Project

 

I need a 3D Scene created from the attached picture. The code must be in C++ language and run in OpenGL in Visual Studio 2019. The solution must be a .zip file containing all the necessary directories and libraries. I will upload the picture I need to recreate the 3D scene from. Please don’t use glut. Here are the requirements from the course:

Directions

Using the image you selected in a previous milestone, you will be creating 3D objects that represent the components and layout of that image. Although you have already begun to complete some of this work in your other milestones, during this project you will be refining and adding to your earlier submissions before bringing everything together. Note that you will be working on your 3D scene in Visual Studio but will also submit a written design decisions document discussing your approach throughout the process.

3D OBJECTS

  1. Create low-polygon 3D representations of real-world objects. Make sure you have at least four completed objects in your 3D scene. At least one of the objects you create should be made using two or more primitive shapes. Note that the object you completed in a previous milestone can count as one of your four. Utilize organized geometry and ensure that polygons (triangles) on each 3D model are well spaced and connected. To minimize complexity and save 3D modeling time, the polygon count for your objects should not exceed 1,000 triangles. As you work, remember to think in terms of simple shapes and ask yourself what primitive 3D shapes go into making up each object in your scene. Four of the following primitive shapes must appear at least once in your creation:
    • Cube
    • Cylinder
    • Plane
    • Pyramid
    • Sphere
    • Torus
  1. Apply accurately projected textures to a 3D model. You must select two objects to texture. Note that you should have already textured one object in a previous milestone. If you use that object here, it will count as one of your two. As you work, the textures you select should be royalty-free images with resolutions of 1024 x 1024 pixels or higher. Please refer to the Sourcing Textures Tutorial, linked in the Supporting Materials section, for guidance on how to locate images that can be used for textures.
  1. Apply lighting to create a polished visualization of 3D models. You must include a minimum of two light sources, and at least one of them should be colored. Note that the light you worked on in a previous milestone counts as one of your two lights. The light sources you create will need to capture all of the objects in the 3D world you are building, meaning they should be positioned at locations that do not cause parts of the objects to appear dark when moving the camera around them. While we recommend that you include a point light for one of your two lights, you may implement a directional light or spotlight if you choose. As you generate lighting, make sure that any lights are designed in a way that helps curate a final polished presentation. You will need to properly implement all components of the Phong shading model, including the following:
    • Ambient
    • Diffuse
    • Specular
  1. Place objects appropriately, using the X, Y, and Z coordinates, relative to one another in the 3D world. As you work, be sure to match the photograph you selected as closely as possible by placing the objects in their proper locations. Note that when you first import code for the objects you created in previous weeks, the objects may overlap, as it is likely that they were all initially placed at 0, 0, 0.

NAVIGATION

  1. Apply horizontal, vertical, and depth camera navigation around the 3D scene. The camera will be traversing the X, Y, and Z axes, and you should ensure it can capture all of the objects in your 3D scene. In a previous milestone, you already created some of this code. It is recommended that you use the code you have already created and then increase the radius of the camera’s orbit so it will correctly encompass all of the objects in the world you are building. You may find it easiest to add each object separately and then adjust the orbit radius or position of the camera each time. As you work, we recommend you use the following input devices:
    • WASD keys: These keys should be used to control the forward, backward, left, and right motion.
    • QE keys: These keys should be used to control the upward and downward movement.
  1. Apply nuanced camera controls to effectively view the 3D objects in the application. This should allow the orientation of the camera to change even though its location has not moved. You should focus first on pitch and yaw, but careful changes can be made to roll, keeping in mind that you may want the upward direction to stay in the same location. As you work, you will also want to code for adjustments in the speed of the movement so a user will have more control over how they explore the objects in the scene. We recommend you use the following input devices:
    • Mouse cursor: This should be used to change the orientation of the camera so it can look up and down or right and left.
    • Mouse scroll: This should be used to adjust the speed of the movement, or the speed the camera travels around the scene.
  1. Create perspective and orthographic displays of the 3D world. Use the tap of a keyboard key to allow a user to change the viewport display of all objects in the scene between orthographic (2D) and perspective (3D) views at will. To accomplish this, you will be switching the function call to retrieve either the perspective or orthographic projection matrix. Note that you will be keeping the camera in the same orientation that you already developed in previous criteria.

Computer Science Homework Help

Create a Template to automate a process

 

By automating the process, it will allow the IT team to deploy these systems on demand for the various departments. This will save time and remove chances of human error for every new deployment.

Create a Template to Automatically Deploy an Email-Receiving Pipeline

Use the instructions in the Getting Started Receiving Email with Amazon SES tutorial as a guide to help understand the process to deploy this service. You can then either use CloudFormation Designer to generate the appropriate template using the associated widget, or you can find the appropriate code on your own.

The template should perform these tasks automatically:

☑️ Set Up Prerequisites
☑️ Verify Your Domain
☑️ Set up a Receipt Rule to Create an S3 Bucket
☑️ Send a Test Message

Test your template by viewing the test message sent. After you have confirmed that it is working, clean up your environment to avoid new charges from AWS. Save your template for the final Cloud Orchestration and Automation Report before moving to the next step. There, you will deploy storage and content delivery systems.

Computer Science Homework Help

Cumberlands Cyber Security Awareness Among University Students Article Review

 

You will write a 2 page review/abstract/summary on an article from a peer- reviewed scholarly journal. This is to assess your ability to select and summarize the research of others, analyze and apply the research of others, and communicate professionally and effectively to their regard. However, the most important rationale for this assignment is for you to see how statistical analysis is presented.Instructions:

  1. Use the library system (https://www.ucumberlands.edu/library) or online catalog to locate a journal article that pertains to your research, thesis, or area of interest. The article you chose should have performed some statistical analysis of gathered data and made an inference using something other than just the average (mean). That is, they can’t just talk about averages; they must have used one of these tests: t-test, chi-square, F-test, Fischer test, ANOVA, MANOVA, ANCOVA, Mann-Whitney, correlation, regression.
  2. Readthearticlethoroughly.
  3. Write a 2-page summary about the article following the given guidelines.

Guidelines:

  1. Your review should include:
    1. Thequestion/problembeingresearchedbytheauthor
    2. The experiment that will answer the question
    3. How they collected data
    4. Analysis of the data (Must identify the statistical test used)
    5. Theirconclusionorfindings

Computer Science Homework Help

PJM 6630 Boston University Wk 3 Touch Screen Kiosks Survey Essay

 

Overview and Rationale

undefined

In this assignment you will develop a survey intended for distribution to family and friends, although you will not actually conduct the survey.

undefined

Week 3 Learning Objectives

undefined

This assignment is directly linked to the following learning outcomes from the course syllabus:

undefined

  • Utilize tools and techniques to elicit requirements to support realization of business goals.
  • Interpret the needs of the business by translating their priorities and producing the requirements for the project team.

Computer Science Homework Help

SJSU Cloud Consumers Discussion

 

Answer 1

The approach uses a machine-learning-based framework to classify cloud consumers based on their behavioral characteristics, including time, location, and computing power. The cloud consumers can be categorized based on their data access requirements by considering the following types of cloud consumers: consumers who access cloud services in bulk, those who access cloud services using a subscription model, those who access cloud services in the cloud, and consumers who access cloud services through a per-service model. The service consumers can be classified based on the types of services they access in the cloud. The service consumers for the cloud consumer that access cloud services through a subscription model. Summary, A variety of cloud applications and services can be accessed and managed in the cloud system through the cloud consumer. The primary advantage of using an access scheme for cloud service usage is that the cloud consumer does not incur additional provisioning and management costs for data center infrastructure. It is essential to understand the differences between cloud storage and network. With the network, the cloud consumer uses the underlying network infrastructure and the underlying data center infrastructure (Thirugnanam et al., 2020). Cloud storage is designed to be the most flexible and cost-effective solution because it can leverage the existing infrastructure in many different situations. Network architecture is a significant concern for cloud consumers because it is essential to realize that cloud services and virtualization are not the same. Cloud providers have the responsibility to build a virtual network that provides high-speed service to their customers. Network topologies can only be used if the network structure for the customers is optimized. Therefore, the provider needs to understand the customers’ requirements to ensure that the network architecture for the customers can meet those requirements. In order to optimize the network architecture, network management and the service delivery team work together to optimize the best possible network architecture for the cloud providers.

This method also facilitates the use of nonparametric measures in future decision-making processes and enhances our ability to detect unexpected statistical events. We propose a machine learning-based approach to analyze the responses to weather models, and specifically, we utilize the model performance measures for prediction accuracy. This approach allows for the use of improved nonparametric statistics and is scalable for large datasets. The results presented in this show that our approach can be applied to many existing data-mining frameworks. The previous presents an example of using ANN to mine eCommerce data using sentiment analysis to understand purchasing behavior. In practice, ANNs are applied to real-world data in order to support real-world business and social problems (Melati et al., 2019).

Reference:

Melati, D., Grinberg, Y., Kamandar Dezfouli, M., Janz, S., Cheben, P., Schmid, J. H., Sánchez-Postigo, A., & Xu, D.-X. (2019). Mapping the global design space of nanophotonic components using machine learning pattern recognition. Nature Communications, 10(1), 4775–4775. https://doi.org/10.1038/s41467-019-12698-1

Thirugnanam, H., Ramesh, M. V., & Rangan, V. P. (2020). Enhancing the reliability of landslide early warning systems by machine learning. Landslides, 17(9), 2231–2246. https://doi.org/10.1007/s10346-020-01453-z

—————————————————————————————————————————————-

Answer 2

The huge amount of data requires powerful methods, algorithms, and techniques that are hard to find, describe, and implement affordably. In the field of big data analytics, one such approach is called big-data processing. Big data processing attempts to make large amounts of existing data available to efficient analytics tools by analyzing it in a scalable way that is suitable..handling large data sets. The big data paradigm encompasses a variety of different technologies that we explore later in this. We begin with the challenges and opportunities people face when adopting the distributed, scalable database model, and we continue with key technologies that are driving these changes. The emergence of distributed, scalable database applications has had profound effects on the world of information. We can now create new and richer services that span the globe without deploying servers and data centers across the globe (Terol et al., 2020).

Moreover, this capability has resulted in some interesting developments in this field. For example, some techniques to achieve high mobility have been proposed, including the Transport of magnetic materials over non-volatile media. For example, in In, the authors proposed a novel mechanism to support magnetic and thermal media transmission over non-conductive surfaces such as those of a radio system over a non-conductive carrier medium. This was achieved by employing a hybrid electronic waveguide. As already stated, such a transmission network would not only be more reliable but, more importantly, more efficient. This transmission network would not only save energy by reducing the amount of power required to transmit a packet, but it would also lead to fewer errors being made in transmission. It is important to point out that this mechanism is, at the same time, more complex and more expensive than sending a message by direct radio or electricity exchange (Xu & Jackson, 2019).

The big data transformation will be performed as a parallel process, and once done, all the data will be stored in a “cloud” of data warehouses, on which each of the data warehouse operations is built. The data warehouse operations are executed in a parallel manner. This is the new way for companies. The big data revolution has provided us with the means to change the way people live. People live without the necessary tools or knowledge to act. When tools and knowledge are insufficient, people behave and exploit others. These are the types of relationships that we want to eliminate. The goal of this book is to help you move beyond the self-serving, ego-driven, selfish, short-sighted, and ineffective approach to security management (Xu & Jackson, 2019).

Reference:

Terol, R. M., Reina, A. R., Ziaei, S., & Gil, D. (2020). A Machine Learning Approach to Reduce Dimensional Space in Large Datasets. IEEE Access, 8, 148181–148192. https://doi.org/10.1109/ACCESS.2020.3012836

Xu, C., & Jackson, S. A. (2019). Machine learning and complex biological data. Genome Biology, 20(1), 76–76. https://doi.org/10.1186/s13059-019-1689-0

Computer Science Homework Help

San Jose State University Handling Categorical Attributes Discussion Responses

 

Answer 1

Two (2) techniques in handling categorical attributes are as follows:

One-Hot Encoding: One-Hot Encoding is the most common and correct way to deal with non-ordinal categorical data as it consists of creating an additional feature for each group of the categorical feature and mark each observation belonging (Value=1) or not (Value=0) to that group. (Zuccarelli, 2020)

Target Encoding: It consists of substituting each group in a categorical feature with the average response in the target variable. The process to obtain the Target Encoding are: Group the data by category, Calculate the average of the target variable per each group and Assign the average to each observation belonging to that group (Zuccarelli, 2020).

Two (2) ways in which continuous attributes differ from categorical attributes are as follows:

Continuous/Quantitative attribute is data where the values can change continuously, and we cannot count the number of different values whereas Categorical attribute, in contrast, is for those aspects of our data where we make a distinction between different groups, and where we typically can list a small number of categories (Kosara, 2013).

Examples of Continuous attribute include weight, price, profits, counts, etc. Basically, anything that can be measure or count is quantitative whereas Categorical attributes includes product type, gender, age group, etc.

References

Kosara, R. (2013). Data: Continuous vs. Categorical.

Zuccarelli, E. (2020). Handling Categorical Data, The Right Way.

—————————————————————————————————————————————-

Answer 2

  • Discuss four (2) techniques in handling categorical attributes?

Categorical attributes including symmetric binary attributes such as Gender, and nominal attributes such as state and level of education. In order to apply association rule mining and extract pattern from the categorical attributes, there is a need to transform them into items. One way of transformation is to create a new binary item for each attribute-value pair, for example, gender can be replaced with Male and female, education can be replaced with education = graduate, education = college etc. One of the items will have a value of 1 the rest will have a value of 0. Another approach is to group related attribute values to smaller number of categories or group less frequent items into “others” category. This approach works well with nominal attributes whose values are not frequent, such as State. The third technique is to remove some high-frequency items before apply standard association rule algorithms, because they corresponds to typical values of an attribute and seldom carry new information about the pattern. The fourth technique is to avoid generating candidate itemsets that contains more than one item from the same attribute because the support count of the itemset will be zero. This approach helps reduce the computation time (Tan et al., 2019).

  • Discuss two (2) ways in which continuous attributes differ from categorical attributes?

Examples of continuous attributes are annual income and age. They need to be handled differently from how we handle the categorical attributes. One method is discretization method, where the adjacent values of a continues attribute are grouped into finite intervals and then the discrete intervals can be mapped to asymmetric binary attributes and the existing association analysis algorithms can be applied. Another way of handling continuous attribute is to transform the data into 0/1 matrix. If the counts exceed a certain threshold the entry will be 1 and otherwise 0. By transforming the continuous attributes into binary dataset, the existing frequent itemset generation algorithms can be applied. However, the association accuracy will be impacted by the threshold value. Some associations will be missed if the threshold is too high and low threshold can result in many spurious associations. The third approach is statistics-based method, where the interested target variable is withheld, the rest categorical and continuous attributes are binarized. Finally, the existing algorithms such as Apriori and FP-growth can be applied to the binarized data to extract frequent itemsets (Tan et al., 2019).

Reference:

Tan, P.-N., Steinbach, M., Kumar, V., & Karpatne, A. (2019). Introduction to data mining (Second). Pearson.

Computer Science Homework Help

San Jose State University K Means Discussion Responses

 

Answer 1

1.What is K-means from a basic standpoint?

K-means is an algorithm used in cluster analysis, which is a type of unsupervised machine learning (Tan et al., 2019). Since there are no labels applied to the data, clustering is described as unsupervised. When class labels must be applied to the data, that signifies a supervised process, and thus it becomes a classification method rather than a type of clustering (Tan et al., 2019). The difference between unsupervised clustering and supervised classification can be shown with an example. Suppose a fruit farm is growing one variety each of grapes, peaches and nectarines and using semi-automated, semi-intelligent picking equipment. A grapevine will grow its fruit in clusters on a stem (a bunch of grapes) in a manner that is unique among the three fruits. The machine has no trouble distinguishing grapes from either of the other two fruits. The grapes can be harvested in clusters with little chance of the grapes getting confused with either the peaches or nectarines, so it can use unsupervised clustering techniques when harvested by the machinery. However, the peach and nectarine trees must have their fruit picked and then manually classified by a human (assuming the trees are mixed in together) since the characteristics of the fruit is so similar. The machine that picks the peaches and nectarines cannot tell the difference between the two (and sometimes neither can I). So some sort of classification is necessary for the peaches and nectarines, which is considered supervised learning. K-means uses a prototype or representative data point for each of the clusters or groups (Tan et al., 2019). The representative data point is selected by an average, which is basically the means or centroid of the data points (Tan et al., 2019). There is one centroid for each group or cluster, and it is represented by K from the name, K-means (Garbade, 2018). K-means does calculations iteratively to select the best placement of the centroids within the clusters (Garbade, 2018). When there are no further changes in placement or the number of desired iterations is complete, the K-means algorithm is complete (Garbade, 2018). What are some examples of how K-means could be used in real life? According to Raghupathi (2018), the K-means algorithm can be used for document classification, identifying high-risk crime locations, customer segmentation, insurance fraud detection, profiling cyber criminals, and more. Can anybody think of other common examples of K-means used in clustering?

2. Is a binary variable the same as a dichotomous variable? Provide scholarly justification.

In general terms, yes, binary and dichotomous variables are the same (Glen, 2014). However, when digging into their definitions in more details, subtle differences can be seen. Glen (2014) described a binary variable as a subtype under the larger category of dichotomous variable. A dichotomous variable has two possible values, such as pass or fail, which represent nominal categories (Glen, 2014). A binary variable also has two values, but is typically represented either in Boolean (such as False or True), or integer (such as 0 or 1) values (Karabiber, 2021). In addition, dichotomous variables can either be discrete or continuous variables, (Glen, 2014). For example, the pass or fail dichotomous variable could have a value of 69.5, but if the professor is nice and rounds up to 70, it is passing (Glen, 2014). Binary variables on the other hand are a discrete variable with no option for a range (Tan et al., 2019). So what are some other examples for thought? Glen (2014) explained that a person can be dead or alive, so that is a discrete dichotomous variable. What happens if a person is on a life support system, however? Would that person then be considered in a continuous dichotomous state of living? And we know the binary descriptor of male or female describes the majority of the population. However, what about those individuals who describe themselves as non-binary? If they are non-binary are they then considered continuous dichotomous or would they be discrete dichotomous?

References

Garbade, M. (2018). Understanding K-means clustering in machine learning. Retrieved from https://towardsdatascience.com/understanding-k-mea…

Glen, S. (2014). Dichotomous variable: Definition. StatisticsHowTo.com: Elementary Statistics for the rest of us! Retrieved from https://www.statisticshowto.com/dichotomous-variab…

Karabiber, F. (2021). Binary variable. Retrieved from https://www.learndatasci.com/glossary/binary-varia…

Raghupathi, K. (2018). Ten interesting use cases for the K-Means algorithm. Retrieved from https://dzone.com/articles/10-interesting-use-case…

Tan, P., Steinbach, M., Karpatne, A., & Kumar, V. (2019). Introduction to Data Mining (2nd Edition). Pearson Education (US).

—————————————————————————————————————————————-

Answer 2

K-means is a learning algorithm that is unsupervised and designed to classify any unlabeled data into a given number of distinct clusters. It is simply a way of assessing any observations with similar characteristics and putting them together. The recommended cluster consists of observations that are more similar in a cluster than the clusters themselves. The algorithm’s aim is to minimize an objective function (Caruso et al., 2021).

2

The dichotomous variable is a natural choice for this analysis as it gives us more information on how people interact. If people want to communicate, there are two things that they can do. The first is to use a personal website, and the second is a public website to communicate with others. There are two types of public communication sites: a site for the whole group or small groups, or an entire organization. Both of these sites are important and are described in a follow-up (Caruso et al., 2021).

The binary variable is the same as a dichotomous variable because there is a chance of choosing the positive answer with the smallest sample size. The sample sizes available for binary variables include the following, A sample of n records A sample of n. n is a random sample of n. If there are no constraints on sample sizes, then samples of n will always be chosen by a priori probability. The sample sizes are not shown (Caruso et al., 2021).

Reference

Caruso, G., Gattone, S. A., Fortuna, F., & Di Battista, T. (2021). Cluster Analysis for mixed data: An application to credit risk evaluation. Socio-Economic Planning Sciences, 73, 100850.