MCQOPTIONS
Saved Bookmarks
This section includes 607 Mcqs, each offering curated multiple-choice questions to sharpen your Computer Science Engineering (CSE) knowledge and support exam preparation. Choose a topic below to get started.
| 501. |
This clustering algorithm merges and splits nodes to help modify nonoptimal partitions. |
| A. | agglomerative clustering |
| B. | expectation maximization |
| C. | conceptual clustering |
| D. | k-means clustering |
| Answer» E. | |
| 502. |
This clustering algorithm initially assumes that each data instance represents a single cluster. |
| A. | agglomerative clustering |
| B. | conceptual clustering |
| C. | k-means clustering |
| D. | expectation maximization |
| Answer» D. expectation maximization | |
| 503. |
With Bayes classifier, missing data items are |
| A. | treated as equal compares. |
| B. | treated as unequal compares. |
| C. | replaced with a default value. |
| D. | ignored. |
| Answer» C. replaced with a default value. | |
| 504. |
This supervised learning technique can process both numeric and categorical input attributes. |
| A. | linear regression |
| B. | bayes classifier |
| C. | logistic regression |
| D. | backpropagation learning |
| Answer» B. bayes classifier | |
| 505. |
This technique associates a conditional probability value with each data instance. |
| A. | linear regression |
| B. | logistic regression |
| C. | simple regression |
| D. | multiple linear regression |
| Answer» C. simple regression | |
| 506. |
Logistic regression is a ________ regression technique that is used to model data having a _____outcome. |
| A. | linear, numeric |
| B. | linear, binary |
| C. | nonlinear, numeric |
| D. | nonlinear, binary |
| Answer» E. | |
| 507. |
The leaf nodes of a model tree are |
| A. | averages of numeric output attribute values. |
| B. | nonlinear regression equations. |
| C. | linear regression equations. |
| D. | sums of numeric output attribute values. |
| Answer» D. sums of numeric output attribute values. | |
| 508. |
Regression trees are often used to model _______ data. |
| A. | linear |
| B. | nonlinear |
| C. | categorical |
| D. | symmetrical |
| Answer» C. categorical | |
| 509. |
Simple regression assumes a __________ relationship between the input attribute and output attribute. |
| A. | linear |
| B. | quadratic |
| C. | reciprocal |
| D. | inverse |
| Answer» B. quadratic | |
| 510. |
The average squared difference between classifier predicted output and actual output. |
| A. | mean squared error |
| B. | root mean squared error |
| C. | mean absolute error |
| D. | mean relative error |
| Answer» B. root mean squared error | |
| 511. |
The correlation coefficient for two real-valued attributes is –0.85. What does this value tell you? |
| A. | the attributes are not linearly related. |
| B. | as the value of one attribute increases the value of the second attribute also increases. |
| C. | as the value of one attribute decreases the value of the second attribute increases. |
| D. | the attributes show a curvilinear relationship. |
| Answer» D. the attributes show a curvilinear relationship. | |
| 512. |
Bootstrapping allows us to |
| A. | choose the same training instance several times. |
| B. | choose the same test set instance several times. |
| C. | build models with alternative subsets of the training data several times. |
| D. | test a model with alternative subsets of the test data several times. |
| Answer» B. choose the same test set instance several times. | |
| 513. |
Data used to optimize the parameter settings of a supervised learner model. |
| A. | training |
| B. | test |
| C. | verification |
| D. | validation |
| Answer» E. | |
| 514. |
The standard error is defined as the square root of this computation. |
| A. | the sample variance divided by the total number of sample instances. |
| B. | the population variance divided by the total number of sample instances. |
| C. | the sample variance divided by the sample mean. |
| D. | the population variance divided by the sample mean. |
| Answer» B. the population variance divided by the total number of sample instances. | |
| 515. |
Selecting data so as to assure that each class is properly represented in both the training and test set. |
| A. | cross validation |
| B. | stratification |
| C. | verification |
| D. | bootstrapping |
| Answer» C. verification | |
| 516. |
The average positive difference between computed and desired outcome values. |
| A. | root mean squared error |
| B. | mean squared error |
| C. | mean absolute error |
| D. | mean positive error |
| Answer» E. | |
| 517. |
Which of the following is a common use of unsupervised clustering? |
| A. | detect outliers |
| B. | determine a best set of input attributes for supervised learning |
| C. | evaluate the likely performance of a supervised learner model |
| D. | determine if meaningful relationships can be found in a dataset |
| Answer» B. determine a best set of input attributes for supervised learning | |
| 518. |
Which statement is true about prediction problems? |
| A. | the output attribute must be categorical. |
| B. | the output attribute must be numeric. |
| C. | the resultant model is designed to determine future outcomes. |
| D. | the resultant model is designed to classify current behavior. |
| Answer» E. | |
| 519. |
Classification problems are distinguished from estimation problems in that |
| A. | classification problems require the output attribute to be numeric. |
| B. | classification problems require the output attribute to be categorical. |
| C. | classification problems do not allow an output attribute. |
| D. | classification problems are designed to predict future outcome. |
| Answer» D. classification problems are designed to predict future outcome. | |
| 520. |
Imagine a Newly-Born starts to learn walking. It will try to find a suitable policy to learn walking after repeated falling and getting up.specify what type of machine learning is best suited? |
| A. | classification |
| B. | regression |
| C. | kmeans algorithm |
| D. | reinforcement learning |
| Answer» E. | |
| 521. |
A nearest neighbor approach is best used |
| A. | with large-sized datasets. |
| B. | when irrelevant attributes have been removed from the data. |
| C. | when a generalized model of the data is desirable. |
| D. | when an explanation of what has been found is of primary importance. |
| Answer» C. when a generalized model of the data is desirable. | |
| 522. |
For a multiple regression model, SST = 200 and SSE = 50. The multiple coefficient of determination is |
| A. | 0.25 |
| B. | 4.00 |
| C. | 0.75 |
| D. | none of the above |
| Answer» C. 0.75 | |
| 523. |
The multiple coefficient of determination is computed by |
| A. | dividing ssr by sst |
| B. | dividing sst by ssr |
| C. | dividing sst by sse |
| D. | none of the above |
| Answer» D. none of the above | |
| 524. |
Another name for an output attribute. |
| A. | predictive variable |
| B. | independent variable |
| C. | estimated variable |
| D. | dependent variable |
| Answer» C. estimated variable | |
| 525. |
The adjusted multiple coefficient of determination accounts for |
| A. | the number of dependent variables in the model |
| B. | the number of independent variables in the model |
| C. | unusually large predictors |
| D. | none of the above |
| Answer» E. | |
| 526. |
A measure of goodness of fit for the estimated regression equation is the |
| A. | multiple coefficient of determination |
| B. | mean square due to error |
| C. | mean square due to regression |
| D. | none of the above |
| Answer» D. none of the above | |
| 527. |
A multiple regression model has |
| A. | only one independent variable |
| B. | more than one dependent variable |
| C. | more than one independent variable |
| D. | none of the above |
| Answer» C. more than one independent variable | |
| 528. |
A multiple regression model has the form: y = 2 + 3x1 + 4x2. As x1 increases by 1 unit (holding x2 constant), y will |
| A. | increase by 3 units |
| B. | decrease by 3 units |
| C. | increase by 4 units |
| D. | decrease by 4 units |
| Answer» D. decrease by 4 units | |
| 529. |
A term used to describe the case when the independent variables in a multiple regression model are correlated is |
| A. | regression |
| B. | correlation |
| C. | multicollinearity |
| D. | none of the above |
| Answer» D. none of the above | |
| 530. |
A regression model in which more than one independent variable is used to predict the dependent variable is called |
| A. | a simple linear regression model |
| B. | a multiple regression models |
| C. | an independent model |
| D. | none of the above |
| Answer» D. none of the above | |
| 531. |
Supervised learning differs from unsupervised clustering in that supervised learning requires |
| A. | at least one input attribute. |
| B. | input attributes to be categorical. |
| C. | at least one output attribute. |
| D. | output attributes to be categorical. |
| Answer» C. at least one output attribute. | |
| 532. |
Supervised learning and unsupervised clustering both require at least one |
| A. | hidden attribute. |
| B. | output attribute. |
| C. | input attribute. |
| D. | categorical attribute. |
| Answer» B. output attribute. | |
| 533. |
Computers are best at learning |
| A. | facts. |
| B. | concepts. |
| C. | procedures. |
| D. | principles. |
| Answer» B. concepts. | |
| 534. |
The process of forming general concept definitions from examples of concepts to be learned. |
| A. | deduction |
| B. | abduction |
| C. | induction |
| D. | conjunction |
| Answer» D. conjunction | |
| 535. |
The effectiveness of an SVM depends upon: |
| A. | selection of kernel |
| B. | kernel parameters |
| C. | soft margin parameter c |
| D. | all of the above |
| Answer» E. | |
| 536. |
What can be major issue in Leave-One-Out-Cross-Validation(LOOCV)? |
| A. | low variance |
| B. | high variance |
| C. | faster runtime compared to k-fold cross validation |
| D. | slower runtime compared to normal validation |
| Answer» C. faster runtime compared to k-fold cross validation | |
| 537. |
A student Grade is a variable F1 which takes a value from A,B,C and D. Which of the following is True in the following case? |
| A. | variable f1 is an example of nominal variable |
| B. | variable f1 is an example of ordinal variable |
| C. | it doesn\t belong to any of the mentioned categories |
| D. | it belongs to both ordinal and nominal category |
| Answer» C. it doesn\t belong to any of the mentioned categories | |
| 538. |
PCA works better if there is 1. A linear structure in the data 2. If the data lies on a curved surface and not on a flat surface 3. If variables are scaled in the same unit |
| A. | 1 and 2 |
| B. | 2 and 3 |
| C. | 1 and 3 |
| D. | 1,2 and 3 |
| Answer» D. 1,2 and 3 | |
| 539. |
You are given sesimic data and you want to predict next earthquake , this is an example of |
| A. | supervised learning |
| B. | reinforcement learning |
| C. | unsupervised learning |
| D. | dimensionality reduction |
| Answer» B. reinforcement learning | |
| 540. |
Prediction is |
| A. | the result of application of specific theory or rule in a specific case |
| B. | discipline in statistics used to find projections in multidimensional data |
| C. | value entered in database by expert |
| D. | independent of data |
| Answer» B. discipline in statistics used to find projections in multidimensional data | |
| 541. |
Which of the folllowing is an example of feature extraction? |
| A. | construction bag of words from an email |
| B. | applying pca to project high dimensional data |
| C. | removing stop words |
| D. | forward selection |
| Answer» C. removing stop words | |
| 542. |
Which of the following is a reasonable way to select the number of principal components "k"? |
| A. | choose k to be the smallest value so that at least 99% of the varinace is retained. - answer |
| B. | choose k to be 99% of m (k = 0.99*m, rounded to the nearest integer). |
| C. | choose k to be the largest value so that 99% of the variance is retained. |
| D. | use the elbow method |
| Answer» B. choose k to be 99% of m (k = 0.99*m, rounded to the nearest integer). | |
| 543. |
The most popularly used dimensionality reduction algorithm is Principal Component Analysis (PCA). Which of the following is/are true about PCA? 1. PCA is an unsupervised method 2. It searches for the directions that data have the largest variance 3. Maximum number of principal components |
| A. | 1 & 2 |
| B. | 2 & 3 |
| C. | 3 & 4 |
| D. | all of the above |
| Answer» E. | |
| 544. |
Feature can be used as a |
| A. | binary split |
| B. | predictor |
| C. | both a and b?? |
| D. | none of the above |
| Answer» D. none of the above | |
| 545. |
A measurable property or parameter of the data-set is |
| A. | training data |
| B. | feature |
| C. | test data |
| D. | validation data |
| Answer» C. test data | |
| 546. |
Following are the descriptive models |
| A. | clustering |
| B. | classification |
| C. | association rule |
| D. | both a and c |
| Answer» E. | |
| 547. |
If machine learning model output doesnot involves target variable then that model is called as |
| A. | descriptive model |
| B. | predictive model |
| C. | reinforcement learning |
| D. | all of the above |
| Answer» B. predictive model | |
| 548. |
In simple term, machine learning is |
| A. | training based on historical data |
| B. | prediction to answer a query |
| C. | both a and b?? |
| D. | automization of complex tasks |
| Answer» D. automization of complex tasks | |
| 549. |
The "curse of dimensionality" referes |
| A. | all the problems that arise when working with data in the higher dimensions, that did not exist in the lower dimensions. |
| B. | all the problems that arise when working with data in the lower dimensions, that did not exist in the higher dimensions. |
| C. | all the problems that arise when working with data in the lower dimensions, that did not exist in the lower dimensions. |
| D. | all the problems that arise when working with data in the higher dimensions, that did not exist in the higher dimensions. |
| Answer» B. all the problems that arise when working with data in the lower dimensions, that did not exist in the higher dimensions. | |
| 550. |
Select the correct answers for following statements. 1. Filter methods are much faster compared to wrapper methods. 2. Wrapper methods use statistical methods for evaluation of a subset of features while Filter methods use cross validation. |
| A. | both are true |
| B. | 1 is true and 2 is false |
| C. | both are false |
| D. | 1 is false and 2 is true |
| Answer» C. both are false | |