Research Snappy
  • Market Research Forum
  • Investment Research
  • Consumer Research
  • More
    • Advertising Research
    • Healthcare Research
    • Data Analysis
    • Top Companies
    • Latest News
No Result
View All Result
Research Snappy
No Result
View All Result

60 Interview Questions On Machine Learning

researchsnappy by researchsnappy
February 9, 2020
in Healthcare Research
0
60 Interview Questions On Machine Learning
400
SHARES
2.4k
VIEWS
Share on FacebookShare on Twitter


We frequently come out with resources
for aspirants and job seekers in data science to help them make a career in
this vibrant field. Cracking interviews especially where understating of machine
learning is needed can be tricky. Here are 60 most commonly asked interview
questions for data scientists, broken into linear regression, logistic
regression and clustering.

Part 1 – Linear Regression  

36 Question

  • What is linear regression?
  • A linear regression is a linear approximation of a causal relationship between two or more variables.
  • It falls under the supervised machine learning algorithms.
  • What is process of carrying out a linear regression?
  • Get sample data
  • Design a model that works on
    that sample
  • Make predictions for the whole
    population
  • How do you represent a simple linear regression?
  • Y = b0 +b1 x1 + e
  • Y – dependent variable
  • X1 – independent variable
  • e – Error term = Y – Y(hat)
  • What is the difference between correlation and regression?
  • Correlation does not apply
    causation. Regression is done to understand the impact of independent variable
    on the dependent variable.
  • Correlation is symmetric
    regrading both the variables p(x,y) = p(y,x). Regression is one way.
  • Correlation does not capture the
    direction of causal relationship. Regression captures the cause and
    effect. 
  • What are the columns in the coefficient table?
  • The coefficient table contains
    the variable name, coefficient, standard error and p-value.
  • What is standard error?
  • Standard error shows the
    accuracy for each variable
  • What is p-value?
  • The p-value shows the
    significance of the variable. It tells us if the variable is useful or
    not. 
  • The H0 is coefficient = 0 and
    the H1 is coefficient ≠ 0
  • If p-value < 0.05 (in most
    of the cases) we reject H0
  • What is OLS?
  • OLS stands for ordinary least square
  • It measures the error between
    the actual Y and predicted Y
  • Lower the error, better is the
    model
  • What are the other regression methods?
  • Generalized least squares
  • Maximum likelihood estimates
  • Bayesian regression
  • Kernel regression
  • Gaussian regression
  • What is TSS, ESS and RSS?
  • TSS stands for Total Sum of Squares. It measures the total variability.
  • TSS = ∑(y – y(mean))2
  • ESS stands for Explained Sum of Squares. It measures the variability that is explained.
  • ESS = ∑(y(pred) – y(mean))2
  • RSS stands for Residual Sum of Squares. It measures the difference between the observed Y and predicted Y.
  • RSS = ∑(y – y(pred))2
  • What is the relationship between TSS, ESS and RSS?
  • TSS = ESS + RSS
  • Total variability = Explained variability + Unexplained variability
  • What is R-Squared?
  • R-Squared is also known as goodness of fit
  • Smaller the RSS, better is the model
  • R-Sq = ESS / TSS = 1 – (RSS / TSS)
  • R-Squared takes a value between 0 and 1.
  • If R-Sq = 0 then the model does not explain any variability
  • If R-Sq = 1 then the model explains entire variability
  • What is adjusted R-Squared?
  • Adjusted R-Squared is a step on R-Squared and adjusts for the number of variables included in the model
  • As we add more variables the explanatory power of the model may increase.
  • Adjusted R-Squared penalizes the model for the number of variables that are used in the model.
  • What is the relationship between R-Squared and Adjusted R-Squared?
  • Adj R-Sq is always lower than the R-Sq
  • Adj R-Sq = 1 – ((1-RSq) * (n-1) / (n-p-1))
  • Where n is the number of observations and p is the number of variables
  • What happens when we add a variable and it increases the R-Sq but decreases the Adj R-Sq?
  • The variable can be omitted since it holds no predictive power
  • We should also look at the p-value of the added variable and confirm our decision
  • What is feature selection?
  • It is a method to simplify the model and improves the speed
  • It is done to avoid too many features
  • p-value in regression coefficient table can be used to drop insignificant variables
  • What is feature scaling?
  • Different variables have different magnitude
  • Feature scaling is done to bring the variables to the same magnitude
  • Standardization is one of the methods used for feature scaling
  • What is standardization?
  • It is also called normalization
  • X (std) = (x – µ) / σ
  • Regardless of the data we will get data with mean 0 and standard deviation of 1
  • What is the interpretation of the weights?
  • In ML coefficients are called weights.
  • A positive weight shows that as feature increases in value, so does Y
  • A negative weight shows that as feature decreases in value, so does Y
  • What is the difference between overfitting and underfitting?
  • Underfitting happens when the
    model has not captured the underlying logic of the data.
  • Overfitting happens when the
    model has focused too much on the training dataset that it cannot understand
    test dataset
  • How to identify if the model is overfitting or underfitting?
  • Underfit model performs bad
    (low accuracy) on training and bad (low accuracy) on test.
  • Overfit model performs good
    (high accuracy) on training and bad (low accuracy) on test.
  • A good model performs good
    (high accuracy) on training and good (high accuracy) on test.
  • What is multiple linear regression?
  • In multiple linear regression
    that are more than one predictor.
  • Good models require multiple
    independent variables in order to address the higher complexity of the problem.
  • Y = b0 +b1 x1 + b2 x2 + … + bk
    xk + e
  • What are the assumptions of linear regression?
  • Linearity
  • No endogeneity
  • Normality and homoscedasticity
  • No autocorrelation
  • No multi-collinearity
  • What happens if the linear regression violates any of its
    assumptions?
  • The biggest mistake you can
    make is to perform a regression that violates one of its assumptions.
  • If the regression assumptions
    are violated, then performing regression analysis will yield incorrect results.
  • What does linearity mean?
  • It means a linear relationship
  • To check if there is linear
    relationship between x and y the simplest thing to do is plot a scatter plot
    between x and y
  • What are the fixes of linearity?
  • If linearity assumption is
    violated, then we can use non-linear regression
  • We can also transform the x
    (exponential transformation / log transformation)
  • What does no endogeneity mean?
  • No endogeneity means no
    relationship between x and ε
  • It may be because we have
    omitted an important predictor from the model
  • What is omitted variable bias?
  • If the modeler forgets to
    include an important predictor in the model
  • It may lead to counter-intuitive
    coefficient signs
  • Once the important variable is
    included rest of the coefficients fall into place
  • What is the assumption of normality?
  • It means the normal
    distribution of the error term
  • The mean of the residuals
    should be zero
  • The standard deviation of the
    residuals should be constant
  • What is the assumption of homoscedasticity?
  • In simple terms it means the
    equal variance
  • There is no relationship
    between the error term and the predicted Y
  • How to prevent heteroscedasticity?
  • It may be due to outliers
  • It may be due to omitted
    variable bias
  • Log transformation
  • What does autocorrelation mean?
  • It is common in time series
    modeling
  • It means that Y(t) is dependent
    on historical values, Y(t-1) or Y(t-2) or … Y(t-k)
  • How to detect autocorrelation?
  • DW test is used to detect
    autocorrelation
  • If DW test statistics is less
    than 1 then there is strong autocorrelation
  • If DW test statistics is close
    to 2 then there is no autocorrelation
  • If DW test statistics is more
    then 3 then there is strong autocorrelation
  • What are the remedies to remove autocorrelation?
  • There is no remedy in linear
    regression
  • The modelers can try different
    models like AR, MA, ARMA or ARIMA
  • What does multicollinearity mean?
  • When two or more variables have
    high correlation
  • If there is perfect
    multicollinearity then standard error will be infinite
  • Imperfect multicollinearity
    means that the correlation is slightly less than 1 or slightly more than -1.
    However imperfect multicollinearity also causes serious issues in the model
  • What are the fixes of multicollinearity?
  • Find the correlation between
    each pair of independent variables
  • If two variables are highly
    correlated, then either drop one of them or transform them into a single
    variable

Part 2 – Logistic Regression 

10 Question

  • What is logistic
    regression?
  • Logistic regression predicts the probability of an event occurring
  • LN (π / (1-π)) = b0 +b1 x1 + b2 x2 + … + bk xk + e
  • LN (π / (1-π)) is called
    log-odds
  • Why cannot we use linear
    regression for dichotomous output?
  • The linear regression is used
    for unbounded output
  • The linear regression does not
    know that the output is bounded between 0 and 1
  • What are the assumptions of logistic regression?
  • Linearity
  • No endogeneity
  • Normality and homoscedasticity
  • No autocorrelation
  • No multi-collinearity
  • What is MLE?
  • MLE stands for maximum
    likelihood estimate
  • It is a function which
    estimates how likely it is that the model at hand describes the real underlying
    relationship of the variables.
  • Bigger the MLE, the higher the
    probability that our model is correct.
  • How to interpret the coefficients of the logistic regression?
  • The predicted variable is “is_promoted”
  • The coeffects can be
    interpreted as

LN(is_promoted
/ (1- is_promoted)) = 9.9311 + 0.9106 x region_lnp + 0.8291 x KPIs_met_lnp +
0.8665 x awards_won_lnp + 0.659 x previous_year_rating + 1.1796 x avg_training_score_lnp

  • What is the sensitivity of the variable “avg_training_score_lnp”
    based on the coefficient table?
  • LN(is_promoted 2 / is_promoted
    1) = 1.1796 x (avg_training_score_lnp 2 – avg_training_score_lnp 1)
  • If the avg_training_score_lnp
    is increased by 1 unit then is_promoted 2 = 3.2531 * is_promoted 1
  • That means the is_promoted
    increases by 3.25 times.
  • c is equivalent to the
    well-known measure ROC.
  • c ranges from 0.5 to 1, where
    0.5 corresponds to the model randomly predicting the response, and a 1
    corresponds to the model perfectly discriminating the response.
  • What is concordance / discordance / tie pairs?
  • Percent Concordant – A pair of observations with different observed responses is said to
    be concordant if the observation with the lower ordered response value (0) has
    a lower predicted mean score than the observation with the higher ordered
    response value (1).
  • Percent Discordant – If the observation with the lower ordered response value has a
    higher predicted mean score than the observation with the higher ordered
    response value, then the pair is discordant.
  • Percent Tied – If a pair of observations with different responses is neither
    concordant nor discordant, it is a tie.
  • How to measure the accuracy of logistic regression?
  • Where the prediction is <
    0.5 there the predicted variable = 0. Where the prediction is >= 0.5 there
    the predicted variable = 1.
  • Confusion matrix is used to
    measure the accuracy of the logistic regression.
  • What is the accuracy and misclassification for the below confusion
    matrix?
  • Accuracy is defined as correct
    classification by the algorithm
  • Based on the above confusion
    matrix, accuracy = (49845 + 4008) / 54808
  • Misclassification is 1 –
    accuracy, misclassification = 1- 98.3% 

Part 3 – Clustering

14 Question

  • What is cluster analysis?
  • Cluster analysis is a multivariate statistical technique that groups
    observations based on their features
  • Observations in a dataset can be divided into different groups and
    sometimes this is very useful
  • What is the goal of
    clustering analysis?
  • The goal of clustering analysis is to maximize the similarity of
    observations within a cluster and maximize the dissimilarity between clusters
  • What is the difference
    between supervised learning and unsupervised learning?
  • In supervised learning we are dealing with labelled data. We know the
    correct values before training our model. Examples are regression and
    classification.
  • In unsupervised learning we do not know the correct values before
    training our model. Examples are clustering.
  • What is Euclidean
    distance?
  • Euclidean distance measure the shortest distance between two points
    (x1,y1) and (x2,y2)
  • ED = SQRT((x2-x1)^2 + (y2-y1)^2)
  • What is centroid?
  • Centroid is the midpoint between two points (x1,y1) and (x2,y2)
  • C = ((x2+x1)/2 , (y2+y1)/2)
  • What is the different
    between exploratory and confirmatory & explanatory analysis?
  • Exploratory analysis:
  • Confirmatory & explanatory analysis:
    • Validate previous research
  • What is K-Means algorithm?
  • We start with selecting the number of clusters we want
  • Next, we specify the cluster seeds – the starting centroids 
  • Then, we assign each point to a centroid
  • We adjust the centroids
  • The above two steps are repeated will we can no longer assign the points
  • What are the pros and cons
    of the K-Means algorithm?
  • Pros:
  • Cons:
    • We need to pick number of clusters
    • Sensitive to initialization
  • How do we select the
    number of clusters?
  • Elbow method is used to determine the number of clusters
  • The objective is to minimize the within cluster sum of squares
  • Based on the below graph
    suggest the number of clusters?
  • At number of clusters = 5 we reach an elbow  
  • Based on the number of clusters = 5 the centroid of the clusters is
    summarized below.
  • How can we make sure that
    K-Means output is not sensitive to initialization?
  • Run multiple K-means with different initialization and compare results
  • Use KMeans++ algorithm to get the most optimal seed
  • What is hierarchical
    clustering?
  • Hierarchical clustering is a method of cluster analysis which seeks to
    build a hierarchy of clusters.
  • Hierarchical clustering came before K-Means clustering
  • One of the examples is animal kingdom.
  • What are the different
    methods of hierarchical clustering?
  • Agglomerative – Also called bottom-up approach. Each observation starts in its own
    cluster and pairs of clusters are merged as one moves up the hierarchy.
  • Divisive – Also called top-down approach. All observations start in one cluster,
    and splits are performed recursively as one moves down the hierarchy.
  • What are the pros and cons
    of the hierarchical clustering?
  • Pros:
    • Shows all possible linkages
    • Give much better understanding of data
    • No need to pre-define the number of clusters
  • Cons:
    • Computationally expensive

Enjoyed this story? Join our Telegram group. And be part of an engaging community.



Previous Post

Wintrust Financial (NASDAQ:WTFC) Price Target Cut to $75.00

Next Post

Consumer NAS Market – Global Size, Share, Trends, Analysis and Forecast 2020 – 2025 |Apple, Dell Technologies, NEC, Seagate Technology, etc – Instant Tech Market News

Next Post
Consumer NAS Market – Global Size, Share, Trends, Analysis and Forecast 2020 – 2025 |Apple, Dell Technologies, NEC, Seagate Technology, etc – Instant Tech Market News

Consumer NAS Market – Global Size, Share, Trends, Analysis and Forecast 2020 – 2025 |Apple, Dell Technologies, NEC, Seagate Technology, etc – Instant Tech Market News

Research Snappy

Category

  • Advertising Research
  • Consumer Research
  • Data Analysis
  • Healthcare Research
  • Investment Research
  • News
  • Top Company News

HPIN International Financial Platform Becomes a New Benchmark for India’s Digital Economy

Top 10 Market Research Companies in the world

3 Best Market Research Certifications in High Demand

  • Privacy Policy
  • Terms of Use
  • Antispam
  • DMCA
  • Contact Us

© 2025 researchsnappy.com

No Result
View All Result
  • Market Research Forum
  • Investment Research
  • Consumer Research
  • More
    • Advertising Research
    • Healthcare Research
    • Data Analysis
    • Top Companies
    • Latest News

© 2025 researchsnappy.com