Practice Databricks Databricks-Certified-Professional-Data-Scientist Exam Questions
Page: 1/28 Total 138 Questions
Question No 1
Feature Hashing approach is "SGD - based classifiers avoid the need to predetermine vector size by
simply picking a reasonable size and shoehorning the training data into vectors of that size" now with
large vectors or with multiple locations per feature in Feature hashing?
Question No 2
What are the advantages of the Hashing Features?
Question No 3
Question - 3: In machine learning, feature hashing, also known as the hashing trick (by analogy to the
kernel trick), is a fast and space - efficient way of vectorizing features (such as the words in a
language), i.e., turning arbitrary features into indices in a vector or matrix. It works by applying a
hash function to the features and using their hash values modulo the number of features as indices
directly, rather than looking the indices up in an associative array. So what is the primary reason of
the hashing trick for building classifiers?
Question No 4
Suppose A, B , and C are events. The probability of A given B , relative to P(|C), is the same as the
probability of A given B and C (relative to P ). That is,
Question No 5
What is the considerable difference between L1 and L2 regularization?
Page: 1/28 Total 138 Questions
Full Access
Databricks-Certified-Professional-Data-Scientist Exam Features