Automatic Liquid handling notes:
Automatic Liquid Handling (ALH) requires the control of two variables: accuracy and precision.
Inaccuracy is defined as the deviation of a measurement away from the tr
Name Sequence TmC CG% nt A T C G Extinction coefficient(l/(molcm) Molecular weight(g/mol) nmol g/OD260
Gap1F ccacatcgctcagacaccat 66.7 55.0 20 6.0 3.0 9.0 2.0 185800.0 5990.9 5.4 32.2
Gap1R
Classification accuracy
The figure shows the value of
the discriminate function
p(Y 1  X)
f (x) log
p(Y 0  X)
across the test examples
The only test error is also the
decision with the lowest
conf
Classifying cancer types
We select a subset of the
genes (more in our feature
selection class later in the
course).
We compute the mean for
each of the genes in each of
the classes
= 1.8
= 0.6
2=
Important points
Problems with estimating full joints
Advantages of Nave Bayes assumptions
Applications to discrete and continuous cases
Problems with Nave Bayes classifiers
Possible problems with Nave
Bayes classifiers: Assumptions
In most cases, the assumption of conditional independence given
the class label is violated
 much more likely to find the word George if we
MLE for Gaussian Nave Bayes
Classifier
For each class we need to estimate one global value (prior) and
two values for each feature (mean and variance)
The prior is computed in the same way we did befo
Gaussian Bayes Classification
To determine the class when using the
Gaussian assumption we need to compute
p(xy):
P(x  y)
1
1/ 2
(2 )

exp (X
1/ 2

P(y
)T
v  x)
1
(X
Once again, we need lots of
Example
Dictionary
Washington
Congress
54. Mccain
55. Obama
56. Nader
j
54= 54(x )
j
55= 55(x )
j
56= 56(x )
=1
=1
=0
Assume we would like to classify documents
as election related or not.
Example: co
Nave Bayes Classifier
Nave Bayes classifiers assume that given the class label (Y) the
attributes are conditionally independent of each other:
p(x  y)
pi (x i  y)
i
Specific model for
atribute i
Pr
Data likelihood
The global likelihood of the data can be expressed as:
L(X,Y) L(X Y)L(Y)
Since the two parts of this product do not share parameters we can
maximize them separately.
For binary att
Feature transformation
How do we encode the set of features (words) in the document?
What type of information do we wish to represent? What can we
ignore?
Most common encoding: Bag of Words
Treat
Bayes decision rule
If we know the conditional probability P(X  Y) we
can determine the appropriate class by using
Bayes rule:
P(y i  x)
P(x  y i)P(y i) def
qi (x)
P(x)
But how do we determine
p(X
Organic Chemistry 334
Syllabus
Fall 2014
Class Meeting Information:
Lecture Days: Tuesday and Thursday
Lecture Times and Location: 8:00  9:50 AM in Hoffmann Hall 109
Tentative Course Outline: We are