ADVANCED MACHINE LEARNING
CLASS PROJECT
PROF. TONY JEBARA
PRESENTATIONS ON |
APR 22, APR 27, APR 29 AND MAY 04 2016 |
WRITE
UP DUE ON |
MAY 7th 2016 BY MIDNIGHT |
1. These are 4-person team projects. It is up to you to form a team of 4 people to explore this effort and to produce a co-authored paper representing the entire team's work. Machine learning is increasingly a multi-person effort at companies (and in academia) so collaboration is a crucial skill.
2. Use the TA and Professor's office hours to discuss your project ideas and make sure that they are reasonable. Be prepared to discuss broadly what you plan to do and are doing, what results you expect, etc. If you are really stuck, look at the topics below and those covered in class to think of a direction and we can also try to help you if you come to office hours.
3.
The final presentations should be in powerpoint or pdf files
and must be no more than 10 minutes long. Since you are in a group of 4 people, you don't all have to present, one person could be the designated speaker for the group and everyone will get the same grade for the presentation and the writeup. We strongly suggest that your team does not have more than 10 slides total. Time yourself to make sure you do not ramble on for more than your allowed time. We will deduct points if you exceed your allotted time and we will also stop you if you go over your allowed time (that is how things work at real conferences).
4. After presentations, submit a write-up
in a two-column conference paper-style document as a Postscript file project.ps
or a Portable Document Format file project.pdf, whichever is more appropriate
and convenient for you to produce. Please do not send your work as a Microsoft
Office document, LaTex source code, or something more exotic. Include images
within your document as figures. Keep your total write-up no longer than 5
pages (two-column) although for 2-person projects you can write up to 8 pages. If you go over the page limits, you will also lose points (that's how conferences enforce limits). To see how to write a good paper and present it, check out this link:
http://www.cs.iastate.edu/~honavar/grad-advice.html
In particular see Simon
Peyton Jones on "How to Write a Good Research Paper". We recommend
using Latex to write up your report: http://www.latex-project.org
Submit
your homework via Courseworks. If unable to, please email it to both the TAs and Instructor.
Please tar.gz everything in your current directory and then send it to us. Make
sure you send us a write up of your results as a postscript or pdf file
containing any figures, tables and equations as well as your Matlab or C code
and scripts as separate files.
For
examples of previous year’s projects, take a look at:
http://www1.cs.columbia.edu/~jebara/6772/proj/
http://www1.cs.columbia.edu/~jebara/6998-01/projects/
(some links may be
broken, just try to follow the ones that work)
PROJECT
DESCRIPTION
Unlike the
assignments, for the projects there is no fixed recipe to follow. Rather, you
are free to pick a topic and direction that you find motivating and to leverage
the tools covered in class. Here are a few themes we suggest as well as a few
papers to look into.
B. Taskar, C. Guestrin, D. Koller http://books.nips.cc/papers/files/nips16/NIPS2003_AA04.pdf
T. Jaakkola,
M. Meila and T. Jebara http://www1.cs.columbia.edu/~jebara/papers/maxent.pdf
T. Jebara
Cutting-Plane Training of Structural SVMs
Joachims, Finlay and Yu. http://www.cs.cornell.edu/People/tj/publications/joachims_etal_09a.pdf
Structured Prediction with Relative Margin
Shivaswamy and Jebara. http://www.cs.columbia.edu/~jebara/papers/icmla09structrmm.pdf
Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data
Lafferty, McCallum and Pereira http://www.cs.columbia.edu/~jebara/6772/papers/crf.pdf
Majorization for CRFs and Latent Likelihoods
Jebara and Choromanska
http://www.cs.columbia.edu/~jebara/papers/nips2012.pdf
J. Wang, T. Jebara and S.F. Chang http://www.cs.columbia.edu/~jebara/papers/icml08.pdf
T. Jebara, J. Wang and S.F. Chang http://www.cs.columbia.edu/~jebara/papers/JebWanCha09.pdf
S. Andrews and T. Jebara http://www1.cs.columbia.edu/~jebara/papers/stu-andrews-workshop-submission-nips2007.pdf
J. Goldberger,
http://www.cs.toronto.edu/~hinton/absps/nca.pdf
B. Shaw and T. Jebara
http://www1.cs.columbia.edu/~jebara/papers/aistatsMVE07.pdf
K. Weingberger, B. Backed and L.
Saul,
http://www.seas.upenn.edu/~kilianw/publications/PDFs/kfactor_aistats05.pdf
S. Bowling, A. Ghodsi and D. Wilkinson,
http://www.machinelearning.org/proceedings/icml2005/papers/009_Action_BowlingEtAl.pdf
C. Bishop, http://www.ncrg.aston.ac.uk/Papers/postscript/NCRG_96_015.ps.Z
S. Roweis and L. Saul, http://www.sciencemag.org/cgi/reprint/290/5500/2323.pdf
S.
Mika et al., http://www.kernelmachines.org/papers/MikSchSmoMueRaeSch99.ps.gz
M. Collins et al, http://www.research.att.com/~dasgupta/pca.pdf
J. Weston
et al., http://www.ai.mit.edu/people/sayan/webPub/feature.ps
T. Jebara and T. Jaakkola, http://www.cs.columbia.edu/~jebara/papers/uai.pdf
H. Lodhi et al, http://www.support-vector.net/papers/string.ps
A kernel between sets of vectors
R. Kondor and T. Jebara
http://www.cs.columbia.edu/~jebara/papers/Kondor,Jebara_point_set.pdf
Probability Product Kernels
T. Jebara, R. Kondor and A. Howard
http://www1.cs.columbia.edu/~jebara/papers/jebara04a.pdf
Exploiting generative models in discriminative classifiers
T. Jaakkola and D. Haussler. http://www.ai.mit.edu/~tommi/papers/gendisc.ps
Density Estimation under Independent Similarly Distributed Sampling Assumptions
T. Jebara, Y. Song and K. Thadani
http://www1.cs.columbia.edu/~jebara/papers/nips07isd.pdf
R. Caruana, http://citeseer.nj.nec.com/10214.html
T. Jebara, http://jmlr.csail.mit.edu/papers/volume12/jebara11a/jebara11a.pdf
J.
Baxter, http://citeseer.nj.nec.com/baxter95learning.html
T. Dietterich and T. Bakiri, ftp.cs.orst.edu/pub/tgd/papers/jair-ecoc.ps.gz
Learning switching
linear models of human motion
V. Pavlovik, et al http://www.cc.gatech.edu/~rehg/Papers/SLDS-NIPS00.pdf
Dynamical Systems
Trees
A. Howard and T. Jebara
http://www1.cs.columbia.edu/~jebara/papers/uai04.pdf
S. Mukherjee et al, http://www.ai.mit.edu/people/girosi/home-page/nnsp97.pdf
M. Brand,
http://www.media.mit.edu/people/brand/papers/brand-chmm.ps.gz
H. Attias, http://research.microsoft.com/~hagaia/uai99.ps
W.
Penny, http://www.fil.ion.ucl.ac.uk/~wpenny/publications/vgbmm.ps
Heskes & Zoeter ftp://ftp.mbfys.kun.nl/pub/snn/pub/reports/Heskes.uai2002.ps.gz
J. Weston, et.
al., http://www.icml2006.org/icml_documents/camera-ready/127_Inference_with_the_U.pdf
D. Zhou, et.
al., http://research.microsoft.com/~denzho/papers/LLGC.pdf
T. Joachims, http://www-ai.cs.uni-dortmund.de/DOKUMENTE/Joachims_99c.ps.gz
P. Shivaswamy and T. Jebara, http://www.cs.columbia.edu/~jebara/papers/nips08.pdf
M. Tipping, ftp.research.microsoft.com/users/mtipping/rvm_nips.ps.gz
Estimating
the Support of a High-Dimensional Distribution.
Scholkopf, et.
al. Microsoft Technical Report, MSR-TR-99-87. 1999.
Come see me for the hardcopy of the paper.
T. Jebara and Y. Bengio, http://www1.cs.columbia.edu/~jebara/papers/snowbird3.pdf
J. Tenenbaum and W. Freeman, http://www.merl.com/reports/docs/TR99-04.pdf
B. Frey and N. Jojic, http://www.psi.toronto.edu/~frey/papers/tmg-cvpr99.ps.Z
T.
Jebara, http://www.cs.columbia.edu/~jebara/papers/permkern.pdf
Transformation Invariance in Pattern Recognition
Simard, et al http://yann.lecun.com/exdb/publis/psgz/simard-00.ps.gz
N.
Friedman et al, http://www.cs.huji.ac.il/~noamm/publications/UAI2001.ps.gz
T. Jebara
and V. Shchogolev. http://www1.cs.columbia.edu/~jebara/papers/bmatching.pdf
S. Arora, S. Rao, U. Vazirani. http://www.cs.princeton.edu/~arora/pubs/arvstoc.pdf
F. Bach and M. Jordan, http://cmm.ensmp.fr/~bach/kernelICA-jmlr.pdf
T. Jebara, http://www.cs.columbia.edu/~jebara/papers/uai08tree.pdf
T. Hofmann, http://www.cs.brown.edu/people/th/papers/Hofmann-UAI99.pdf
Or... any topic you can convince us
about and involves advanced machine learning techniques! In particular, this would
be a method published in a top machine learning conference in the past 15 years.
Feel free
to also bring new papers to the list below and suggest them as well. Places to
look for papers include recent machine learning conferences such as:
and
some machine learning journals like the Journal of Machine Learning Research,
Journal of Artificial Intelligence Research, Machine Learning, Pattern
Recognition, Neural Computation, IEEE Transactions on Pattern
Analysis and Machine Intelligence and so forth. Many recent articles from
these compilations are available online or in the library. You can find copies
of the papers (postscript and pdf) through Citeseer, a popular search engine
for computer science publications: http://citeseer.nj.nec.com/cs
What are examples of bad choices for projects? Anything that only
involves easy algorithms from the introductory class (just logistic
regression, just SVMs, just HMMs, just perceptrons, just EM for
mixtures of Gaussians, just junction tree algorithm, etc.). These are
ok methods to use as baselines to compare with while you develop or
implement a better method but the whole point is to go beyond the
things we learned about in 4771. Also, do not waste too much time
setting up and presenting on the domain of your problem (say
motivating and setting up a problem from finance, genomics,
etc.). This course is about the machine learning side of things and
not about the domains.
Potential
datasets on which to try some of your learning algorithms:
http://www1.ics.uci.edu/~mlearn/MLRepository.html
Stanford Large Network Dataset Collection:
http://snap.stanford.edu/data/
http://www.cs.toronto.edu/~delve/
http://www-personal.buseco.monash.edu.au/~hyndman/TSDL/