Beruflich Dokumente
Kultur Dokumente
Dept. of Computer Science Dept. of Computer Science
Dept. of Computer Science Dept. of Computer Science
Types of reviews Conference reviewing
• Conferences: • Typical review load: 10 papers/reviewer
– one-shot
– accept/reject • Reading and understanding 10 papers takes 10
– few modifications hours of quality time
• Journals:
• Most reviewers do not have this time, so they
– iteration (major revision – minor revision – accept)
apply filtering and do not waste time on papers
– significant rewrites to improve quality
that are not acceptable anyway
Dept. of Computer Science Dept. of Computer Science
Dept. of Computer Science Dept. of Computer Science
Reject Accept
• Yes/No questions: • Matter of degree:
– Is the paper complete (checklist) ? – Is the work novel? Are these just someone else’s
– Is the result correct? ideas in a different notion?
– Did you learn something from it? (difficult to satisfy) – Is the problem important?
– Is the work significant and difficult to obtain?
• If any of these is no, reject
• Yes, useful for ranking (weak/strong accept)
Dept. of Computer Science Dept. of Computer Science
Dept. of Computer Science Dept. of Computer Science
Journal reviewing 2. How to write a paper
• Journals allow for iterations • Checklist
– Problem
• Same filters as for conferences, but important to – Solution
help author improve the paper – Example
– Evaluation
• Can expect significant rewrites/additional work
• Title, abstract, introduction, methodology, results,
• Most journal submissions are eventually discussion, conclusion
published somewhere
Dept. of Computer Science Dept. of Computer Science
Title Abstract
• Title (cover the most important information): • Abstract:
– New problem (new perspective) – Problem?
– New findings – Solution or methodology?
– New methodology – Results?
• Example:
• Example: – Higher-order structural dependencies such as
– Higher-order structural topic modeling in document networks coauthor relations exist widely in document networks.
(Zeng, 2010) However, their use in topic modeling problems is
– Physical signals for protein-DNA recognition (Zeng, 2009) significantly hampered in practice by the intractable
– Cascade Markov random fields for stroke extraction of Chinese complexity of modeling and optimizing the long-range
characters (Zeng, 2010) topic dependence. (Background)
Dept. of Computer Science Dept. of Computer Science
Abstract Abstract
• Based on Markov random fields, we address the • To achieve the approximate inference of the
higher-order structural topic modeling (HSTM) best topic configuration, we develop a hybrid
problem with learned higher-order clique belief propagation and Gibbs sampling algorithm
potentials that capture complex topic structures. for HSTM. Experimental results on DBLP and
We observe that higher-order topic structures MEDLINE document datasets show that the
encountered in document networks are often higher-order structural dependencies among
sparse, i.e., many topic configurations of a coauthors and documents can enhance the
higher-order clique are equally unlikely and thus overall topic summarization performance
need not to be encouraged throughout the compared with state-of-the-art topic models for
network. (Problem) only pairwise relations. (Methodology + Results)
Dept. of Computer Science Dept. of Computer Science
Introduction Methodology
• Literature review • The key part of your paper:
– Why this problem is significant? – Theory
– Recent solutions? – Mathematical formulation
– What inspires this idea? – Algorithms
– Computational issues
• Deep understanding of your problem helps a lot – Implementation issues
Dept. of Computer Science Dept. of Computer Science
Results and discussion 3. Case study
• Data sets and performance measures • ICML 2010, Phase I, two reviewers, pass initial
filtering
• Experimental settings • Reviewer 1:
– The author proposed a novel topic model that
incorporates higher-order relationship among
• Comparison with state-of-the-art methods documents/authors. ……
– The author then proposed a new inference algorithm
which integrates belief propagation and Gibbs
• When your method is better and when your
sampling.
method does not work
Dept. of Computer Science Dept. of Computer Science
Dept. of Computer Science Dept. of Computer Science
Responses to reviewer 1 (Continued) Reviewer 2
• Generally, higher-order MRFs scale badly since • The model seemed quite similar to Daume's MRTF (both
K^N topic configurations need to be considered. in structure, and in results). Daume's model seemed to
have computational issues, so anything the current work
But we find that in real document networks the
did to resolve that would be a plus.
propagated higher-order topic patterns are
SPARSE in Fig 3. Sparsity not only reveals the
• I didn't exactly follow the justification for the model in the
specificity of higher-order relations in topic introduction. I would have been helped by a specific
modeling, but also significantly reduces the example (in words) of where higher-order relations are
complexity of learning and inference (see Sec 3). better than pair-wise relations.
As a result, our model scales well with large
number of topics in practice.
Dept. of Computer Science Dept. of Computer Science
Dept. of Computer Science Dept. of Computer Science
Responses to reviewer 2 (Continued) Responses to reviewer 2 (Continued)
• Fig 5 shows the log likelihood per word, where 0.1
corresponds to a significant improvement.
• Fig 1 shows an example why higher-order
• Previous work (Chang 09) used the total log likelihood,
relations provide more specific constraints (since which scaled per word likelihood by the total number of
simply using {y5,y1}, {y6,y1} to represent the words. (Chang 09) shows an improved link prediction
coauthor relation of {y1,y5,y6} cannot distinguish than (Nallapati 08), so the comparison with (Chang 09) is
with the situation of y5 and y6 separately sufficient to demonstrate the improved link prediction.
coauthored with y1 writing two different papers). • Because we compare the extracted topic features for
document classification with pairwise topic models, we
need not to show the SVM results using original word
vectors.
Dept. of Computer Science Dept. of Computer Science
Any questions?
Thank you!
Dept. of Computer Science Dept. of Computer Science