I've created a systematic approach that aims at assessing certain aspects about papers that I'm reviewing.
The first phase (Context and Understanding) is meant to cut to the core of the underlying research that is being presented. It is also hoped that by analysing the papers in this fashion, justification and reflection can be undertaken when considering individual aspects.
The second phase (Methodological Issues) aims to consider the threats to the research's validity with a view to spotting problems within the research process itself.
The example column is based on a review done in Reviewing A Fast Learning Algorithm for Deep Belief Nets.
| I | Phase 1: Context and Understanding |
||
| # | Aspect | Description | Example |
| 1 | Title |
Title of the research paper |
A Fast Learning Algorithm for Deep Belief Nets |
| 2 | Topic | The general topic of the paper | Restricted Boltzmann Machines |
| 3 | Topic Details | More additional details, if needed, to better articulate the description |
A description of an algorithm that, when applied to an existing DBNs (which comprise RBMs), will improve its performance. |
| 4 | Year | Year of publication | 2006 |
| 5 | Research Area | The particular research area this topic falls under | Deep Learning (Deep Neural Networks) |
| 6 | Type of research paper | Is it a survey or a topic-specific paper | Specific topic in the research area |
| 7 | Author(s) | The authors of the paper | Hinton, G.E., Osindero, S. and Teh, Y.-W |
| 8 | Research Question | The research question the paper is addressing |
How can the performance of DBNs be improved? How can the effects of 'explaining away' be eliminated in DNNs? |
| 9 | Problem |
The problem the paper is going to solve |
The phenomenon in DBNs of 'explaining away' |
| 10 | Problem Analysis | More detailed analysis of the problem | |
| 11 | Research Deliverables, Promises and/or Objectives |
Deliberables/objectives typically suggest that data and a specific technique must be used to |
|
| 12 | Aim of research |
In what perspective does the paper aim to solve the problem? |
Describe |
| 13 | Research Aim Analysis |
To describe how an algorithm works to solve the phenomenon of explaining away, which reduces the performance of DBNs. It also explains how the algorithm can solve the problem, and explains what the phenomenon of 'explaining away' is |
|
| 14 | Philosophical worldview | What view do the authors have, eg, Objectivist | Postpositive (Objectivist), Pragmatist |
| 15 | Philosophical world view analysis |
Details, justification for the indication of a particular philosophical view. |
The author uncovers why DBNs have lower than expected performance, which is due to a phenomenon called 'explaining away'. In this respect, they figure out what it is and how to fix it, thereby revealing a previously unknown approach to this problem. This, therefore, the researchers employ a postpositivist approach to knowledge, accepting that there is knowledge/truth that must be discovered, which is what this research does. Equally, the research is also very pragmatic as demonstrated by the implementation of the algorithm in a DBN that is shown and can be of general use, particularly in this case in classificying ot generating digit characters, but such that this approach could be applied to any generative model with any data. |
| 16 | Research Design Philosophy (Empirical or Non) | What is the general philosophical approach to the design of the research undertaken in this paper? | Empirical/Quantitative (Gathering of observed/experienced data) |
| 17 | Research Designs |
Specific research design based on the underlying philosophy of the research design (Empirical or Non-Empirical) |
|
| 18 | Research Design Analysis |
The research uses experiments to showcase the results of the algorithm developed. It is therefore based on experimental and empirical observation. The Neural network used is in the testing inherently a statistical model that results in quantitative outcomes, which serve to model its results. As the approach is primarily the showcase of both the description and implementation of the algorithm, as applied to an existing approach (DBNs), it could be argued that this research is also a design science piece, as it is the development and performance of the resulting model that is explored. |
|
| 19 | Research Methods |
The process or manner in which research is conducted. Each paper has a single method of research. |
Experiment (Quantitative measurement) |
| 20 | Analysis of Research Methods |
What does the method consist of, and what are its characteristics? |
The research uses experiments that use existing character digit datasets as input the the model, and the results are then evaluated. The model produced quantitative results that represent the model, which therefore makes this research an application of quantitative measurement based on experimenting with the model (which is enhanced by the algorithm developed in this research). |
| 21 | Reasoning Process |
Where are the conclusions drawn from? Typically, interpretivist approaches use induction to start with the data to derive a theory/understanding. |
Deduction - Theory first, data later |
| 22 | Reasoning Process Analysis |
Where are the conclusions drawn from? Data or theory |
The underlying theory of 'explaining away' is discussed, and an algorithm that aims to eliminate it is implemented in a model that is tested using a dataset (MNIST). In this way, a theoretical idea is tested or proven using data. Therefore, this is a deductive process initiated with a theory and supported by an implementation of an algorithm based on that theory and applied to a model, which is evaluated with data to test whether the reasoning behind the theory is correct. |
| 23 | Answer/Solution to the problem | How does the paper solve the problem that is stated? | DBNs should be pretrained layer-by-layer, followed by fine-tuning to substantially improve learning. |
| 24 | Analysis of the Answer/Solution to the problem | Characteristics of the solution. |
DBNs should be pretrained layer-by-layer using an unsupervised algorithm, followed by a fine-tuning supervised algorithm to substantially improve learning. Use a fast learning algorithm based on complementary pairs to configure DBNs for training |
| 25 | Modes of inquiry | Scientific: Based on research procedure and empirical testing (Objectivist and subjectivist) Non-scientific: Belief, tradition, intuition | Scientific |
| 26 | Mode of Enquiry Analysis |
The science research, basing its approach on statistical methods, specifically the actual use and implementation thereof, uses a neural network, trained with the developed algorithm that is tested and shown to produce observable indications that the algorithm solves the problem presented in the research. The approach is objective, as it uses existing datasets to show the effectiveness of the algorithm (when applied to the model). The algorithm is also by definition repeatable, making this easily testable and therefore the results are not based on subjective or varying procedures. This means this process can be automated. |
|
| 27 | Data (acquired) | 1. What is the acquired material? 2. What is the data collection process? 3. What are the characteristics of this data? |
MNIST dataset: 10,000 character digits, grey-scale 32x32 images used to train the neural network (pre-trained with a new algorithm) The input dataset for the model testing is the MNIST dataset of character digits, which is a repository of 2D images that are well-known and used by researchers for the classification of character digits. The output of the model is numerical data that indicates/predicts the classification of the input data belonging to specific classes of digit characters.
|
| 28 | Analysis of Research Techniques/Tools | What are the techniques/Tools and their characteristics? |
The main technique is experimentation using a neural network with 3 hidden layers and using the developed pre-training algorithm to test how well it removes the 'explaining away' in order to improve the network's inference performance. The MNIST data set is used as training data for the network. The algorithm that is developed is applied to a DNN model (DBN), and it is then tested by experimenting on the model to see the performance that results. The results in an error rate of 1..25% in comparison to the closest rival, which is SVM at 1.4%. |
| 29 | Research techniques (data collection and analysis) | How data is gathered, analysed and inferences drawn | Experiments, Statistical analysis, Neural Networks |
| 30 | Information (Analysis) | 1. What does the data tell us? 2. What meaning is established by analysing the data? 3. What is the nature and the characteristics of this information? |
The MNIST data is processed using the neural network, resulting in output from the neural network (model). The results from the neural network show that using a pre-training algorithm that configures/trains each layer using complementary pairs improves the performance of DBNs, i.e reduces the error discrepancy in predicted vs actual outputs
|
| 31 | Knowledge (general) | 1. How has the information obtained been useful, particularly in a wider sense? 2. What is the nature and characteristics of this knowledge? |
The phenomenon of 'explaining away' that occurs in DNNs (of stacked RMBs) restricts their performance. Using complementary pairs to configure/train each layer to establish initial weights removes 'explaining away' and results in a better-performing neural network.
|
| 32 | Parent/Origin | The origin or parent paper that this paper was referenced. For example, a Survey paper | A survey of deep neural network architectures and their applications |
| 33 | Peer reviewed? | Yes | |
| 34 | Relevance, Contribution, Originality and Novelty | 1. What explicit theoretical dilemma and conceptual underpinning (not just gap) in CS has inspired this research? (See problem) 2. How would truth or fact be evaluated in this research? 3. Which concepts/models/assumptions from the discipline (CS) are integrated into my methodology? 4. What are the implications of this research for the wider field? (see research impact) 5. How do the results challenge or support existing approaches to truth/ fact and theory? |
A key aspect is that Hinton et al have identified and understood exactly what the problem of explaining away is, and so were able to create an algorithm to circumvent it. The improvements to the Performance/Learning of/neural networks as a result of this new algorithm improve the performance of all DBNs, and therefore have a great/wide applicability to all domains that use DBNs. The results of the paper are very generalizable. Another particularly interesting aspect is that the paper shows a way to determine what the model learnt by generating an image based on the learnt weights to 'see' what and how it learned the dataset.
|
| 35 | Popularity | Refer to aspects such as the vastness of citations and collaborations, etc | High |
| 36 | Popularity Analysis | What indicates the popularity? Number of citations? |
IEEE reports 3888 citations while ACM reports 3280 citations. This suggests that this is a very popular piece of research. There is also credence given to the fact that the researcher (Hinton) has an important influence on the AI research community in general. |
| 37 | Research Conclusions | Analysis of the conclusions. 1. Are they subjective? 2. Do they have credibility? | |
| 38 | Causation vs Correlation analysis | How are variables, causation and correlation dealt with in the research? |
Variables:
After using various types of comparative learning algorithms in comparison to the research's approach (which uses the greedy-layer-to-layer pre-training algorithm), the same dataset is used (MNIST) throughout, therefore only the approach to learning changes. This means each model's discrepancy error is evaluated until the lowest value is found to see which model causes the discrepancy value to be the lowest. The neural network configuration is unchanged. |
| 39 | Researcher's notes | Notes that were taken during the reading of the paper. (NB: see existing notes written on/in the papers also) |
|
| 40 | Researcher's TODO | These are actions that the researcher would like to take as a result of reading the paper |
|
| II | Phase 2: Methodological Issues (Subjectivity/Objectivity) | ||
|---|---|---|---|
| Threats to validity | |||
| 41 |
Construct Validity |
Analysis of the approach to measuring the construct under investigation. Any threats, oversights, assumptions or naivety or other risks that might affect the construct validity of this research. |
No obvious flaws |
| 42 |
Internal Validity |
Analysis of cause-and-effect within the research (Truth in the research). Any threats, oversights, assumptions or naivety or other risks that might affect the Internal validity of this research? |
See below |
|
42.1) Research Correctness |
How was subjectivity removed from the outcome, or how is objectivity ensured? E.g.. How reproducible is the research and why? |
Objectivity
Subjectivity/Specificity No obvious flaws |
|
|
42.2) Research technique |
How is subjectivity removed using the technique? What aspects of the research technique are too specific and risk degrading/compromising the real truth? |
Objectivity: Algorithms used in this research are repeatable and inherently automatable. This means all parts of the process, i.e, data, model, and algorithm, are non-varying in nature and therefore can be replicated/verified by third parties. The comparison of alternate models' performance on the NMIST dataset is suitable for evaluating how the pre-trained model's performance compares to those models that do not use it. The research techniques fit the requirements of this research. Subjectivity/Specificity:
|
|
|
42.3) Research techniques vs research question
|
Are the research techniques appropriate given the research question? |
Objectivity: Varying the application of the learning algorithm while keeping other parameters constant, i.e the common data (MNIST) and the design of the neural network design means that it is simple to evaluate the effect of varying the single variable, i.e, the application of the algorithm, making only the algorithm the independent variable. This supports the research question as the neural networks' output/performance (error function) of the classification task directly indicates if the neural network worked better than other results from other models that had not used the pre-training algorithm. Subjectivity/Specificity:
|
|
|
42.4.) Conclusion vs methods |
Are the conclusions appropriate given the methods used? What risks jeopardise the truth of the conclusions? |
Objectivity: Using experimental results based on empirical testing, observation, and comparison supports the conclusion that the research's specific approach is better than the other approaches that were tested. Subjectivity/Specificity:
|
|
| 43 |
External Validity |
Analysis of generalizability (Truth in real-life). Any threats, oversights, assumptions or naivety or other risks that might affect the External validity of this research? |
There is no evidence presented that this approach has or will generalise well to wider applications. |
| 44 |
Data Validity |
Analysis of the data gathered to represent the construct being described. Any threats, oversights, assumptions or naivety or other risks that might affect the Data validity of this research. |
See below |
| 44.1) Data subjectivity (specificity/narrowness) | How is data subjectivity eliminated? How specific is the data and its use/nature itself, limiting the impact of the research or the purported truths that the research suggests by using it. |
Objectivity: Image data for a neural network classification task is appropriate for evaluating the learning of a neural network for the classification of this data against known classification labels. The research data is also a well-known dataset that is often used for testing classification performance in models, and so it is appropriate for this type of research. Subjectivity/Specificity:
|
|
|
44.2) Data vs Research Question |
How well does the data support the research question? |
Objectivity: Varying the application of the learning algorithm while keeping other parameters constant, i.e the common data (MNIST) and the design of the neural network design means that it is simple to evaluate the effect of varying the single variable, i.e, the application of the algorithm, making only the algorithm the independent variable. This supports the research question as the neural networks' output/performance (error function) of the classification task directly indicates if the neural network worked better than other results from other models that had not used the pre-training algorithm. Subjectivity/Specificity:
|
|
| 45 |
Summary of general risks to validity |
General concerns, risks, limitations and assumptions. |
The paper is very technical
|
|
45.1) Credibility concerns
|
Are there noticeable flaws or limitations in the research? This can be high-level level such as a lack of references, large gaps in bibliography, dependency on specific authors (lack of balance) or other missing or inadequate observations/aspects that make the research less credible. |
Objectivity: There are gaps in the referenced papers; however, as this paper tests a new algorithm using experimentation and comparison with other approaches, the literature is less influential. In this respect, the literature is relatively objective. Subjectivity:
|
|
To take the information gathered and put it into a more presentable form.
An example of this approach is Reviewing A Fast Learning Algorithm for Deep Belief Nets. The outline of which looks like this:
This research review approach is currently under development, but you can use it if you'd like a place to start.