I've created a systematic approach that aims at assessing certain aspects about papers that I'm reviewing.

The first phase (Context and Understanding) is meant to cut to the core of the underlying research that is being presented. It is also hoped that by analysing the papers in this fashion, justification and reflection can be undertaken when considering individual aspects.

The second phase (Methodological Issues) aims to consider the threats to the research's validity with a view to spotting problems within the research process itself.

The example column is based on a review done in Reviewing A Fast Learning Algorithm for Deep Belief Nets

 

I Phase 1: Context and Understanding
# Aspect Description Example
1 Title

Title of the research paper

A Fast Learning Algorithm for Deep Belief Nets

2 Topic The general topic of the paper Restricted Boltzmann Machines
3 Topic Details More additional details, if needed, to better articulate the description

A description of an algorithm that, when applied to an existing DBNs (which comprise RBMs), will improve its performance.

4 Year Year of publication 2006
5 Research Area The particular research area this topic falls under Deep Learning (Deep Neural Networks)
6 Type of research paper Is it a survey or a topic-specific paper Specific topic in the research area
7 Author(s) The authors of the paper Hinton, G.E., Osindero, S. and Teh, Y.-W
8 Research Question The research question the paper is addressing

How can the performance of DBNs be improved? How can the effects of 'explaining away' be eliminated in DNNs?

9 Problem

The problem the paper is going to solve

The phenomenon in DBNs of 'explaining away'
10 Problem Analysis More detailed analysis of the problem  
11 Research Deliverables, Promises and/or Objectives

Deliberables/objectives typically suggest that data and a specific technique must be used to

  1. Show how using 'complementary priors' removes explaining away.

  2. Derive an unsupervised learning algorithm that uses complementary priors

  3. Describe a hybrid neural network that uses associative memory and 3 hidden layers

  4. Use an algorithm on a neural network to prove the algorithm's effectiveness

  5. The performance of the model is FAST and ACCURATE

12 Aim of research

In what perspective does the paper aim to solve the problem?

Describe
13 Research Aim Analysis  

To describe how an algorithm works to solve the phenomenon of explaining away, which reduces the performance of DBNs.

It also explains how the algorithm can solve the problem, and explains what the phenomenon of 'explaining away' is

14 Philosophical worldview What view do the authors have, eg, Objectivist Postpositive (Objectivist), Pragmatist
15 Philosophical world view analysis  

Details, justification for the indication of a particular philosophical view.

The author uncovers why DBNs have lower than expected performance, which is due to a phenomenon called 'explaining away'. In this respect, they figure out what it is and how to fix it, thereby revealing a previously unknown approach to this problem. This, therefore, the researchers employ a postpositivist approach to knowledge, accepting that there is knowledge/truth that must be discovered, which is what this research does.

Equally, the research is also very pragmatic as demonstrated by the implementation of the algorithm in a DBN that is shown and can be of general use, particularly in this case in classificying ot generating digit characters, but such that this approach could be applied to any generative model with any data.

16 Research Design Philosophy (Empirical or Non) What is the general philosophical approach to the design of the research undertaken in this paper?  Empirical/Quantitative (Gathering of observed/experienced data)
17 Research Designs

Specific research design based on the underlying philosophy of the research design (Empirical or Non-Empirical)

  • Design Science Research (Empirical)
  •  Experimental (Empirical)
  • Statistical Modelling (Empirical)
  • Algorithms design (Non-Empirical)
18 Research Design Analysis  

The research uses experiments to showcase the results of the algorithm developed. It is therefore based on experimental and empirical observation.

The Neural network used is in the testing inherently a statistical model that results in quantitative outcomes, which serve to model its results.

As the approach is primarily the showcase of both the description and implementation of the algorithm, as applied to an existing approach (DBNs), it could be argued that this research is also a design science piece, as it is the development and performance of the resulting model that is explored.

19 Research Methods

The process or manner in which research is conducted. Each paper has a single method of research.

Experiment (Quantitative measurement)
20 Analysis of Research Methods

What does the method consist of, and what are its characteristics?

The research uses experiments that use existing character digit datasets as input the the model, and the results are then evaluated. The model produced quantitative results that represent the model, which therefore makes this research an application of quantitative measurement based on experimenting with the model (which is enhanced by the algorithm developed in this research).

21 Reasoning Process

Where are the conclusions drawn from? Typically, interpretivist approaches use induction to start with the data to derive a theory/understanding.

Deduction - Theory first, data later
22 Reasoning Process Analysis

Where are the conclusions drawn from? Data or theory

The underlying theory of 'explaining away' is discussed, and an algorithm that aims to eliminate it is implemented in a model that is tested using a dataset (MNIST). In this way, a theoretical idea is tested or proven using data. Therefore, this is a deductive process initiated with a theory and supported by an implementation of an algorithm based on that theory and applied to a model, which is evaluated with data to test whether the reasoning behind the theory is correct.

23 Answer/Solution to the problem How does the paper solve the problem that is stated? DBNs should be pretrained layer-by-layer, followed by fine-tuning to substantially improve learning.
24 Analysis of the Answer/Solution to the problem Characteristics of the solution.

DBNs should be pretrained layer-by-layer using an unsupervised algorithm, followed by a fine-tuning supervised algorithm to substantially improve learning.

Use a fast learning algorithm based on complementary pairs to configure DBNs for training

25 Modes of inquiry Scientific:  Based on research procedure and empirical testing (Objectivist and subjectivist) Non-scientific: Belief, tradition, intuition Scientific
26 Mode of Enquiry Analysis  

The science research, basing its approach on statistical methods, specifically the actual use and implementation thereof, uses a neural network, trained with the developed algorithm that is tested and shown to produce observable indications that the algorithm solves the problem presented in the research.

The approach is objective, as it uses existing datasets to show the effectiveness of the algorithm (when applied to the model). The algorithm is also by definition repeatable, making this easily testable and therefore the results are not based on subjective or varying procedures. This means this process can be automated.

27 Data (acquired) 1. What is the acquired material? 2. What is the data collection process? 3. What are the characteristics of this data?

 

MNIST dataset: 10,000 character digits, grey-scale 32x32 images used to train the neural network (pre-trained with a new algorithm)

The input dataset for the model testing is the MNIST dataset of character digits, which is a repository of 2D images that are well-known and used by researchers for the classification of character digits. The output of the model is numerical data that indicates/predicts the classification of the input data belonging to specific classes of digit characters.

 

28 Analysis of Research Techniques/Tools What are the techniques/Tools and their characteristics?
  • Contrastive divergence algorithm

  • Gibbs sampling (Hidden Markov Monteo carlo method)

  • Greedy-layer-by-layer algorithm (complementary pairs)

  • DBN (Deep Belief Network)

  • Various learning algorithms (Backpropagation, SVMs, squared error and online updates, LeNet5 CNN, cross entropy, etc.)

  • Generation of an image using a learnt model

The main technique is experimentation using a neural network with 3 hidden layers and using the developed pre-training algorithm to test how well it removes the 'explaining away' in order to improve the network's inference performance. The MNIST data set is used as training data for the network.

The algorithm that is developed is applied to a DNN model (DBN), and it is then tested by experimenting on the model to see the performance that results. The results in an error rate of 1..25% in comparison to the closest rival, which is SVM at 1.4%.

29 Research techniques (data collection and analysis) How data is gathered, analysed and inferences drawn Experiments, Statistical analysis, Neural Networks
30 Information (Analysis) 1. What does the data tell us? 2. What meaning is established by analysing the data? 3. What is the nature and the characteristics of this information?

The MNIST data is processed using the neural network, resulting in output from the neural network (model). 

The results from the neural network show that using a pre-training algorithm that configures/trains each layer using complementary pairs improves the performance of DBNs, i.e reduces the error discrepancy in predicted vs actual outputs

 

31 Knowledge (general) 1. How has the information obtained been useful, particularly in a wider sense? 2. What is the nature and characteristics of this knowledge?

The phenomenon of 'explaining away' that occurs in DNNs (of stacked RMBs) restricts their performance.

Using complementary pairs to configure/train each layer to establish initial weights removes  'explaining away' and results in a better-performing neural network.

 

32 Parent/Origin The origin or parent paper that this paper was referenced. For example, a Survey paper A survey of deep neural network architectures and their applications
33 Peer reviewed?   Yes
34 Relevance, Contribution, Originality and Novelty 1. What explicit theoretical dilemma and conceptual underpinning (not just gap) in CS has inspired this research? (See problem) 2. How would truth or fact be evaluated in this research? 3. Which concepts/models/assumptions from the discipline (CS) are integrated into my methodology? 4. What are the implications of this research for the wider field? (see research impact) 5. How do the results challenge or support existing approaches to truth/ fact and theory?

 

A key aspect is that Hinton et al have identified and understood exactly what the problem of explaining away is, and so were able to create an algorithm to circumvent it. 

The improvements to the Performance/Learning of/neural networks as a result of this new algorithm improve the performance of all DBNs, and therefore have a great/wide applicability to all domains that use DBNs. The results of the paper are very generalizable.

Another particularly interesting aspect is that the paper shows a way to determine what the model learnt by generating an image based on the learnt weights to 'see' what and how it learned the dataset. 

 

35 Popularity Refer to aspects such as the vastness of citations and collaborations, etc High
36 Popularity Analysis What indicates the popularity? Number of citations?

IEEE reports 3888 citations while ACM reports 3280 citations. This suggests that this is a very popular piece of research.

There is also credence given to the fact that the researcher (Hinton) has an important influence on the AI research community in general.

IEEE: https://ieeexplore.ieee.org/document/6796673

ACM: https://dl.acm.org/doi/10.1162/neco.2006.18.7.1527

37 Research Conclusions Analysis of the conclusions. 1. Are they subjective? 2. Do they have credibility?  
38 Causation vs Correlation analysis How are variables, causation and correlation dealt with in the research?

Variables:

  1. Learning approach (with or without greedy layer-by-layer algorithm)
  2. MNIST dataset
  3. Neural network configuration

After using various types of comparative learning algorithms in comparison to the research's approach (which uses the greedy-layer-to-layer pre-training algorithm), the same dataset is used (MNIST) throughout, therefore only the approach to learning changes. This means each model's discrepancy error is evaluated until the lowest value is found to see which model causes the discrepancy value to be the lowest. The neural network configuration is unchanged.

39 Researcher's notes Notes that were taken during the reading of the paper. (NB: see existing notes written on/in the papers also)
  • It's interesting to think of an observation as a result (situation/response) of possible causes (weighted inputs/events)

    1. This might be a way to model dealing with stimulus and response pairs as events

40 Researcher's TODO These are actions that the researcher would like to take as a result of reading the paper
  1. Learn more about generative models

    1. It seems that I'm more aware of discriminatory models.

  2. Find out what maximum likelihood and contrastive divergence learning are

  3. What is a function of a discrete variable?

  4. Understand what the problem of explaining away really is

 

II Phase 2: Methodological Issues (Subjectivity/Objectivity)
  Threats to validity
41

Construct Validity

Analysis of the approach to measuring the construct under investigation. Any threats, oversights, assumptions or naivety or other risks that might affect the construct validity of this research.

No obvious flaws
42

Internal Validity

Analysis of cause-and-effect within the research (Truth in the research). Any threats, oversights, assumptions or naivety or other risks that might affect the Internal validity of this research?

See below
 

42.1) Research Correctness

How was subjectivity removed from the outcome, or how is objectivity ensured? E.g.. How reproducible is the research and why?

Objectivity

    1. The same constant data is used as was used by others (MNIST)

    2. The research uses an objective measure of performance (inference error) using the dataset shows that its error rate is better with this model than with previous models.

Subjectivity/Specificity

No obvious flaws

 

42.2) Research technique

How is subjectivity removed using the technique? What aspects of the research technique are too specific and risk degrading/compromising the real truth?

 

Objectivity:

Algorithms used in this research are repeatable and inherently automatable. This means all parts of the process, i.e, data, model, and algorithm, are non-varying in nature and therefore can be replicated/verified by third parties. 

The comparison of alternate models' performance on the NMIST dataset is suitable for evaluating how the pre-trained model's performance compares to those models that do not use it. The research techniques fit the requirements of this research.

Subjectivity/Specificity:

  1. The research techniques are limited to only used on 2D image data (pixels)
  2. Only a 3-layer neural network is used 

 

 

42.3) Research techniques vs research question

 

Are the research techniques appropriate given the research question? 

 

Objectivity:

Varying the application of the learning algorithm while keeping other parameters constant, i.e the common data (MNIST) and the design of the neural network design means that it is simple to evaluate the effect of varying the single variable, i.e, the application of the algorithm, making only the algorithm the independent variable. 

This supports the research question as the neural networks' output/performance (error function) of the classification task directly indicates if the neural network worked better than other results from other models that had not used the pre-training algorithm. 

Subjectivity/Specificity:

  1. The performance is only measured using data that reflects 2D character images. Larger images or more complex images are not assessed.
  2. Only a specific configuration/design of the DBN (3-layer is used) to remove the effects of 'explaining away'

 

 

42.4.) Conclusion vs methods

Are the conclusions appropriate given the methods used? What risks jeopardise the truth of the conclusions?

 

Objectivity:

Using experimental results based on empirical testing, observation, and comparison supports the conclusion that the research's specific approach is better than the other approaches that were tested.

Subjectivity/Specificity:

  1. Experiments were based on only 2D image data (pixels) so the conclusions can only be representative of character-based image data

 

43

External Validity

Analysis of generalizability (Truth in real-life). Any threats, oversights, assumptions or naivety or other risks that might affect the External validity of this research?

There is no evidence presented that this approach has or will generalise well to wider applications.
44

Data Validity

Analysis of the data gathered to represent the construct being described. Any threats, oversights, assumptions or naivety or other risks that might affect the Data validity of this research.

See below

  44.1) Data subjectivity (specificity/narrowness) How is data subjectivity eliminated? How specific is the data and its use/nature itself, limiting the impact of the research or the purported truths that the research suggests by using it.

Objectivity:

Image data for a neural network classification task is appropriate for evaluating the learning of a neural network for the classification of this data against known classification labels.

The research data is also a well-known dataset that is often used for testing classification performance in models, and so it is appropriate for this type of research.

Subjectivity/Specificity:

  1. Only the MNIST dataset is used, so the data used to represent the solution presented in this research is limited.

  2. This limits the research's outcomes and approaches to dealing with small geometric character recognition.

 

44.2) Data vs Research Question

How well does the data support the research question?

Objectivity:

Varying the application of the learning algorithm while keeping other parameters constant, i.e the common data (MNIST) and the design of the neural network design means that it is simple to evaluate the effect of varying the single variable, i.e, the application of the algorithm, making only the algorithm the independent variable. 

This supports the research question as the neural networks' output/performance (error function) of the classification task directly indicates if the neural network worked better than other results from other models that had not used the pre-training algorithm. 

Subjectivity/Specificity:

  1. Only 2D character digits, pixel information is used to show how the techniques in the research improve inference performance. 
45

Summary of general risks to validity

General concerns, risks, limitations and assumptions.

The paper is very technical

  1. It relies on an understanding of many different ideas and processes such draw deeply on existing knowledge.

  2. Those inexperienced researchers may find it difficult to validate construct and internal validity without being well acquainted with the theory, algorithms and approaches discussed.

 

45.1) Credibility concerns

 

Are there noticeable flaws or limitations in the research? This can be high-level level such as a lack of references, large gaps in bibliography, dependency on specific authors (lack of balance) or other missing or inadequate observations/aspects that make the research less credible.

Objectivity:

There are gaps in the referenced papers; however, as this paper tests a new algorithm using experimentation and comparison with other approaches, the literature is less influential. In this respect, the literature is relatively objective.

Subjectivity:

  1. The research is 19 years old (2025) and techniques here could be outdated or have been improved by subsequent research possibly making this research deprecated.

 

To take the information gathered and put it into a more presentable form.

An example of this approach is Reviewing A Fast Learning Algorithm for Deep Belief Nets. The outline of which looks like this:

This research review approach is currently under development, but you can use it if you'd like a place to start.