Reading Development project

From CCN Wiki
Revision as of 14:55, 27 April 2020 by Chris (talk | contribs)
Jump to navigation Jump to search

The Reading Development project uses an open dataset collected between 2008-2013 in James Booth's Developmental Child Neuroscience Lab at Northwestern. The data are available at openneuro.org.

General Notes

To be added later.

Constrained Classifier Project

Participants

28 Children (14 TD, 14 RD) with either strong or weak (incl. dyslexia diagnosis) reading skills and no other diagnosis and data from 2 time points. Children were scored according to Z score on 3 reading measures (WATT, Pseudoword decoding, something else), with an additional -1 or -0.5 assigned for a clinical dyslexia diagnosis (-1) or reading difficulty diagnosis (-.5). I selected the children with the top and bottom 14 summed scores.

Preprocessing

Anatomical surfaces reconstructed using T1w images at both early and late time points. Functional data from VV runs (word and nonword) preprocessed with both 6mm and 2mm blurring, Siemens slice time.

Functional Masking Using GLMA

A GLMA was carried out, contrasting all lexical vs. baseline to create group-level functional masks for TD and RD (p<.001, cluster P<.05). Union of masks was calculated. Masks were subdivided along template anatomical boundaries using custom scripts intersection clusters with template region boundaries. These regions were further subdivided using mris_divide_whatever.

Time Series Extraction

The 2mm-smoothed data were detrended using detrend_wmcsf.sh to remove linear trend, motion and wm and csf signal as nuisance regressors. The surfacetimecourses.sh shell script was used to extract time series in fsaverage space to plaintext files.

Functional Connectivity Estimation

Cross-mutual information (XMI) was used to calculate a functional connectivity matrix for each functional run (8 runs/participant) using a custom MATLAB script.

MATLAB FC Pseudocode

for each subject S:
   for each time series file:
      load the files for lh and rh
      merge lh and rh into a single ts matrix
      eliminate timepoints at spikes (3SD from mean)
      initialize empty adjacency matrix A
      for each ordered pair of regions (i,j) in upper triangle of A:
         calculate the number of bins required for the cross-mutual-information calculation
         calculate the cross-mutual-information for time series in regions i and j and store in A[i,j]
    append A to collection of adjacency matrices for S
 save all adjacency matrices for S to a .mat file

FC Pattern Generation

Upper triangle of XMI matrix was extracted as a vector and normalized (Z Score) and scaled to fall between 0 and 1. The resulting distribution was very positively skewed so the square root of these values was calculated to make the distribution more normal. The scaled FC vectors for each of the 8 runs per participant were tagged for lexicality (word=1, pseudoword=0) at element n-1, and reading skill (RD=0, TD=1) at element n. This generated a matrix of 224 (28 subjects * 8 runs) by 6788 (6786 connections + lexicality code + group code). The matrix was written to a single .csv file.