spatsoc
can be used in social network analysis to generate gambit of the group format data from GPS relocation data, perform data stream randomization and generate group by individual matrices.
Gambit of the group format data is generated using the grouping functions:
group_times
group_pts
group_lines
group_polys
Data stream randomization is performed using the randomizations
function.
Group by individual matrices are generated using the get_gbi
function.
Note: edge list generating functions are also available and are described in the vignette Using edge list generating functions and dyad_id.
spatsoc provides users with one temporal (group_times
) and three spatial (group_pts
, group_lines
, group_polys
) functions to generate gambit of the group data from GPS relocations. Users can consider spatial grouping at three different scales combined with an appropriate temporal grouping threshold. The gambit of the group data is then used to generate a group by individual matrix and build the network.
spatsoc
expects a data.table
for all DT
arguments and date time columns to be formatted POSIXct
.
## Load packages
library(spatsoc)
library(data.table)
library(asnipe)
library(igraph)
## Read data as a data.table
DT <- fread(system.file("extdata", "DT.csv", package = "spatsoc"))
## Cast datetime column to POSIXct
DT[, datetime := as.POSIXct(datetime)]
## Calculate the year of the relocation
DT[, yr := year(datetime)]
Next, we will group relocations temporally with group_times
and spatially with one of group_pts
, group_lines
, group_polys
. Note: these are mutually exclusive, only select one spatial grouping function at a time.
group_pts
Point based grouping by calculating distance between relocations in each timegroup. Depending on species and study system, relevant temporal and spatial grouping thresholds are used. In this case, relocations within 5 minutes and 50 meters are grouped together.
## Temporal groups
group_times(DT, datetime = 'datetime', threshold = '5 minutes')
## Spatial groups
group_pts(
DT,
threshold = 50,
id = 'ID',
coords = c('X', 'Y'),
timegroup = 'timegroup'
)
group_lines
Line based grouping by measuring intersection of, optionally buffered, trajectories for each individual in each timegroup. Longer temporal thresholds are used to measure, for example, intersecting daily trajectories.
# EPSG code for relocations
utm <- 32736
## Group relocations by julian day
group_times(DT, datetime = 'datetime', threshold = '1 day')
## Group lines for each individual and julian day
group_lines(
DT,
threshold = 50,
projection = utm,
id = 'ID',
coords = c('X', 'Y'),
timegroup = 'timegroup',
sortBy = 'datetime'
)
group_polys
Polygon based grouping by generating home ranges using adehabitatHR
and measuring intersection or proportional overlap. Longer temporal thresholds are used to create seasonal, monthly, yearly home ranges.
# EPSG code for relocations
utm <- 32736
## Option 1: area = FALSE and home range intersection 'group' column added to DT
group_polys(
DT,
area = FALSE,
hrType = 'mcp',
hrParams = list(percent = 95),
projection = utm,
id = 'ID',
coords = c('X', 'Y')
)
## Option 2: area = TRUE
# results must be assigned to a new variable
# data.table returned has ID1, ID2 and proportion and area overlap
areaDT <- group_polys(
DT,
area = TRUE,
hrType = 'mcp',
hrParams = list(percent = 95),
projection = utm,
id = 'ID',
coords = c('X', 'Y')
)
Once we’ve generated groups using group_times
and one of the spatial grouping functions, we can generate a group by individual matrix.
The following code chunk showing get_gbi
can be used for outputs from any of group_pts
, group_lines
or group_polys(area = FALSE)
. For the purpose of this vignette however, we will consider the outputs from group_pts
(2. a)) for the following code chunk.
Note: we show this example creating the group by individual matrix and network for only 2016 to illustrate how spatsoc
can be used for simpler data with no splitting of temporal or spatial subgroups (e.g.: yearly, population). See the random network section for how to use spatsoc
in social network analysis for multi-year or other complex data.
get_gbi
## Subset DT to only year 2016
subDT <- DT[yr == 2016]
## Generate group by individual matrix
# group column generated by spatsoc::group_pts
gbiMtrx <- get_gbi(DT = subDT, group = 'group', id = 'ID')
Note: spatsoc::get_gbi
is identical in function to asnipe::get_group_by_individual
, but is more efficient (some benchmarks measuring >10x improvements) thanks to data.table::dcast
.
asnipe::get_network
Next, we can use asnipe::get_network
to build the observed social network. Ensure that the argument “data_format” is “GBI”. Use other arguments that are relevant to your analysis, here we calculate a Simple ratio index.
## Generate observed network
net <- get_network(gbiMtrx,
data_format = "GBI",
association_index = "SRI")
Three types of data stream randomization are provided by spatsoc
’s randomizations
function:
The results of randomizations
must be assigned. The function returns the id
and datetime
columns provided (and anything provided to splitBy
). In addition, columns ‘observed’ and ‘iteration’ are returned indicating observed rows and which iteration rows correspond to (where 0 is the observed).
As with spatial grouping functions, these methods are mutually exclusive. Pick one type
and rebuild the network after randomization.
Note: the coords
argument is only required for trajectory type randomization, since after randomizing with this method, the ‘coords’ are needed to redo spatial grouping (with group_pts
, group_lines
or group_polys
).
type = 'step'
'step'
randomizes identities of relocations between individuals within each time step. The datetime
argument expects an integer group generated by group_times
. The group
argument expects the column name of the group generated from the spatial grouping functions.
Four columns are returned when type = 'step'
along with id
, datetime
and splitBy
columns:
# Calculate year column to ensure randomization only occurs within years since data spans multiple years
DT[, yr := year(datetime)]
## Step type randomizations
# providing 'timegroup' (from group_times) as datetime
# splitBy = 'yr' to force randomization only within year
randStep <- randomizations(
DT,
type = 'step',
id = 'ID',
group = 'group',
coords = NULL,
datetime = 'timegroup',
iterations = 3,
splitBy = 'yr'
)
type = 'daily'
'daily'
randomizes identities of relocations between individuals within each day. The datetime
argument expects a datetime POSIXct
format column.
Four columns are returned when type = 'daily'
along with id
, datetime
and splitBy
columns:
# Calculate year column to ensure randomization only occurs within years since data spans multiple years
DT[, yr := year(datetime)]
## Daily type randomizations
# splitBy = 'yr' to force randomization only within year
randDaily <- randomizations(
DT,
type = 'daily',
id = 'ID',
group = 'group',
coords = NULL,
datetime = 'datetime',
splitBy = 'yr',
iterations = 20
)
type = 'trajectory'
'trajectory'
randomizes daily trajectories within individuals (Spiegel et al. 2016). The datetime
argument expects a datetime POSIXct
format column.
Five columns are returned when type = 'trajectory'
along with id
, datetime
and splitBy
columns:
# Calculate year column to ensure randomization only occurs within years since data spans multiple years
DT[, yr := year(datetime)]
## Trajectory type randomization
randTraj <- randomizations(
DT,
type = 'trajectory',
id = 'ID',
group = NULL,
coords = c('X', 'Y'),
datetime = 'datetime',
splitBy = 'yr',
iterations = 20
)
Once we’ve randomized the data stream with randomizations
, we can build the random network.
We will use the get_gbi
function directly when type
is either ‘step’ or ‘daily’. For type = 'trajectory'
, we will recalculate spatial groups with one of group_pts
, group_lines
, group_polys
for the randomized data. In this case, the example shows group_pts
.
Since we want to create a group by individual matrix for each random iteration (and in this case, each year), we will use mapply
to work on subsets of the randomized data.
Note: building the random networks depends on the type
used and therefore, the following chunks are mutually exclusive. Use the one that corresponds to the randomization type you used above.
type = 'step'
randomizations
with type = 'step'
returns a ‘randomID’ which should be used instead of the observed ‘ID’ to generate the group by individual matrix.
After get_gbi
, we use asnipe::get_network
to build the random network.
## Create a data.table of unique combinations of iteration and year, excluding observed rows
iterYearLs <- unique(randStep[!(observed), .(iteration, yr)])
## Generate group by individual matrix
# for each combination of iteration number and year
# 'group' generated by spatsoc::group_pts
# 'randomID' used instead of observed ID (type = 'step')
gbiLs <- mapply(FUN = function(i, y) {
get_gbi(randStep[iteration == i & yr == y],
'group', 'randomID')
},
i = iterYearLs$iter,
y = iterYearLs$yr,
SIMPLIFY = FALSE
)
## Generate a list of random networks
netLs <- lapply(gbiLs, FUN = get_network,
data_format = "GBI", association_index = "SRI")
type = 'daily'
randomizations
with type = 'step'
returns a ‘randomID’ which should be used instead of the observed ‘ID’ to generate the group by individual matrix.
After get_gbi
, we use asnipe::get_network
to build the random network.
In this case, we will generate a fake column representing a “population” to show how we can translate the mapply
chunk above to three (or more variables).
## Generate fake population
randDaily[, population := sample(1:2, .N, replace = TRUE)]
## Create a data.table of unique combinations of iteration, year, and population, excluding observed rows
iterYearLs <- unique(randStep[!(observed), .(iteration, yr, population)])
## Generate group by individual matrix
# for each combination of iteration number and year
# 'group' generated by spatsoc::group_pts
# 'randomID' used instead of observed ID (type = 'step')
gbiLs <- mapply(FUN = function(i, y, p) {
get_gbi(randDaily[iteration == i &
yr == y & population == p],
'group', 'randomID')
},
i = iterYearLs$iter,
y = iterYearLs$yr,
p = iterYearLs$population,
SIMPLIFY = FALSE
)
## Generate a list of random networks
netLs <- lapply(gbiLs, FUN = get_network,
data_format = "GBI", association_index = "SRI")
type = 'trajectory'
randomizations
with type = 'trajectory'
returns a random date time which should be used instead of the observed date time to generate random gambit of the group data.
First, we pass the randomized data to group_times
using the random date time for datetime
.
After get_gbi
, we use asnipe::get_network
to build the random network.
## Randomized temporal groups
# 'datetime' is the randomdatetime produced by randomizations(type = 'trajectory')
group_times(randTraj, datetime = 'randomdatetime', threshold = '5 minutes')
## Randomized spatial groups
# 'iteration' used in splitBy to ensure only points within each iteration are grouped
group_pts(randTraj, threshold = 50, id = 'ID', coords = c('X', 'Y'),
timegroup = 'timegroup', splitBy = 'iteration')
## Create a data.table of unique combinations of iteration and year, excluding observed rows
iterYearLs <- unique(randStep[!(observed), .(iteration, yr)])
## Generate group by individual matrix
# for each combination of iteration number and year
# 'group' generated by spatsoc::group_pts
# 'ID' used since datetimes were randomized within individuals
gbiLs <- mapply(FUN = function(i, y) {
get_gbi(randTraj[iteration == i & yr == y],
'group', 'ID')
},
i = iterYearLs$iter,
y = iterYearLs$yr,
SIMPLIFY = FALSE
)
## Generate a list of random networks
netLs <- lapply(gbiLs, FUN = get_network,
data_format = "GBI", association_index = "SRI")
Finally, we can calculate some network metrics. Please note that there are many ways of interpreting, analyzing and measuring networks, so this will simply show one option.
To calculate observed network metrics, use the network (net
) produced in 4. from 2016 data.
## Generate graph
g <- graph.adjacency(net, 'undirected',
diag = FALSE, weighted = TRUE)
## Metrics for all individuals
observed <- data.table(
centrality = evcent(g)$vector,
strength = graph.strength(g),
degree = degree(g),
ID = names(degree(g)),
yr = subDT[, unique(yr)]
)
With the list of random networks from 6., we can generate a list of graphs with igraph::graph.adjacency
(for example) and calculate random network metrics.
This example uses the netLs
generated by 6. a) which was split by year and iteration.
## Generate graph and calculate network metrics
mets <- lapply(seq_along(netLs), function(n) {
g <- graph.adjacency(netLs[[n]], 'undirected',
diag = FALSE, weighted = TRUE)
data.table(
centrality = evcent(g)$vector,
strength = graph.strength(g),
degree = degree(g),
ID = names(degree(g)),
iteration = iterYearLs$iter[[n]],
yr = iterYearLs$yr[[n]]
)
})
## Metrics for all individuals across all iterations and years
random <- rbindlist(mets)
## Mean values for each individual and year
meanMets <- random[, lapply(.SD, mean), by = .(ID, yr),
.SDcols = c('centrality', 'strength', 'degree')]
Instead of calculating observed and random metrics separately (shown in 7. and 8.), we can calculate metrics for both at the same time and compare.
This chunk expects the outputs from 5. a), skipping steps 6.-8.
Note: by removing the !(observed)
subset from randStep
performed in 6. a), we will include observed rows where iteration == 0
. This will return a gbiLs
where the observed and random rows are included in the same data.table
.
## Create a data.table of unique combinations of iteration and year, including observed and random rows
iterYearLs <- unique(randStep[, .(iteration, yr)])
## Generate group by individual matrix
# for each combination of iteration and year
# 'group' generated by spatsoc::group_pts
# 'randomID' used instead of observed ID (type = 'step')
gbiLs <- mapply(FUN = function(i, y) {
get_gbi(randStep[iteration == i & yr == y],
'group', 'randomID')
},
i = iterYearLs$iter,
y = iterYearLs$yr,
SIMPLIFY = FALSE
)
## Generate a list of random networks
netLs <- lapply(gbiLs, FUN = get_network,
data_format = "GBI", association_index = "SRI")
## Generate graph and calculate network metrics
mets <- lapply(seq_along(netLs), function(n) {
g <- graph.adjacency(netLs[[n]], 'undirected',
diag = FALSE, weighted = TRUE)
data.table(
centrality = evcent(g)$vector,
strength = graph.strength(g),
ID = names(degree(g)),
iteration = iterYearLs$iter[[n]],
yr = iterYearLs$yr[[n]]
)
})
## Observed and random for all individuals across all iterations and years
out <- rbindlist(mets)
## Split observed and random
out[, observed := ifelse(iteration == 0, TRUE, FALSE)]
## Mean values for each individual and year, by observed/random
meanMets <- out[, lapply(.SD, mean), by = .(ID, yr, observed),
.SDcols = c('centrality', 'strength')]