_id
stringlengths 36
36
| text
stringlengths 5
665k
| marker
stringlengths 3
6
| marker_offsets
sequence | label
stringlengths 28
32
|
---|---|---|---|---|
237e1229-3682-475d-898c-7619607f3c7c |
Each gene was divided into 200 bp bins and levels of pol-II activity
were computed at each time point
as the total weighted count of reads overlapping each bin, where each
read was weighted by how many basepairs in the read overlap the bin, as follows. Only
uniquely mapped reads were used. For any read that at least partially
overlaps a bin, the number of basepairs overlapping the bin was added
into the activity level of the bin. For any read spanning multiple
bins, the number of basepairs overlapping each bin were added into activity of
that bin. The Genomatix mapping software provides alignment scores
(values between 0 and 1; with our threshold only between 0.92 and 1)
for mapping reads to the genome; for any read
having alignment score less than 1, the number of overlapping
basepairs added to each bin was multiplied by the alignment score.
A noise removal was then done: a noise level was computed as the
average activity level in 74 manually selected regions from
Chromosome 1 that were visually determined to be inactive over the
measurement time points, as follows.
The regions were divided into 200 bp bins, and
total weighted counts of reads overlapping each bin (each read
weighted by the number of basepairs overlapping the bin) were
computed in the same way as for the genes in the previous step. For
each time point, the noise level was computed as the average
activity level over all bins from all 74 regions. The computed
noise level was subtracted from the mean of each bin in each gene,
thresholding the result at zero. A list of the empty regions
used
is included as a Supplementary Dataset S1.
As the number of ChIP-seq reads collected overall for pol-II varies
between time points, a robust normalisation was done. After the
previous noise removal step, for each gene \(g\) at each time \(t\)
we compute the mean of the remaining activity (activity level after
noise removal) over bins of the gene, denoted as \(r_{gt}\) .
The activity levels are weighted counts of basepairs from reads
overlapping the gene; we select genes having sufficient activity,
that is, at least \(5\cdot 200\) overlapping basepairs from reads over
each 200 bp bin of the gene, on average over the bins. For each gene \(g\) let
\(T_{g} = \lbrace t^{\prime }\in \lbrace 5,10,\ldots ,1280\mathrm {\;min}\rbrace | r_{gt^{\prime }}>5\cdot 200\rbrace \)
denote those time points (except the first time point) where the gene has
sufficient activity.
For each time point we
compute a normalisation factor of [1]}
\(C_t = \textrm {Median}_g \left\lbrace \frac{r_{gt}}{\textrm {GeomMean}_{t^{\prime }} \lbrace r_{gt^{\prime }}\rbrace } \right\rbrace \;.\)
where \(\textrm {Median}_g\lbrace \cdot \rbrace \) denotes median over genes and\(\textrm {GeomMean}_{t^{\prime }}\lbrace r_{gt^{\prime }}\rbrace =(\prod _{t^{\prime }\in T_g} r_{gt^{\prime }})^{1/|T_g|}\) is the
geometric mean over the time points having sufficient activity for
gene \(g\) . The median is computed for time points after the first time
point; for the first time point \(t=0\mathrm {\;min}\)
we set \(C_t=1\) .
The factor \(C_t\) normalises all the gene activity levels (weighted read counts) at a time point downwards if
genes at that time point have unusually many reads, exceeding their
(geometric) mean activity level, and normalises upwards if gene
activity levels fall
under their mean activity level.
Lastly, time series summaries were computed for pol-II at each gene.
For each gene at each time point \(t\) , the mean activity level
(weighted read-count) of pol-II
over bins in the 20% section of the gene nearest to transcription end
was computed, normalised by \(C_t\) . This measured pol-II level
represents transcriptional activity that had successfully passed
through the gene to the transcription end site; it is expected to
correspond better with mRNA production rate than pol-II activity at the
transcription start of the gene, since pol-II near the
transcription start site
can be in the active or
inactive state and after activation may require a significant time for
transcription to complete.
For a small number of genes where the active mRNA transcripts covered only
part of the gene, we considered the area from the first active exon to
last active exon, and summarised the gene using the 20% section
nearest to the end of the area. Active transcripts were defined here
as transcripts with a mean of more than 1.1 assigned counts in the
BitSeq posterior expression estimates. BitSeq uses a prior that
assigns 1 “pseudo-count” per transcript, so the active transcripts
were only required to have minimal posterior expression that was
distinguishable from the prior. A list of active transcripts
is included as Supplementary Dataset S2.
Lastly, for mathematical convenience, for each pol-II time series we
subtracted from all time points the minimum value over the time points.
| [1] | [
[
2516,
2519
]
] | https://openalex.org/W2152239989 |
a6bdb79e-1cab-41de-a84a-f73542f3c8c5 | The linear differential equation (REF ) and
its linear solution operator (REF ) are similar to
those used previously in [1]}, [2]}, [3]}
except for the added delay. As in the previous works, the linearity
of the solution operator permits exact joint Gaussian process (GP)
modelling over \(p(t)\) and \(m(t)\) .
| [1] | [
[
120,
123
]
] | https://openalex.org/W2124584833 |
f73e05ef-d9a3-413d-8b76-5ef4f41cc0a1 | The linear differential equation (REF ) and
its linear solution operator (REF ) are similar to
those used previously in [1]}, [2]}, [3]}
except for the added delay. As in the previous works, the linearity
of the solution operator permits exact joint Gaussian process (GP)
modelling over \(p(t)\) and \(m(t)\) .
| [2] | [
[
126,
129
]
] | https://openalex.org/W2125373697 |
543533bc-59d5-4c4a-958d-3505de184843 | For each gene, we model the pol-II activity time series
in a nonparametric fashion by applying a GP prior over the
shapes of the time series.
Previous similar GP models [1]}, [2]}, [3]}
have used a squared exponential covariance function for \(p(t)\) , as
that allows derivation of all the shared covariances in closed form.
This covariance has the limitation that it is stationary, and
functions following it revert to zero away from data. These
properties severely degrade its performance on our highly unevenly
sampled data. To avoid this, we model \(p(t)\) as an
integral of a function having a GP prior with a squared
exponential covariance: then the posterior
mean of \(p(t)\) tends to remain constant between observed data.
That is, we model
\(p(t)=p_0 + \int _{u=0}^t v(t) \mathrm {d}t\)
| [1] | [
[
169,
172
]
] | https://openalex.org/W2124584833 |
76bb2ed1-5562-413d-81a8-243794401c5f | For each gene, we model the pol-II activity time series
in a nonparametric fashion by applying a GP prior over the
shapes of the time series.
Previous similar GP models [1]}, [2]}, [3]}
have used a squared exponential covariance function for \(p(t)\) , as
that allows derivation of all the shared covariances in closed form.
This covariance has the limitation that it is stationary, and
functions following it revert to zero away from data. These
properties severely degrade its performance on our highly unevenly
sampled data. To avoid this, we model \(p(t)\) as an
integral of a function having a GP prior with a squared
exponential covariance: then the posterior
mean of \(p(t)\) tends to remain constant between observed data.
That is, we model
\(p(t)=p_0 + \int _{u=0}^t v(t) \mathrm {d}t\)
| [2] | [
[
175,
178
]
] | https://openalex.org/W2125373697 |
3b34371a-bd20-4124-8349-f89dd79433ff | Given the data and the priors for the parameters, we apply fully
Bayesian inference with Hamiltonian Monte Carlo (HMC) sampling [1]} to
obtain samples from the posterior distribution of the parameters. HMC
is a MCMC algorithm that uses gradients of
the target distribution to simulate a Hamiltonian dynamical system
with an energy function based on the target distribution. This allows
taking long steps while maintaining a high acceptance rate in the
sampling.
| [1] | [
[
128,
131
]
] | https://openalex.org/W2059448777 |
7a96a816-66c8-4af1-8f08-63bfb90d563a | We run 4 parallel chains starting from different random initial states
to allow convergence checking. We use the HMC implementation from
NETLAB toolkit in Matlab with momentum persistence and number of leap
frog steps \(\tau = 20\) which were found to work well in all cases.
The step length \(\epsilon \) is tuned separately for every model (see
below). After tuning, each chain is run for 10000 iterations. The
samples are then thinned by a factor of 10, and the first half of the
samples are discarded, leaving 500 samples from each chain, 2000 in
all. Convergence is monitored using the potential scale reduction
factor \(\sqrt{\hat{R}}\) [1]}. \(\sqrt{\hat{R}}\) is
computed separately for each variable, and if any of them is greater
than 1.2, the result is discarded and a new sample obtained in a
similar manner. The 9 genes that did not converge after 10 iterations of
this process were removed from further analysis. In most cases these had severely
multimodal delay distributions that were difficult to sample from and
would have made further analysis difficult.
| [1] | [
[
646,
649
]
] | https://openalex.org/W2148534890 |
889554f9-7bfa-4fa1-823f-9479872c3230 | The synthetic data were generated by fitting a GP with the MLP
covariance [1]} to the Pol-II measurements of the gene TIPARP
(ENSG00000163659), and numerically solving the mRNA level using
Eq. (REF ) with the GP posterior mean as \(p(t)\) .
The parameters used were: \(\Delta \in \lbrace 0, 10, 20, 30\rbrace \mathrm {\;min}\) , \(t_{1/2} =\log (2) / \alpha \in \lbrace 2, 4, 8, 16, 32, 64\rbrace \mathrm {\;min}\) , \(\beta _0 = 0.005\) ,
\(\beta = 0.03\) , \(m_0 = 0.008/\alpha \) . The parameter values were
chosen empirically to get profiles that approximately fitted the
actual mRNA observations while looking reasonable and informative
across the entire range of parameter values.
| [1] | [
[
74,
77
]
] | https://openalex.org/W1746819321 |
21ef01bd-afcb-4d01-bbe6-713c2d34c1cc | Introduction.
The statistics of individual speckle patterns created by coherent illumination of an optically random medium is well understood. It is widely accepted that as the real and imaginary parts of the scattered field are uncorrelated Gaussian random variables, the field amplitude follows a Rayleigh distribution and the light intensity follows a negative exponential distribution [1]}. The joint statistics of the scattered light is a more challenging problem. Various types of correlations in space, time or frequency may exist in the scattered fields, depending on the problem geometry and disorder strength [2]}, [3]}. A particularly relevant configuration involves imaging through a scattering slab, where the goal is to reconstruct the image from the transmitted scattered light. The strongest and most evident of the field correlations, the memory effect (ME) [4]}, [3]}, has proven to be useful in such a situation for example allowing to reconstruct the image from a simple autocorrelation calculation of the speckle pattern [6]}, [7]}, [8]}. Recently more subtle long-range mesoscopic correlations emerging in strongly scattering media [9]} have been used to retrieve the image of a hidden object [10]}. However, it is well known that correlations capture only a fraction of the total statistical dependence between random variables.
| [1] | [
[
389,
392
]
] | https://openalex.org/W137864573 |
e80038b9-398f-443f-8291-4ea7f8db99b9 | As we go from surface to the center of a neutron star, at
sufficiently high densities, the matter is expected to undergo a
transition from hadronic matter, where the quarks are confined
inside the hadrons, to a state of deconfined quarks. Finally,
there are up, down and strange quarks in the quark matter. Other
quarks have high masses and do not appear in this state.
Glendenning has shown that a proper construction of the
hadron-quark phase transition inside the neutron stars implies the
coexistence of nucleonic matter and quark matter over a finite
range of the pressure. Therefore, a mixed hadron-quark phase
exists in the neutron star and its energy is lower than those of
the quark matter and nucleonic matter [1]}. These show
that we can consider a neutron star as composed of a hadronic
matter layer, a mixed phase of quarks and hadrons and, in core, a
quark matter.
Recent Chandra observations also imply that the objects RX
J185635-3754 and 3C58 could be neutron stars with the quark core
[2]}.
| [1] | [
[
720,
723
]
] | https://openalex.org/W2073220558 |
b449e642-6b34-4481-b1da-163640581471 | In our calculations, the equation of state of hot nucleonic matter
is determined using the lowest order constrained variational
(LOCV) method as follows
[1]}, [2]}, [3]}, [4]}, [5]}, [6]}, [7]}, [8]}. We adopt a trail
wave function as
\(\psi =F\phi ,\)
| [2] | [
[
159,
162
]
] | https://openalex.org/W2061136215 |
9084b508-c8dd-4b1b-9823-310dfbcd30b7 | In our calculations, the equation of state of hot nucleonic matter
is determined using the lowest order constrained variational
(LOCV) method as follows
[1]}, [2]}, [3]}, [4]}, [5]}, [6]}, [7]}, [8]}. We adopt a trail
wave function as
\(\psi =F\phi ,\)
| [3] | [
[
165,
168
]
] | https://openalex.org/W1991212332 |
82867e4a-659e-4fa3-93ed-db9fe10962ca | In our calculations, the equation of state of hot nucleonic matter
is determined using the lowest order constrained variational
(LOCV) method as follows
[1]}, [2]}, [3]}, [4]}, [5]}, [6]}, [7]}, [8]}. We adopt a trail
wave function as
\(\psi =F\phi ,\)
| [4] | [
[
171,
174
]
] | https://openalex.org/W2071489567 |
4efa1d94-9bfe-46b4-af1b-0feb5ffbfb89 | In our calculations, the equation of state of hot nucleonic matter
is determined using the lowest order constrained variational
(LOCV) method as follows
[1]}, [2]}, [3]}, [4]}, [5]}, [6]}, [7]}, [8]}. We adopt a trail
wave function as
\(\psi =F\phi ,\)
| [6] | [
[
183,
186
]
] | https://openalex.org/W2015172574 |
afbd7c2f-3073-478c-aac4-dfdd2be27dee | In our calculations, the equation of state of hot nucleonic matter
is determined using the lowest order constrained variational
(LOCV) method as follows
[1]}, [2]}, [3]}, [4]}, [5]}, [6]}, [7]}, [8]}. We adopt a trail
wave function as
\(\psi =F\phi ,\)
| [7] | [
[
189,
192
]
] | https://openalex.org/W2044584848 |
1f859c7d-c1c5-481d-8ea8-fe7474b8d25e | In our calculations, the equation of state of hot nucleonic matter
is determined using the lowest order constrained variational
(LOCV) method as follows
[1]}, [2]}, [3]}, [4]}, [5]}, [6]}, [7]}, [8]}. We adopt a trail
wave function as
\(\psi =F\phi ,\)
| [8] | [
[
195,
198
]
] | https://openalex.org/W2047803897 |
8a06c512-3ed4-4dc0-9c36-59d025ddb92b | The procedure to calculate the nucleonic matter has been fully discussed in the Refs.
[1]}, [2]}.
| [2] | [
[
92,
95
]
] | https://openalex.org/W2061136215 |
d6090bd9-0de3-4102-819e-b5f226708452 | We use the MIT bag model for the quark matter calculations. In
this model, the energy density is the kinetic energy of quarks
plus a bag constant (\({\cal B}\) ) which is interpreted as the
difference between the energy densities of non interacting quarks
and interacting ones [1]},
\({\cal E}_{tot} = {\cal E}_u + {\cal E}_d + {\cal E}_s + {\cal B},\)
| [1] | [
[
277,
280
]
] | https://openalex.org/W4242783256 |
c1bd2c03-3c09-4060-bb0b-07de5257d8fa | For the mixed phase, where the fraction of space occupied by quark
matter smoothly increases from zero to unity, we have a mixture of
hadrons, quarks and electrons. In the mixed phase, according the
Gibss equilibrium condition, the temperatures, pressures and
chemical potentials of the hadron phase (H) and quark phase (Q)
are equal [1]}. Here, for each temperature we let the
pressure to be an independent variable.
| [1] | [
[
334,
337
]
] | https://openalex.org/W2073220558 |
e52e8ebe-4b40-4d0e-b76f-3b6775bfb31a | To obtain \(\mu ^H_p\) and \(\mu ^H_n\) for the hadronic matter in
mixed phase, we use the semiempirical mass formula
[1]}, [2]}, [3]},
\(E=T(n,x)+V_0(n)+(1-2x)^2V_2(n),\)
| [3] | [
[
132,
135
]
] | https://openalex.org/W2057910081 |
fbed2da0-7d34-479e-8a0b-6eb673979bd6 | The structure of neutron star is determined by numerically
integrating the Tolman-Oppenheimer-Volkoff equation (TOV)
[1]}, [2]}, [3]}, [4]},
\(\frac{dP}{dr}=-\frac{G[{\cal E}(r)+\frac{P(r)}{c^2}][m(r)+\frac{4\pi r^3 P(r)}{c^2}]}{r^2[1-\frac{2Gm(r)}{rc^2}]},\)
\(\frac{dm}{dr}=4\pi r^2{\cal E}(r),\)
| [1] | [
[
117,
120
]
] | https://openalex.org/W2498126760 |
d3d1c884-baeb-4654-a279-469c1ade3f35 | The structure of neutron star is determined by numerically
integrating the Tolman-Oppenheimer-Volkoff equation (TOV)
[1]}, [2]}, [3]}, [4]},
\(\frac{dP}{dr}=-\frac{G[{\cal E}(r)+\frac{P(r)}{c^2}][m(r)+\frac{4\pi r^3 P(r)}{c^2}]}{r^2[1-\frac{2Gm(r)}{rc^2}]},\)
\(\frac{dm}{dr}=4\pi r^2{\cal E}(r),\)
| [2] | [
[
123,
126
]
] | https://openalex.org/W1649945517 |
592cb1e6-e0c7-4874-98e9-16718fbc388c | In our calculations for the structure of hot neutron star with the
quark core, we use the following equations of state: (i) Below the
density of \(0.05\ fm^3\) , we use the equation of state calculated
by Baym [1]}. (ii) From the density of \(0.05\ fm^3\) up to
the density where the mixed phase starts, we use the equation of
state of pure hadron phase calculated in section REF . (iii)
In the range of densities in which there is the mixed phase, we
use the equation of state calculated in section REF . (iv)
Beyond the density of end point of the mixed phase, we use the
equation of state of pure quark phase calculated in section
REF .
All calculations are done for \({\cal B}=90\ MeV fm^{-3}\) at two
different temperatures \(T=10\) and \(20\ MeV\) . Our results are as
follows.
| [1] | [
[
210,
213
]
] | https://openalex.org/W4235382658 |
00f7f695-dfd2-4fcb-a629-7a10c89f4ba5 | To address such problems in abstractive summarization, several works have been proposed to improve the factual consistency of summarization. As shown in Table REF , existing works can be roughly categorized into two classes: fact-input methods [1]}, [2]} which aim to encode the information of facts in the article, and post-edit methods [3]}, [4]} which seek to correct the factual errors after decoding. What's more, there are some integrated works which perform both improvements [5]}.
These methods usually need to modify the architecture of the model, adding additional encoding modules or post-edit modules.
<FIGURE> | [1] | [
[
244,
247
]
] | https://openalex.org/W2963676814 |
92c1de2d-8518-4ad6-80b7-05448df12e5e | To address such problems in abstractive summarization, several works have been proposed to improve the factual consistency of summarization. As shown in Table REF , existing works can be roughly categorized into two classes: fact-input methods [1]}, [2]} which aim to encode the information of facts in the article, and post-edit methods [3]}, [4]} which seek to correct the factual errors after decoding. What's more, there are some integrated works which perform both improvements [5]}.
These methods usually need to modify the architecture of the model, adding additional encoding modules or post-edit modules.
<FIGURE> | [3] | [
[
338,
341
]
] | https://openalex.org/W3102645206 |
d68fd688-4e94-4095-865e-c1a3cc9abc03 | Contrastive learning on encoder (CoEnc) calculates the contrastive loss described in [1]}. CoEnc first encodes the article and summaries (ground truth and negative samples), then make the encoded representation of the article and the ground truth summary closer, and make that of the article and the factual inconsistency summary apart. As shown in Figure REF , the motivation of encoding both article and summary on the encoder is to catch and encode fact information. Given the article, the encoder can only distinguish the ground truth summary from the very similar negative summary by catching the common correct fact in the article-summary pair. It can be also explained from the view of data augmentation. Similar to the crops and rotations of images in SimCLR [2]}, we can regard the article and summary as two kinds of “data augmentation” on the fact. CO2Sum is designed to catch the fact behind the augmentation.
| [2] | [
[
767,
770
]
] | https://openalex.org/W3005680577 |
901c71ef-49de-4203-a5c0-715c670599ed |
QAGS [1]}: QAGS generates questions about named entities and noun phrases in the predicted summary using a trained QG (Question Generation) model, then uses a QA (Question Answering) model to find answers to questions from the corresponding article. QAGS calculates token-level F1 similarity between QA results and asked entities or noun phrase in the summary as the final score.
QuestEval [2]}: Compared to QAGS, QuestEval considers the situation of unanswerable questions. What's more, QuestEval does not calculate answer similarity, but scores the precision and recall apart, then QuestEval gives a weighted F1 score.
Close Scheme Fact Triple [3]}: Fact Triple based metrics score the precision between summary extracted triple and article extracted triple. The triples \((Subject, Relation, Object)\) are extracted using Named Entity Recognition (NER) and Relation Extraction (RE) models. These triples are structured data of factual information and can be used to evaluate factual consistency.
Open Scheme Fact Triple [3]}: Open Scheme Fact Triple is similar to the close ones, but the relationship in the fact triple is text span instead of classified relation label.
| [3] | [
[
649,
652
],
[
1028,
1031
]
] | https://openalex.org/W2947681066 |
b37ab257-61fd-41c7-94e1-c0c181a92f44 | We use factsummhttps://github.com/Huffon/factsumm [1]}, OpenIEhttps://github.com/philipperemy/stanford-openie-python [2]} and official provided codehttps://github.com/ThomasScialom/QuestEval [3]} to build evaluation system. On the use of trained models, we choose FLAIR [4]} for NER, LUKE [5]} for RE, T5 [6]} and Roberta [7]} for QA and QG. We calculate all Fact Triple metrics only on oracle sentences in article and summaries, since there is no need to calculate triple precision on those redundant sentences in the article.
| [2] | [
[
117,
120
]
] | https://openalex.org/W2251913848 |
e04672b2-2bf0-4284-b0f4-1134d2a2f3e8 |
Random: randomly pick and replace words in the ground truth summary to construct negative samples.
NP: identify noun phrases in the summary and replace the words in the phrase.
NER: perform Named Entity Recognition on the summary and construct negative samples with entity-level replacements.
LFN: the proposed language model based construction method in CO2Sum.
LFN (DN): similar to the improvement of Roberta [1]} over BERT [2]}, we perform dynamic (DN) negative sample construction during training.
| [2] | [
[
431,
434
]
] | https://openalex.org/W2896457183 |
4fff9dae-4fcf-45ad-9b5d-5875b98a6b50 | In this section, we study the loss function in the CoDec. We compare the results of PM max-margin loss with the original loss (Vanilla) described in [1]}. Besides, we attempt another gated-weighting method (Gated) that dynamically calculates the weight of different positions. It uses a Linear Gate Unit [2]} to calculate the weights based on the hidden state of the decoder. The results are shown in Table REF . The gated method does not perform better than the vanilla, but position masked loss outperforms vanilla on all metrics. We assume that it is too difficult for a model to learn the different weights of positions. A simple mask can stabilize the training and performs better.
<FIGURE><FIGURE> | [2] | [
[
304,
307
]
] | https://openalex.org/W2964265128 |
aa6fa84c-696b-4729-8f7e-4099f59bc303 | Most existing methods for improving fact consistency can be divided into fact-input-based methods and post-edit-based methods.
Fact-input-based methods focus on enhancing the representation of facts in the source article or incorporating commonsense knowledge, which is useful to facilitate summarization systems understanding the facts for reducing consistent error.
[1]} introduce FTSum to reduce consistent error by applying the encoder to incorporate the fact description.
[2]} aim to incorporate entailment knowledge into the summarization model.
Post-edit based method aims to apply a post-edit on the model-generated summaries for obtaining more factual-consistent summarization.
[3]} propose a fact corrector, which corrects the factual error in the model-generated summary in an iterative and auto-regressive manner.
[4]} propose a neural-based corrector module to address the factual inconsistent issue by identifying and correcting factual errors in generated summaries.
[5]} explore to model the facts in the source article with knowledge graphs based on a neural network.
[6]} study contrast candidate generation and selection to correct the extrinsic fact hallucinations in a post-edit manner.
Comparing with the above works, we aim to improve factual consistency through contrastive learning without introducing extra parameters.
| [1] | [
[
368,
371
]
] | https://openalex.org/W2963676814 |
59cd90bd-1f37-4d9a-a694-382219db96c5 | Most existing methods for improving fact consistency can be divided into fact-input-based methods and post-edit-based methods.
Fact-input-based methods focus on enhancing the representation of facts in the source article or incorporating commonsense knowledge, which is useful to facilitate summarization systems understanding the facts for reducing consistent error.
[1]} introduce FTSum to reduce consistent error by applying the encoder to incorporate the fact description.
[2]} aim to incorporate entailment knowledge into the summarization model.
Post-edit based method aims to apply a post-edit on the model-generated summaries for obtaining more factual-consistent summarization.
[3]} propose a fact corrector, which corrects the factual error in the model-generated summary in an iterative and auto-regressive manner.
[4]} propose a neural-based corrector module to address the factual inconsistent issue by identifying and correcting factual errors in generated summaries.
[5]} explore to model the facts in the source article with knowledge graphs based on a neural network.
[6]} study contrast candidate generation and selection to correct the extrinsic fact hallucinations in a post-edit manner.
Comparing with the above works, we aim to improve factual consistency through contrastive learning without introducing extra parameters.
| [3] | [
[
687,
690
]
] | https://openalex.org/W3102489149 |
f114c42f-8206-4596-81f4-044a594d00a3 | Most existing methods for improving fact consistency can be divided into fact-input-based methods and post-edit-based methods.
Fact-input-based methods focus on enhancing the representation of facts in the source article or incorporating commonsense knowledge, which is useful to facilitate summarization systems understanding the facts for reducing consistent error.
[1]} introduce FTSum to reduce consistent error by applying the encoder to incorporate the fact description.
[2]} aim to incorporate entailment knowledge into the summarization model.
Post-edit based method aims to apply a post-edit on the model-generated summaries for obtaining more factual-consistent summarization.
[3]} propose a fact corrector, which corrects the factual error in the model-generated summary in an iterative and auto-regressive manner.
[4]} propose a neural-based corrector module to address the factual inconsistent issue by identifying and correcting factual errors in generated summaries.
[5]} explore to model the facts in the source article with knowledge graphs based on a neural network.
[6]} study contrast candidate generation and selection to correct the extrinsic fact hallucinations in a post-edit manner.
Comparing with the above works, we aim to improve factual consistency through contrastive learning without introducing extra parameters.
| [4] | [
[
826,
829
]
] | https://openalex.org/W3102645206 |
3459e464-0d64-4e71-b268-2ee671be3232 | The orthogonal and unitary calculi allow for the systematic study of functors from either the category of real inner product spaces, or the category of complex inner product spaces, to the category of based topological spaces. The motivating examples are \(\mathrm {BO}(-) \colon V \longmapsto \mathrm {BO}(V)\) , where \(\mathrm {BO}(V)\) is the classifying space of the orthogonal group of \(V\) , and \(\mathrm {BU}(-)\colon W \longmapsto \mathrm {BU}(W)\) where \(\mathrm {BU}(W)\) is the classifying space of the unitary group of \(W\) . The foundations of orthogonal calculus were originally developed by Weiss in [1]}, and later converted to a model category theoretic framework by Barnes and Oman in [2]}. From the unitary calculus perspective, it has long been known to the experts, with the foundations and model category framework developed by the author in [3]}.
| [1] | [
[
623,
626
]
] | https://openalex.org/W4239559895 |
7bf5ee4a-2b35-43c9-b8c7-bce6275a749c | The orthogonal and unitary calculi allow for the systematic study of functors from either the category of real inner product spaces, or the category of complex inner product spaces, to the category of based topological spaces. The motivating examples are \(\mathrm {BO}(-) \colon V \longmapsto \mathrm {BO}(V)\) , where \(\mathrm {BO}(V)\) is the classifying space of the orthogonal group of \(V\) , and \(\mathrm {BU}(-)\colon W \longmapsto \mathrm {BU}(W)\) where \(\mathrm {BU}(W)\) is the classifying space of the unitary group of \(W\) . The foundations of orthogonal calculus were originally developed by Weiss in [1]}, and later converted to a model category theoretic framework by Barnes and Oman in [2]}. From the unitary calculus perspective, it has long been known to the experts, with the foundations and model category framework developed by the author in [3]}.
| [2] | [
[
711,
714
]
] | https://openalex.org/W2034710066 |
d84fb767-a4bb-4c4d-975b-1e55556ab617 | We start with the categories of spectra in Section , and use the change of group functors of Mandell and May [1]}, to construct Quillen adjunctions between spectra with an action of \(\mathrm {O}(n)\) , \(\mathrm {U}(n)\) and \(\mathrm {O}(2n)\) respectively. We utilise the Quillen equivalence between orthogonal and unitary spectra of [2]} to show that these change of group functors interact in a homotopically meaningful way with the change of model functor induced by realification.
| [1] | [
[
109,
112
]
] | https://openalex.org/W2075488415 |
84189aac-5c29-4c97-9c75-bec1acac3733 | In Section we move to comparing the intermediate categories. These are categories \(\mathrm {O}(n)\mathcal {E}_n^\mathbf {O}\) and \(\mathrm {U}(n)\mathcal {E}_n^\mathbf {U}\) constructed by Barnes and Oman [1]}, and the author [2]}, which act as an intermediate in the zig-zag of Quillen equivalences for orthogonal and unitary calculus respectively. For this, we introduce two new intermediate categories, \(\mathrm {O}(n)\mathcal {E}_n^\mathbf {U}\) and \(\mathrm {U}(n)\mathcal {E}_{2n}^\mathbf {O}\) between the standard intermediate categories. These are the standard intermediate categories with restricted group actions through the subgroup inclusions \(\mathrm {O}(n) \hookrightarrow \mathrm {U}(n)\) and \(\mathrm {U}(n) \hookrightarrow \mathrm {O}(2n)\) . We exhibit Quillen equivalences between these intermediate categories and the standard intermediate categories, completing the picture using change of group functors from [3]}. The resulting diagram of intermediate categories is as follows,
\(@C=4em{\mathrm {O}(n)\mathcal {E}_n^\mathbf {O} @<-1ex>[r]_{r^*}^\sim & \mathrm {O}(n)\mathcal {E}_n^\mathbf {U} @<1ex>[r]^{\mathrm {U}(n)_+\wedge _{\mathrm {O}(n)}(-)}@<-1ex>[l]_{r_!} & \mathrm {U}(n)\mathcal {E}_n^\mathbf {U} @<1ex>[l]^{\iota ^*}@<-1ex>[r]_{c^*} & \mathrm {U}(n)\mathcal {E}_{2n}^\mathbf {O} @<1ex>[r]^{\mathrm {O}(2n)_+\wedge _{\mathrm {U}(n)}(-)} @<-1ex>[l]_{c_!}^\sim & \mathrm {O}(2n)\mathcal {E}_{2n}^\mathbf {O}. @<1ex>[l]^{\kappa ^*}}\)
| [1] | [
[
210,
213
]
] | https://openalex.org/W2034710066 |
c3295b9d-b52d-479c-a9bb-09debbaf0751 | In Section we move to comparing the intermediate categories. These are categories \(\mathrm {O}(n)\mathcal {E}_n^\mathbf {O}\) and \(\mathrm {U}(n)\mathcal {E}_n^\mathbf {U}\) constructed by Barnes and Oman [1]}, and the author [2]}, which act as an intermediate in the zig-zag of Quillen equivalences for orthogonal and unitary calculus respectively. For this, we introduce two new intermediate categories, \(\mathrm {O}(n)\mathcal {E}_n^\mathbf {U}\) and \(\mathrm {U}(n)\mathcal {E}_{2n}^\mathbf {O}\) between the standard intermediate categories. These are the standard intermediate categories with restricted group actions through the subgroup inclusions \(\mathrm {O}(n) \hookrightarrow \mathrm {U}(n)\) and \(\mathrm {U}(n) \hookrightarrow \mathrm {O}(2n)\) . We exhibit Quillen equivalences between these intermediate categories and the standard intermediate categories, completing the picture using change of group functors from [3]}. The resulting diagram of intermediate categories is as follows,
\(@C=4em{\mathrm {O}(n)\mathcal {E}_n^\mathbf {O} @<-1ex>[r]_{r^*}^\sim & \mathrm {O}(n)\mathcal {E}_n^\mathbf {U} @<1ex>[r]^{\mathrm {U}(n)_+\wedge _{\mathrm {O}(n)}(-)}@<-1ex>[l]_{r_!} & \mathrm {U}(n)\mathcal {E}_n^\mathbf {U} @<1ex>[l]^{\iota ^*}@<-1ex>[r]_{c^*} & \mathrm {U}(n)\mathcal {E}_{2n}^\mathbf {O} @<1ex>[r]^{\mathrm {O}(2n)_+\wedge _{\mathrm {U}(n)}(-)} @<-1ex>[l]_{c_!}^\sim & \mathrm {O}(2n)\mathcal {E}_{2n}^\mathbf {O}. @<1ex>[l]^{\kappa ^*}}\)
| [3] | [
[
944,
947
]
] | https://openalex.org/W2075488415 |
d055ae69-d195-47fe-93d4-e1e9b8511910 | The category \(\mathcal {E}_0^\mathbf {O}\) is category of input functors for orthogonal calculus as studied by Weiss and Barnes and Oman [1]}, [2]}. Moreover \(\mathcal {E}_0^\mathbf {U}\) is the category of input functors for unitary calculus, studied by the author in [3]}. These input categories are categories of diagram spaces as in [4]} hence they can be equipped with a projective model structure.
| [1] | [
[
139,
142
]
] | https://openalex.org/W4239559895 |
fc1ecf49-4b1e-4e84-a134-0b1334658e79 | The category \(\mathcal {E}_0^\mathbf {O}\) is category of input functors for orthogonal calculus as studied by Weiss and Barnes and Oman [1]}, [2]}. Moreover \(\mathcal {E}_0^\mathbf {U}\) is the category of input functors for unitary calculus, studied by the author in [3]}. These input categories are categories of diagram spaces as in [4]} hence they can be equipped with a projective model structure.
| [2] | [
[
145,
148
]
] | https://openalex.org/W2034710066 |
8c9c4505-b813-4cce-8fee-57cb0b866676 | The category \(\mathcal {E}_0^\mathbf {O}\) is category of input functors for orthogonal calculus as studied by Weiss and Barnes and Oman [1]}, [2]}. Moreover \(\mathcal {E}_0^\mathbf {U}\) is the category of input functors for unitary calculus, studied by the author in [3]}. These input categories are categories of diagram spaces as in [4]} hence they can be equipped with a projective model structure.
| [4] | [
[
341,
344
]
] | https://openalex.org/W2086997195 |
40147305-f885-4934-908d-5d29019b0e32 | Arguably the most important class of functors in orthogonal and unitary calculi are the \(n\) -polynomial functors, and in particular the \(n\) -th polynomial approximation functor. Here we give a short overview of these functors, for full details on these functors see [1]}, [2]}, [3]}.
| [1] | [
[
270,
273
]
] | https://openalex.org/W4239559895 |
f12275ec-379c-41b0-ac5b-b0da9c85b7cc | Arguably the most important class of functors in orthogonal and unitary calculi are the \(n\) -polynomial functors, and in particular the \(n\) -th polynomial approximation functor. Here we give a short overview of these functors, for full details on these functors see [1]}, [2]}, [3]}.
| [2] | [
[
276,
279
]
] | https://openalex.org/W2034710066 |
fb817270-e1c1-4c09-add2-fb9d1d65073b | Since an \(n\) -polynomial functor is \((n+1)\) -polynomial, see [1]}, these polynomial approximation functors assemble into a Taylor tower approximating a given input functor. Moreover there is a model structure on \(\mathcal {E}_0\) which captures the homotopy theory of \(n\) -polynomial functors.
| [1] | [
[
65,
68
]
] | https://openalex.org/W4239559895 |
cdbca8f4-1140-4d41-a723-9ae62128df76 | Proposition 2.5 ([1]}, [2]})
There is a cellular proper topological model structure on \(\mathcal {E}_0\) where a map \(f\colon E \rightarrow F\) is a weak equivalence if \(T_nf\colon T_nE \rightarrow T_nF\) is a levelwise weak equivalence, the cofibrations are the cofibrations of the projective model structure and the fibrations are levelwise fibrations such that
\({E [r]^f [d]_{\eta _E} & F [d]^{\eta _F} \\T_nE [r]_{T_nf} & T_n F}\)
| [1] | [
[
17,
20
]
] | https://openalex.org/W2034710066 |
8b53fb8c-ee82-4b7c-9099-462995e8f439 | Proposition 2.7 ([1]}, [2]})
There is a topological model structure on \(\mathcal {E}_0\) where the weak equivalences are those maps \(f\) such that \(D_nf\) is a weak equivalence in \(\mathcal {E}_0\) , the fibrations are the fibrations of the \(n\) -polynomial model structure and the cofibrations are those maps with the left lifting property with respect to the acyclic fibrations. The fibrant objects are \(n\) -polynomial and the cofibrant-fibrant objects are the projectively cofibrant \(n\) -homogeneous functors.
| [1] | [
[
17,
20
]
] | https://openalex.org/W2034710066 |
519d46f4-b116-4f80-b76b-a3ae1c3b1904 | In [1]}, Weiss constructs a zig-zag of equivalences between the category of \(n\) -homogeneous functors (up to homotopy) and the homotopy category of spectra with an action of \(\mathrm {O}(n)\) . In [2]}, Barnes and Oman put this zig-zag into a model category theoretic framework via a zig-zag of Quillen equivalences between the \(n\) -homogeneous model structure on \(\mathcal {E}_0^\mathbf {O}\) , and spectra with an action of \(\mathrm {O}(n)\) . This zig-zag moves through an intermediate category, denote \(\mathrm {O}(n)\mathcal {E}_n^\mathbf {O}\) . In [3]}, the author constructs a similar zig-zag of Quillen equivalences between the unitary \(n\) -homogeneous model structure and spectra with an action of \(\mathrm {U}(n)\) . We give an overview of the construction of these intermediate categories and how they relate to spectra and the \(n\) -homogenous model structure.
| [1] | [
[
3,
6
]
] | https://openalex.org/W4239559895 |
d36d32a8-5b37-417d-b405-de0ddf3fd78c | In [1]}, Weiss constructs a zig-zag of equivalences between the category of \(n\) -homogeneous functors (up to homotopy) and the homotopy category of spectra with an action of \(\mathrm {O}(n)\) . In [2]}, Barnes and Oman put this zig-zag into a model category theoretic framework via a zig-zag of Quillen equivalences between the \(n\) -homogeneous model structure on \(\mathcal {E}_0^\mathbf {O}\) , and spectra with an action of \(\mathrm {O}(n)\) . This zig-zag moves through an intermediate category, denote \(\mathrm {O}(n)\mathcal {E}_n^\mathbf {O}\) . In [3]}, the author constructs a similar zig-zag of Quillen equivalences between the unitary \(n\) -homogeneous model structure and spectra with an action of \(\mathrm {U}(n)\) . We give an overview of the construction of these intermediate categories and how they relate to spectra and the \(n\) -homogenous model structure.
| [2] | [
[
200,
203
]
] | https://openalex.org/W2034710066 |
18440549-2077-4c01-af6d-1eb0976a5c8a | Let \(n\mathbb {S}\) be the functor given by \(V \longmapsto S^{nV}\) where \(nV := \mathbb {F}^n \otimes _\mathbb {F}V\) . By [1]} and [2]} the intermediate categories are equivalent to a category of \(n\mathbb {S}\) -modules and hence come equipped with an \(n\) -stable model structure similar to the stable model structure on spectra. The weak equivalences of the \(n\) -stable model structure are given by \(n\pi _*\) -isomorphisms. Theses are defined via the structure maps of objects in \(\mathrm {Aut}(n)\mathcal {E}_n\) , and as such have slightly different forms depending on whether one is in the orthogonal or unitary setting.
| [1] | [
[
129,
132
]
] | https://openalex.org/W2034710066 |
76319983-df13-43b2-8825-7593fa9945ba | Proposition 2.14 ([1]}, [2]})
There is a cofibrantly generated, proper, topological model structure on the category \(\mathrm {Aut}(n)\mathcal {E}_n\) , where the weak equivalences are the \(n\pi _*\) -isomorphisms, the cofibrations are those maps with the left lifting property with respect to all levelwise acyclic fibrations and the fibrations are those levelwise fibrations \(f\colon X \rightarrow Y\) such that the diagram
\({X(V) [r] [d] & \Omega ^{nW}X(V \oplus W) [d] \\Y(V) [r] & \Omega ^{nW}Y(V \oplus W).}\)
| [1] | [
[
18,
21
]
] | https://openalex.org/W2034710066 |
0e8071a3-282f-4fc6-803d-90229e47e5df | The orthogonal case is similar, full details may be found in [1]}.
| [1] | [
[
61,
64
]
] | https://openalex.org/W2034710066 |
1defc0aa-19c7-4b56-b6a6-fcc317b64634 | Proposition 2.16 ([1]})
There is a Quillen equivalence
\({(\beta _n)_!:\mathrm {O}(n)\mathcal {E}_n^\mathbf {O} @<0.7ex>[r] &@<0.7ex>[l] \mathsf {Sp}^\mathbf {O}[\mathrm {O}(n)]:(\beta _n)^*}\)
| [1] | [
[
18,
21
]
] | https://openalex.org/W2034710066 |
37583206-78d3-4cf4-b168-7414de42c1be | Combining this adjunction with a change of group action from [1]} provides an adjunction
\({\operatorname{\mathrm {res}}_0^n/\mathrm {Aut}(n):\mathrm {Aut}(n)\mathcal {E}_n @<0.7ex>[r] &@<0.7ex>[l] \mathcal {E}_0:\operatorname{\mathrm {ind}}_0^n \varepsilon ^*}.\)
| [1] | [
[
61,
64
]
] | https://openalex.org/W2075488415 |
27c19ee1-26ce-4a58-8dbc-6f8427360208 | Proposition 2.18 ([1]}, [2]})
The adjoint pair
\({\operatorname{\mathrm {res}}_0^n/\mathrm {Aut}(n):\mathrm {Aut}(n)\mathcal {E}_n @<0.7ex>[r] &@<0.7ex>[l] n\operatorname{--homog--}\mathcal {E}_0:\operatorname{\mathrm {ind}}_0^n\varepsilon ^*}\)
| [1] | [
[
18,
21
]
] | https://openalex.org/W2034710066 |
eff0da6b-b918-45d9-bbd2-dbead0678a91 | Proposition 2.19 ([1]},[2]})
Let \(F \in \mathcal {E}_0\) be \(n\) -homogeneous for some \(n >0\) . Then \(F\) is levelwise weakly equivalent to the functor defined as
\(U \longmapsto \Omega ^\infty [(S^{nU} \wedge \Psi _F^n)_{h\mathrm {Aut}(n)}].\)
| [1] | [
[
18,
21
]
] | https://openalex.org/W4239559895 |
5a7b973c-95d3-4736-a1fc-063bccd2808e | When two functors agree to a given order, their Taylor tower agree to a prescribed level. The first result in that direction is the unitary analogue of [1]}.
| [1] | [
[
152,
155
]
] | https://openalex.org/W4239559895 |
5c3089fd-1790-4c03-af63-b68a736fa192 | Lemma 2.21 ([1]},[2]})
let \(p \colon G \rightarrow F\) be a map in \(\mathcal {E}_0\) . Suppose that there is \(b \in \mathbb {Z}\) such that \(p_U\colon G(U) \rightarrow F(U)\) is \(((n+1)\dim _\mathbb {R}(U) - b)\) -connected for all \(U \in J_0\) with \(\dim _\mathbb {F}(U) \ge \rho \) . Then
\(\tau _n(p)_U \colon \tau _n (G(U)) \rightarrow \tau _n(F(U))\)
| [1] | [
[
12,
15
]
] | https://openalex.org/W4239559895 |
725a69a7-6c3c-497f-b696-827d786e5479 | Lemma 2.22 ([1]},[2]})
If \(p \colon F \rightarrow G\) is an order \(n\) agreement, then \(T_k F \rightarrow T_k G\) is a levelwise weak equivalence for \(k \le n\) .
| [1] | [
[
12,
15
]
] | https://openalex.org/W4239559895 |
50009f15-6c5d-496e-b338-3e37f3389027 | Let \(F\) be an \(n\) -homogeneous orthogonal functor. Then by the characterisation, Proposition REF , \(F\) is levelwise weakly equivalent to the functor
\(V \longmapsto \Omega ^\infty [(S^{\mathbb {R}^n \otimes _\mathbb {R}V} \wedge \Psi _F^n)_{h\mathrm {O}(n)}]\)
where \(\Psi _F^n\) is an orthogonal spectrum with an \(O(n)\) -action. It follows that pre-realification of \(F\) is levelwise weakly equivalent to the functor
\(W \longmapsto \Omega ^\infty [(S^{\mathbb {R}^n \otimes _\mathbb {R}W_\mathbb {R}} \wedge \Psi ^n_F)_{h\mathrm {O}(n)}].\)
Using the derived change of group functor, we construct an orthogonal spectrum with an action of \(\mathrm {U}(n)\) ,
\(\mathrm {U}(n)_+ \wedge _{\mathrm {O}(n)}^{L} \Psi _F^n := \mathrm {U}(n)_+ \wedge _{\mathrm {O}(n)} (E\mathrm {O}(n)_+ \wedge \Psi _F^n).\)
By the classification of \(n\) -homogeneous unitary functors, Proposition REF , there is an \(n\) -homogeneous functor \(F^{\prime }\) associated to the above spectrum, given by
\(W \longmapsto \Omega ^\infty [(S^{n \otimes _W̏} \wedge (\mathrm {U}(n)_+ \wedge _{\mathrm {O}(n)} (E\mathrm {O}(n)_+ \wedge \Psi _F^n)))_{h\mathrm {U}(n)}].\)
By [1]}, \(F^{\prime }(W)\) is isomorphic to
\(\Omega ^\infty [\mathrm {U}(n)_+ \wedge _{\mathrm {O}(n)} ((\iota ^*S^{n \otimes _W̏} \wedge (E\mathrm {O}(n)_+ \wedge \Psi _F^n)))_{h(\mathrm {U}(n)}].\)
The \(\mathrm {U}(n)\) -action on \(\mathrm {U}(n)_+ \wedge _{\mathrm {O}(n)} ((\iota ^*S^{n \otimes _W̏} \wedge (E\mathrm {O}(n)_+ \wedge \Psi _F^n)))\) is free (\((E\mathrm {O}(n)_+\) is a free \(\mathrm {O}(n)\) -space), hence taking homotopy orbits equates to taking strict orbits. Hence there is an isomorphism
\(F^{\prime }(W) \cong \Omega ^\infty [(\mathrm {U}(n)_+ \wedge _{\mathrm {O}(n)} ((\iota ^*S^{n \otimes _W̏} \wedge (E\mathrm {O}(n)_+ \wedge \Psi _F^n))))/\mathrm {U}(n)].\)
The strict \(\mathrm {U}(n)\) -orbits of the spectrum \(\mathrm {U}(n)_+ \wedge _{\mathrm {O}(n)} ((\iota ^*S^{n \otimes _W̏} \wedge (E\mathrm {O}(n)_+ \wedge \Psi _F^n))\) are isomorphic to the \(\mathrm {O}(n)\) -orbits of the spectrum, \(\iota ^*S^{n \otimes _W̏} \wedge (E\mathrm {O}(n)_+ \wedge \Psi _F^n))\) , hence \(F^{\prime }(W)\) is isomorphic to
\(\Omega ^\infty [(\iota ^*S^{n \otimes _W̏} \wedge (E\mathrm {O}(n)_+ \wedge \Psi _F^n)))/\mathrm {O}(n)].\)
This last is precisely
\(\Omega ^\infty [(\iota ^*S^{n \otimes _W̏} \wedge \Psi _F^n)_{h\mathrm {O}(n)}]\)
as homotopy orbits is the left derived functor of strict orbits and smashing with \(E\mathrm {O}(n)_+\) is a cofibrant replacement in the projective model structure.
Since the action of \(\mathrm {O}(n)\) on \(\iota ^*S^{n \otimes _W̏}\) is equivalent to the \(\mathrm {O}(n)\) action on \(S^{\mathbb {R}^n \otimes _\mathbb {R}W_\mathbb {R}}\) and the one-point compactification are isomorphic, the above infinite loop space is isomorphic to
\(\Omega ^\infty [(S^{\mathbb {R}^n \otimes _\mathbb {R}W_\mathbb {R}} \wedge \Psi _F^n)_{h\mathrm {O}(n)}].\)
By the characterisation of \(n\) -homogeneous orthogonal functors, we see that this is levelwise weakly equivalent to \(F(W_\mathbb {R}) = (r^*F)(W).\)
| [1] | [
[
1166,
1169
]
] | https://openalex.org/W2075488415 |
8b54c81b-7689-4d80-a5fa-142307906e26 | We argue by induction on the polynomial degree. The case \(n=0\) follows by definition. Assume the map \(r^*T_{n-1}^\mathbf {O} F \rightarrow T_{n-1}^\mathbf {U}(r^*T_{n-1}^\mathbf {O} F)\) is a levelwise weak equivalence. There is a homotopy fibre sequence
\(T_n^\mathbf {O} F \longrightarrow T_{n-1}^\mathbf {O} F \longrightarrow R_n^\mathbf {O} F\)
where \(R_n^\mathbf {O} F\) is \(n\) -homogeneous, since \(F\) satisfies the conditions of [1]}. Lemma REF implies that \(r^* R_n^\mathbf {O} F\) is \(n\) -homogeneous in \(\mathcal {E}_0^\mathbf {U}\) , and in particular \(n\) -polynomial. As homotopy fibres of maps between \(n\) -polynomial objects are \(n\) -polynomial, the homotopy fibre of the map \(r^*T_{n-1}^\mathbf {O} F \rightarrow r^* R_n^\mathbf {O} F\) is \(n\) -polynomial. Computation of homotopy fibres is levelwise, hence the homotopy fibre in question is \(r^* T_n^\mathbf {O} F\) , and it follows that
\(r^* T_n^\mathbf {O} F \longrightarrow T_n^\mathbf {U}(r^* T_n^\mathbf {O} F)\)
is a levelwise weak equivalence.
| [1] | [
[
448,
451
]
] | https://openalex.org/W4239559895 |
e4c19355-c6f1-4ebf-a76b-accaa5c9cb87 | The acyclic fibrations in the \(n\) –polynomial model structure are levelwise, hence both terms in the composite, and resultingly their composite, preserves these. It suffices by [1]} to show that the composite preserves fibrations between fibrant objects, which are the levelwise fibrations by [2]}. It hence suffices to show that \(F\) preserves fibrant objects. Let \(F\) be a \(n\) –polynomial orthogonal functor, then \(\mathrm {red}(F)\) is also \(n\) –polynomial, and reduced. An application of Theorem REF implies the result.
| [2] | [
[
295,
298
]
] | https://openalex.org/W1583122470 |
2825ac31-7e00-48d1-97d5-5976dee8b1e0 | The notion of agreement plays a central role in the theory or orthogonal and unitary calculus, for example it is crucial to the proof that the \(n\) -th polynomial approximation in \(n\) -polynomial, see [1]}. The pre-realification and pre-complexification functors behave well with respect to functors which agree to a certain order.
| [1] | [
[
204,
207
]
] | https://openalex.org/W4239559895 |
6158a9f6-e00d-41e9-bd26-082f31840430 | We have constructed a Quillen adjunction between the orthogonal and unitary \(n\) -homogeneous model structures. To give a complete comparison of the theories we must address the comparisons between the other two categories in the zig-zag of Quillen equivalences of Barnes and Oman [1]} and the author [2]}. We start by addressing the relationship between the categories of spectra. For this, we recall the definitions and model structures involved.
| [1] | [
[
282,
285
]
] | https://openalex.org/W2034710066 |
0a05370c-d6e6-45a6-a674-ddf8906dda09 | compare [1]}.
| [1] | [
[
8,
11
]
] | https://openalex.org/W2075488415 |
e1946ea6-6ab7-40a2-8aef-b31c07426919 | These categories also come with projective and stable model structures constructed analogously to those of Proposition REF . These new intermediate categories will now act as intermediate categories between the standard intermediate categories of orthogonal and unitary calculus. Further, the new intermediate categories equipped with their \(n\) -stable model structure are Quillen equivalent to spectra with an appropriate group action. The proofs of the following two results follow similarly to [1]} and [2]}.
| [1] | [
[
499,
502
]
] | https://openalex.org/W2034710066 |
6a555566-74ef-41dd-b36b-c0538abb7878 | where the isomorphism follows from [1]}.
| [1] | [
[
35,
38
]
] | https://openalex.org/W2075488415 |
203637de-ef05-49dc-b95c-f4a04a88560c | By the zig-zag of Quillen equivalences, [1]} the composite
\({L}\operatorname{\mathrm {res}}_0^n/\mathrm {O}(n) \circ {R}(\beta _n)^*\)
applied to an orthogonal spectrum \(\Theta \) with an action of \(\mathrm {O}(n)\) , is levelwise weakly equivalent to the functor \(F\) , defined by the formula
\(V \longmapsto \Omega ^\infty [(S^{\mathbb {R}^n \otimes V} \wedge \Theta )_{h\mathrm {O}(n)}].\)
This functor is \(n\) –homogeneous, hence also reduced, so \({R}r^*\mathrm {red}(F)\) is levelwise weakly equivalent to \(\mathbb {R}r^*F\) . The zig-zag of Quillen equivalences from unitary calculus, [2]}, together with inflating \(\Theta \) to a spectrum with an action of \(\mathrm {U}(n)\) gives a similar characterisation in terms of an \(n\) -homogeneous functor. The result then follows by Proposition REF .
| [1] | [
[
40,
43
]
] | https://openalex.org/W2034710066 |
61ccc52d-9145-40e1-bd57-48a46af10baa | Let \(F\) be an \(n\) -homogeneous orthogonal functor. Then by the characterisation, Proposition REF , \(F\) is levelwise weakly equivalent to the functor
\(V \longmapsto \Omega ^\infty [(S^{\mathbb {R}^n \otimes _\mathbb {R}V} \wedge \Psi _F^n)_{h\mathrm {O}(n)}]\)
where \(\Psi _F^n\) is an orthogonal spectrum with an \(O(n)\) -action. It follows that pre-realification of \(F\) is levelwise weakly equivalent to the functor
\(W \longmapsto \Omega ^\infty [(S^{\mathbb {R}^n \otimes _\mathbb {R}W_\mathbb {R}} \wedge \Psi ^n_F)_{h\mathrm {O}(n)}].\)
Using the derived change of group functor, we construct an orthogonal spectrum with an action of \(\mathrm {U}(n)\) ,
\(\mathrm {U}(n)_+ \wedge _{\mathrm {O}(n)}^{L} \Psi _F^n := \mathrm {U}(n)_+ \wedge _{\mathrm {O}(n)} (E\mathrm {O}(n)_+ \wedge \Psi _F^n).\)
By the classification of \(n\) -homogeneous unitary functors, Proposition REF , there is an \(n\) -homogeneous functor \(F^{\prime }\) associated to the above spectrum, given by
\(W \longmapsto \Omega ^\infty [(S^{n \otimes _W̏} \wedge (\mathrm {U}(n)_+ \wedge _{\mathrm {O}(n)} (E\mathrm {O}(n)_+ \wedge \Psi _F^n)))_{h\mathrm {U}(n)}].\)
By [1]}, \(F^{\prime }(W)\) is isomorphic to
\(\Omega ^\infty [\mathrm {U}(n)_+ \wedge _{\mathrm {O}(n)} ((\iota ^*S^{n \otimes _W̏} \wedge (E\mathrm {O}(n)_+ \wedge \Psi _F^n)))_{h(\mathrm {U}(n)}].\)
The \(\mathrm {U}(n)\) -action on \(\mathrm {U}(n)_+ \wedge _{\mathrm {O}(n)} ((\iota ^*S^{n \otimes _W̏} \wedge (E\mathrm {O}(n)_+ \wedge \Psi _F^n)))\) is free (\((E\mathrm {O}(n)_+\) is a free \(\mathrm {O}(n)\) -space), hence taking homotopy orbits equates to taking strict orbits. Hence there is an isomorphism
\(F^{\prime }(W) \cong \Omega ^\infty [(\mathrm {U}(n)_+ \wedge _{\mathrm {O}(n)} ((\iota ^*S^{n \otimes _W̏} \wedge (E\mathrm {O}(n)_+ \wedge \Psi _F^n))))/\mathrm {U}(n)].\)
The strict \(\mathrm {U}(n)\) -orbits of the spectrum \(\mathrm {U}(n)_+ \wedge _{\mathrm {O}(n)} ((\iota ^*S^{n \otimes _W̏} \wedge (E\mathrm {O}(n)_+ \wedge \Psi _F^n))\) are isomorphic to the \(\mathrm {O}(n)\) -orbits of the spectrum, \(\iota ^*S^{n \otimes _W̏} \wedge (E\mathrm {O}(n)_+ \wedge \Psi _F^n))\) , hence \(F^{\prime }(W)\) is isomorphic to
\(\Omega ^\infty [(\iota ^*S^{n \otimes _W̏} \wedge (E\mathrm {O}(n)_+ \wedge \Psi _F^n)))/\mathrm {O}(n)].\)
This last is precisely
\(\Omega ^\infty [(\iota ^*S^{n \otimes _W̏} \wedge \Psi _F^n)_{h\mathrm {O}(n)}]\)
as homotopy orbits is the left derived functor of strict orbits and smashing with \(E\mathrm {O}(n)_+\) is a cofibrant replacement in the projective model structure.
Since the action of \(\mathrm {O}(n)\) on \(\iota ^*S^{n \otimes _W̏}\) is equivalent to the \(\mathrm {O}(n)\) action on \(S^{\mathbb {R}^n \otimes _\mathbb {R}W_\mathbb {R}}\) and the one-point compactification are isomorphic, the above infinite loop space is isomorphic to
\(\Omega ^\infty [(S^{\mathbb {R}^n \otimes _\mathbb {R}W_\mathbb {R}} \wedge \Psi _F^n)_{h\mathrm {O}(n)}].\)
By the characterisation of \(n\) -homogeneous orthogonal functors, we see that this is levelwise weakly equivalent to \(F(W_\mathbb {R}) = (r^*F)(W).\)
| [1] | [
[
1166,
1169
]
] | https://openalex.org/W2075488415 |
ec524d83-4cff-40c6-9118-616eb4472427 | We argue by induction on the polynomial degree. The case \(n=0\) follows by definition. Assume the map \(r^*T_{n-1}^\mathbf {O} F \rightarrow T_{n-1}^\mathbf {U}(r^*T_{n-1}^\mathbf {O} F)\) is a levelwise weak equivalence. There is a homotopy fibre sequence
\(T_n^\mathbf {O} F \longrightarrow T_{n-1}^\mathbf {O} F \longrightarrow R_n^\mathbf {O} F\)
where \(R_n^\mathbf {O} F\) is \(n\) -homogeneous, since \(F\) satisfies the conditions of [1]}. Lemma REF implies that \(r^* R_n^\mathbf {O} F\) is \(n\) -homogeneous in \(\mathcal {E}_0^\mathbf {U}\) , and in particular \(n\) -polynomial. As homotopy fibres of maps between \(n\) -polynomial objects are \(n\) -polynomial, the homotopy fibre of the map \(r^*T_{n-1}^\mathbf {O} F \rightarrow r^* R_n^\mathbf {O} F\) is \(n\) -polynomial. Computation of homotopy fibres is levelwise, hence the homotopy fibre in question is \(r^* T_n^\mathbf {O} F\) , and it follows that
\(r^* T_n^\mathbf {O} F \longrightarrow T_n^\mathbf {U}(r^* T_n^\mathbf {O} F)\)
is a levelwise weak equivalence.
| [1] | [
[
448,
451
]
] | https://openalex.org/W4239559895 |
873bc9ce-fc96-423e-bacb-88fde79e3dfd | The acyclic fibrations in the \(n\) –polynomial model structure are levelwise, hence both terms in the composite, and resultingly their composite, preserves these. It suffices by [1]} to show that the composite preserves fibrations between fibrant objects, which are the levelwise fibrations by [2]}. It hence suffices to show that \(F\) preserves fibrant objects. Let \(F\) be a \(n\) –polynomial orthogonal functor, then \(\mathrm {red}(F)\) is also \(n\) –polynomial, and reduced. An application of Theorem REF implies the result.
| [2] | [
[
295,
298
]
] | https://openalex.org/W1583122470 |
20d38e3a-d77c-47df-b074-6fce7b7fadec | The notion of agreement plays a central role in the theory or orthogonal and unitary calculus, for example it is crucial to the proof that the \(n\) -th polynomial approximation in \(n\) -polynomial, see [1]}. The pre-realification and pre-complexification functors behave well with respect to functors which agree to a certain order.
| [1] | [
[
204,
207
]
] | https://openalex.org/W4239559895 |
4cf27916-a6d0-4336-a606-32bfcf030c10 | We have constructed a Quillen adjunction between the orthogonal and unitary \(n\) -homogeneous model structures. To give a complete comparison of the theories we must address the comparisons between the other two categories in the zig-zag of Quillen equivalences of Barnes and Oman [1]} and the author [2]}. We start by addressing the relationship between the categories of spectra. For this, we recall the definitions and model structures involved.
| [1] | [
[
282,
285
]
] | https://openalex.org/W2034710066 |
db9b6bbc-117d-4939-9a14-ab149b71437b | compare [1]}.
| [1] | [
[
8,
11
]
] | https://openalex.org/W2075488415 |
769e3527-e453-4f89-98a8-61ec30911928 | These categories also come with projective and stable model structures constructed analogously to those of Proposition REF . These new intermediate categories will now act as intermediate categories between the standard intermediate categories of orthogonal and unitary calculus. Further, the new intermediate categories equipped with their \(n\) -stable model structure are Quillen equivalent to spectra with an appropriate group action. The proofs of the following two results follow similarly to [1]} and [2]}.
| [1] | [
[
499,
502
]
] | https://openalex.org/W2034710066 |
2181f64d-7200-4ec6-91d7-055dfe27a120 | where the isomorphism follows from [1]}.
| [1] | [
[
35,
38
]
] | https://openalex.org/W2075488415 |
b0855016-27d8-4693-ba77-50723aa9e165 | By the zig-zag of Quillen equivalences, [1]} the composite
\({L}\operatorname{\mathrm {res}}_0^n/\mathrm {O}(n) \circ {R}(\beta _n)^*\)
applied to an orthogonal spectrum \(\Theta \) with an action of \(\mathrm {O}(n)\) , is levelwise weakly equivalent to the functor \(F\) , defined by the formula
\(V \longmapsto \Omega ^\infty [(S^{\mathbb {R}^n \otimes V} \wedge \Theta )_{h\mathrm {O}(n)}].\)
This functor is \(n\) –homogeneous, hence also reduced, so \({R}r^*\mathrm {red}(F)\) is levelwise weakly equivalent to \(\mathbb {R}r^*F\) . The zig-zag of Quillen equivalences from unitary calculus, [2]}, together with inflating \(\Theta \) to a spectrum with an action of \(\mathrm {U}(n)\) gives a similar characterisation in terms of an \(n\) -homogeneous functor. The result then follows by Proposition REF .
| [1] | [
[
40,
43
]
] | https://openalex.org/W2034710066 |
8be61fa0-2ba6-462c-9e82-0c3b83709155 | These categories also come with projective and stable model structures constructed analogously to those of Proposition REF . These new intermediate categories will now act as intermediate categories between the standard intermediate categories of orthogonal and unitary calculus. Further, the new intermediate categories equipped with their \(n\) -stable model structure are Quillen equivalent to spectra with an appropriate group action. The proofs of the following two results follow similarly to [1]} and [2]}.
| [1] | [
[
499,
502
]
] | https://openalex.org/W2034710066 |
d9c60347-3403-4dab-a8c1-2d470b426e3e | where the isomorphism follows from [1]}.
| [1] | [
[
35,
38
]
] | https://openalex.org/W2075488415 |
729488d8-56b5-4a8a-bca0-07bd7c9f778b | By the zig-zag of Quillen equivalences, [1]} the composite
\({L}\operatorname{\mathrm {res}}_0^n/\mathrm {O}(n) \circ {R}(\beta _n)^*\)
applied to an orthogonal spectrum \(\Theta \) with an action of \(\mathrm {O}(n)\) , is levelwise weakly equivalent to the functor \(F\) , defined by the formula
\(V \longmapsto \Omega ^\infty [(S^{\mathbb {R}^n \otimes V} \wedge \Theta )_{h\mathrm {O}(n)}].\)
This functor is \(n\) –homogeneous, hence also reduced, so \({R}r^*\mathrm {red}(F)\) is levelwise weakly equivalent to \(\mathbb {R}r^*F\) . The zig-zag of Quillen equivalences from unitary calculus, [2]}, together with inflating \(\Theta \) to a spectrum with an action of \(\mathrm {U}(n)\) gives a similar characterisation in terms of an \(n\) -homogeneous functor. The result then follows by Proposition REF .
| [1] | [
[
40,
43
]
] | https://openalex.org/W2034710066 |
f032747c-5eac-418b-9556-3006178ab0f2 | Advances in the ability to successfully train very deep neural networks have been key to improving performance in image recognition, language modeling, and many other domains [1]}, [2]}, [3]}, [4]}, [5]}, [6]}. Graph Neural Networks (GNNs) are a family of deep networks that operate on graph structured data by iteratively passing learned messages over the graph's structure [7]}, [8]}, [9]}, [10]}. While GNNs are very effective in a wide variety of tasks [11]}, [12]}, [13]}, deep GNNs, which perform more than 5-10 message passing steps, have not typically yielded better performance than shallower GNNs [14]}, [15]}, [16]}. While in principle deep GNNs should have greater expressivity and ability to capture complex functions, it has been proposed that in practice “oversmoothing” [17]} and “bottleneck effects” [18]} limit the potential benefits of deep GNNs. The purpose of this work is to reap the benefits of deep GNNs while avoiding such limitations.
| [18] | [
[
817,
821
]
] | https://openalex.org/W3034190530 |
20440e8e-e71f-464e-96ee-7c0ffbcb379d | Oversmoothing is a proposed phenomenon where a GNN's latent node representations become increasing similar over successive steps of message passing [1]}. Once these representations are oversmoothed, adding further steps does not add expressive capacity, and so performance does not improve. Bottleneck effects are thought to limit the ability of a deep GNN to communicate information over long ranges, because as the number of steps increases and causes the receptive fields to grow, the intermediate nodes must propagate information between larger and larger sets of peripheral nodes, limiting the ability to precisely transmit individual pieces of information [2]}. We speculate that these are reasons why previous strong GNN results on the molecular property benchmarks we study here, Open Catalyst 2020 (OC20) [3]} and QM9 [4]}, report depths of only 4 [5]} and 8 [6]}, even with large hyperparameter searches.
| [2] | [
[
662,
665
]
] | https://openalex.org/W3034190530 |
fc5a2298-6a8a-4f4c-a8e7-56ba197e71e6 | Scaling Graph Neural Networks with Depth. Recent work has aimed to understand why it is challenging to realise the benefits of training very deep GNNs [1]}. A key contribution has been the analysis of “oversmoothing” which describes how all node features become almost identical in GCNs after a few layers. Since first being noted in [2]} oversmoothing has been studied extensively and regularisation techniques have been suggested to overcome it [3]}, [4]}, [5]}, [6]}, [7]}, but these studies have not shown consistent improvements as depth increases. Another insight has been the analysis of the bottleneck effect in [8]} which may be alleviated by using a fully connected graph. Finally, deep GNNs have been trained using techniques developed for CNNs [9]}, [10]}, but with no conclusive improvement from depth.
| [8] | [
[
620,
623
]
] | https://openalex.org/W3034190530 |
f43d9699-4709-425d-96a2-9e74b27e6321 | An alternative approach to embedding symmetries is to design a rotation equivariant neural network that predicts quantities such as forces directly [1]}, [2]}, [3]}, [4]}, [5]}, [6]}, [7]}. However, the benefits of such models are typically seen on smaller data (e.g. [7]}), and such approaches have yet to achieve SOTA performance on large quantum chemistry datasets. Not all competitive models embed rotation symmetries—notably ForceNet [9]} does not enforce rotation equivariance.
| [6] | [
[
178,
181
]
] | https://openalex.org/W2970663744 |
f78f802e-9c86-44c6-8045-ead0b4c44fef | Processor. The processor consists of "blocks", where each block contain a stack of Interaction Networks (each one known as a "message passing layer") with identical structure but different weights. The node and edge functions are 3 layer MLPs activated with shifted softplus (\(ssp(x)=\ln (0.5e^x + 0.5)\) , [1]}), followed by layer normalisation [2]}. Residual connections on both the nodes and edges are added to each message passing layer. We recurrently apply each block, a process we we call "block iteration". We calculate the number of message passing layers by multiplying the number of blocks by the block size, for example 10 applications of a block of size 10 corresponds to 100 (\(10\times 10\) ) message passing layers.
| [2] | [
[
347,
350
]
] | https://openalex.org/W3037932933 |
1b9f8ffb-4dee-4e63-9146-adcf9f672a3d | The ability to adapt flexibly and efficiently to novel environments is one of the most distinctive and compelling features of the human mind. It has been suggested that humans do so by learning internal models which not only contain abstract representations of the world, but also encode generalisable, structural relationships within the environment [1]}. This latter aspect, it is conjectured, is what allows humans to adapt efficiently and selectively. Recent efforts have been made to mimic this kind of representation in machine learning. World models [2]} aim to capture the dynamics of an environment by distilling past experience into a parametric predictive model. Advances in latent variable models have enabled the learning of world models in a compact latent space [2]}, [4]}, [5]}, [6]}, [7]} from high-dimensional observations such as images. Whilst these models have enabled agents to act in complex environments via planning [5]}, [9]} or learning parametric policies [10]}, [2]}, structurally adapting to changes in the environment remains a significant challenge. The consequence of this limitation is particularly pronounced when deploying learning agents to environments, where distribution shifts occur. As such, we argue that it is beneficial to build structural world models that afford modular and efficient adaptation, and that causal modeling offers a tantalising prospect to discover such structure from observations.
| [1] | [
[
351,
354
]
] | https://openalex.org/W2854445389 |
08de4475-f3eb-4937-85e2-c73fd1fc082c | The ability to adapt flexibly and efficiently to novel environments is one of the most distinctive and compelling features of the human mind. It has been suggested that humans do so by learning internal models which not only contain abstract representations of the world, but also encode generalisable, structural relationships within the environment [1]}. This latter aspect, it is conjectured, is what allows humans to adapt efficiently and selectively. Recent efforts have been made to mimic this kind of representation in machine learning. World models [2]} aim to capture the dynamics of an environment by distilling past experience into a parametric predictive model. Advances in latent variable models have enabled the learning of world models in a compact latent space [2]}, [4]}, [5]}, [6]}, [7]} from high-dimensional observations such as images. Whilst these models have enabled agents to act in complex environments via planning [5]}, [9]} or learning parametric policies [10]}, [2]}, structurally adapting to changes in the environment remains a significant challenge. The consequence of this limitation is particularly pronounced when deploying learning agents to environments, where distribution shifts occur. As such, we argue that it is beneficial to build structural world models that afford modular and efficient adaptation, and that causal modeling offers a tantalising prospect to discover such structure from observations.
| [2] | [
[
557,
560
],
[
777,
780
],
[
991,
994
]
] | https://openalex.org/W2890208753 |
c8989c27-f64c-4371-908f-7b9a8d54bb3b | The ability to adapt flexibly and efficiently to novel environments is one of the most distinctive and compelling features of the human mind. It has been suggested that humans do so by learning internal models which not only contain abstract representations of the world, but also encode generalisable, structural relationships within the environment [1]}. This latter aspect, it is conjectured, is what allows humans to adapt efficiently and selectively. Recent efforts have been made to mimic this kind of representation in machine learning. World models [2]} aim to capture the dynamics of an environment by distilling past experience into a parametric predictive model. Advances in latent variable models have enabled the learning of world models in a compact latent space [2]}, [4]}, [5]}, [6]}, [7]} from high-dimensional observations such as images. Whilst these models have enabled agents to act in complex environments via planning [5]}, [9]} or learning parametric policies [10]}, [2]}, structurally adapting to changes in the environment remains a significant challenge. The consequence of this limitation is particularly pronounced when deploying learning agents to environments, where distribution shifts occur. As such, we argue that it is beneficial to build structural world models that afford modular and efficient adaptation, and that causal modeling offers a tantalising prospect to discover such structure from observations.
| [4] | [
[
783,
786
]
] | https://openalex.org/W2963430173 |
a5482cb5-c76d-471b-ab37-9619845f13f3 | The ability to adapt flexibly and efficiently to novel environments is one of the most distinctive and compelling features of the human mind. It has been suggested that humans do so by learning internal models which not only contain abstract representations of the world, but also encode generalisable, structural relationships within the environment [1]}. This latter aspect, it is conjectured, is what allows humans to adapt efficiently and selectively. Recent efforts have been made to mimic this kind of representation in machine learning. World models [2]} aim to capture the dynamics of an environment by distilling past experience into a parametric predictive model. Advances in latent variable models have enabled the learning of world models in a compact latent space [2]}, [4]}, [5]}, [6]}, [7]} from high-dimensional observations such as images. Whilst these models have enabled agents to act in complex environments via planning [5]}, [9]} or learning parametric policies [10]}, [2]}, structurally adapting to changes in the environment remains a significant challenge. The consequence of this limitation is particularly pronounced when deploying learning agents to environments, where distribution shifts occur. As such, we argue that it is beneficial to build structural world models that afford modular and efficient adaptation, and that causal modeling offers a tantalising prospect to discover such structure from observations.
| [5] | [
[
789,
792
],
[
941,
944
]
] | https://openalex.org/W2900152462 |
94d7b642-d49d-423a-b3b6-76349f867e75 | The ability to adapt flexibly and efficiently to novel environments is one of the most distinctive and compelling features of the human mind. It has been suggested that humans do so by learning internal models which not only contain abstract representations of the world, but also encode generalisable, structural relationships within the environment [1]}. This latter aspect, it is conjectured, is what allows humans to adapt efficiently and selectively. Recent efforts have been made to mimic this kind of representation in machine learning. World models [2]} aim to capture the dynamics of an environment by distilling past experience into a parametric predictive model. Advances in latent variable models have enabled the learning of world models in a compact latent space [2]}, [4]}, [5]}, [6]}, [7]} from high-dimensional observations such as images. Whilst these models have enabled agents to act in complex environments via planning [5]}, [9]} or learning parametric policies [10]}, [2]}, structurally adapting to changes in the environment remains a significant challenge. The consequence of this limitation is particularly pronounced when deploying learning agents to environments, where distribution shifts occur. As such, we argue that it is beneficial to build structural world models that afford modular and efficient adaptation, and that causal modeling offers a tantalising prospect to discover such structure from observations.
| [6] | [
[
795,
798
]
] | https://openalex.org/W2786019934 |
45119972-e403-4b67-a703-dab9a4614348 | The ability to adapt flexibly and efficiently to novel environments is one of the most distinctive and compelling features of the human mind. It has been suggested that humans do so by learning internal models which not only contain abstract representations of the world, but also encode generalisable, structural relationships within the environment [1]}. This latter aspect, it is conjectured, is what allows humans to adapt efficiently and selectively. Recent efforts have been made to mimic this kind of representation in machine learning. World models [2]} aim to capture the dynamics of an environment by distilling past experience into a parametric predictive model. Advances in latent variable models have enabled the learning of world models in a compact latent space [2]}, [4]}, [5]}, [6]}, [7]} from high-dimensional observations such as images. Whilst these models have enabled agents to act in complex environments via planning [5]}, [9]} or learning parametric policies [10]}, [2]}, structurally adapting to changes in the environment remains a significant challenge. The consequence of this limitation is particularly pronounced when deploying learning agents to environments, where distribution shifts occur. As such, we argue that it is beneficial to build structural world models that afford modular and efficient adaptation, and that causal modeling offers a tantalising prospect to discover such structure from observations.
| [7] | [
[
801,
804
]
] | https://openalex.org/W2889347284 |
89255aa9-01c1-4d1e-8fe6-71752bae0391 | The ability to adapt flexibly and efficiently to novel environments is one of the most distinctive and compelling features of the human mind. It has been suggested that humans do so by learning internal models which not only contain abstract representations of the world, but also encode generalisable, structural relationships within the environment [1]}. This latter aspect, it is conjectured, is what allows humans to adapt efficiently and selectively. Recent efforts have been made to mimic this kind of representation in machine learning. World models [2]} aim to capture the dynamics of an environment by distilling past experience into a parametric predictive model. Advances in latent variable models have enabled the learning of world models in a compact latent space [2]}, [4]}, [5]}, [6]}, [7]} from high-dimensional observations such as images. Whilst these models have enabled agents to act in complex environments via planning [5]}, [9]} or learning parametric policies [10]}, [2]}, structurally adapting to changes in the environment remains a significant challenge. The consequence of this limitation is particularly pronounced when deploying learning agents to environments, where distribution shifts occur. As such, we argue that it is beneficial to build structural world models that afford modular and efficient adaptation, and that causal modeling offers a tantalising prospect to discover such structure from observations.
| [9] | [
[
947,
950
]
] | https://openalex.org/W3025660841 |
4c794f34-046f-4a06-a4e4-21bd5fd8c69e | The ability to adapt flexibly and efficiently to novel environments is one of the most distinctive and compelling features of the human mind. It has been suggested that humans do so by learning internal models which not only contain abstract representations of the world, but also encode generalisable, structural relationships within the environment [1]}. This latter aspect, it is conjectured, is what allows humans to adapt efficiently and selectively. Recent efforts have been made to mimic this kind of representation in machine learning. World models [2]} aim to capture the dynamics of an environment by distilling past experience into a parametric predictive model. Advances in latent variable models have enabled the learning of world models in a compact latent space [2]}, [4]}, [5]}, [6]}, [7]} from high-dimensional observations such as images. Whilst these models have enabled agents to act in complex environments via planning [5]}, [9]} or learning parametric policies [10]}, [2]}, structurally adapting to changes in the environment remains a significant challenge. The consequence of this limitation is particularly pronounced when deploying learning agents to environments, where distribution shifts occur. As such, we argue that it is beneficial to build structural world models that afford modular and efficient adaptation, and that causal modeling offers a tantalising prospect to discover such structure from observations.
| [10] | [
[
984,
988
]
] | https://openalex.org/W2995298643 |
51fde515-9424-45c5-9124-91d45fd4d3a1 | Causality plays a central role in understanding distribution changes, which can be modelled as causal interventions [1]}. The Sparse Mechanism Shift hypothesis [1]}, [3]} (SMS) states that naturally occurring shifts in the data distribution can be attributed to sparse and local changes in the causal generative process. This implies that many causal mechanisms remain invariant across domains [4]}, [5]}, [6]}. In this light, learning a causal model of the environment enables agents to reason about distribution shifts and to exploit the invariance of learnt causal mechanisms across different environments. Hence, we posit that world models with a causal structure can facilitate modular transfer of knowledge. To date, however, methods for causal discovery [7]}, [8]}, [9]}, [10]}, [11]} require access to abstract causal variables to learn causal models from data. These are not typically available in the context of world model learning, where we wish to operate directly on high-dimensional observations.
| [1] | [
[
116,
119
],
[
160,
163
]
] | https://openalex.org/W3135588948 |
283f40a6-fcb1-423c-bae1-4d045cd5a4a0 | Causality plays a central role in understanding distribution changes, which can be modelled as causal interventions [1]}. The Sparse Mechanism Shift hypothesis [1]}, [3]} (SMS) states that naturally occurring shifts in the data distribution can be attributed to sparse and local changes in the causal generative process. This implies that many causal mechanisms remain invariant across domains [4]}, [5]}, [6]}. In this light, learning a causal model of the environment enables agents to reason about distribution shifts and to exploit the invariance of learnt causal mechanisms across different environments. Hence, we posit that world models with a causal structure can facilitate modular transfer of knowledge. To date, however, methods for causal discovery [7]}, [8]}, [9]}, [10]}, [11]} require access to abstract causal variables to learn causal models from data. These are not typically available in the context of world model learning, where we wish to operate directly on high-dimensional observations.
| [3] | [
[
166,
169
]
] | https://openalex.org/W2914607694 |
01556cf1-3fc5-4feb-9c40-2f5f0894f1b3 | Causality plays a central role in understanding distribution changes, which can be modelled as causal interventions [1]}. The Sparse Mechanism Shift hypothesis [1]}, [3]} (SMS) states that naturally occurring shifts in the data distribution can be attributed to sparse and local changes in the causal generative process. This implies that many causal mechanisms remain invariant across domains [4]}, [5]}, [6]}. In this light, learning a causal model of the environment enables agents to reason about distribution shifts and to exploit the invariance of learnt causal mechanisms across different environments. Hence, we posit that world models with a causal structure can facilitate modular transfer of knowledge. To date, however, methods for causal discovery [7]}, [8]}, [9]}, [10]}, [11]} require access to abstract causal variables to learn causal models from data. These are not typically available in the context of world model learning, where we wish to operate directly on high-dimensional observations.
| [4] | [
[
394,
397
]
] | https://openalex.org/W2144020560 |
27adb62f-9203-4ec5-a581-fca74fb2e168 | Causality plays a central role in understanding distribution changes, which can be modelled as causal interventions [1]}. The Sparse Mechanism Shift hypothesis [1]}, [3]} (SMS) states that naturally occurring shifts in the data distribution can be attributed to sparse and local changes in the causal generative process. This implies that many causal mechanisms remain invariant across domains [4]}, [5]}, [6]}. In this light, learning a causal model of the environment enables agents to reason about distribution shifts and to exploit the invariance of learnt causal mechanisms across different environments. Hence, we posit that world models with a causal structure can facilitate modular transfer of knowledge. To date, however, methods for causal discovery [7]}, [8]}, [9]}, [10]}, [11]} require access to abstract causal variables to learn causal models from data. These are not typically available in the context of world model learning, where we wish to operate directly on high-dimensional observations.
| [5] | [
[
400,
403
]
] | https://openalex.org/W1905064697 |
3304177f-2e17-4cdd-b12b-9dc87b8ac588 | Causality plays a central role in understanding distribution changes, which can be modelled as causal interventions [1]}. The Sparse Mechanism Shift hypothesis [1]}, [3]} (SMS) states that naturally occurring shifts in the data distribution can be attributed to sparse and local changes in the causal generative process. This implies that many causal mechanisms remain invariant across domains [4]}, [5]}, [6]}. In this light, learning a causal model of the environment enables agents to reason about distribution shifts and to exploit the invariance of learnt causal mechanisms across different environments. Hence, we posit that world models with a causal structure can facilitate modular transfer of knowledge. To date, however, methods for causal discovery [7]}, [8]}, [9]}, [10]}, [11]} require access to abstract causal variables to learn causal models from data. These are not typically available in the context of world model learning, where we wish to operate directly on high-dimensional observations.
| [6] | [
[
406,
409
]
] | https://openalex.org/W879220392 |
911c1c88-36d0-4c96-afae-1a3d5402b709 | Causality plays a central role in understanding distribution changes, which can be modelled as causal interventions [1]}. The Sparse Mechanism Shift hypothesis [1]}, [3]} (SMS) states that naturally occurring shifts in the data distribution can be attributed to sparse and local changes in the causal generative process. This implies that many causal mechanisms remain invariant across domains [4]}, [5]}, [6]}. In this light, learning a causal model of the environment enables agents to reason about distribution shifts and to exploit the invariance of learnt causal mechanisms across different environments. Hence, we posit that world models with a causal structure can facilitate modular transfer of knowledge. To date, however, methods for causal discovery [7]}, [8]}, [9]}, [10]}, [11]} require access to abstract causal variables to learn causal models from data. These are not typically available in the context of world model learning, where we wish to operate directly on high-dimensional observations.
| [7] | [
[
761,
764
]
] | https://openalex.org/W1524326598 |
3688fb92-0d60-4ac6-a71b-8376d066b415 | Causality plays a central role in understanding distribution changes, which can be modelled as causal interventions [1]}. The Sparse Mechanism Shift hypothesis [1]}, [3]} (SMS) states that naturally occurring shifts in the data distribution can be attributed to sparse and local changes in the causal generative process. This implies that many causal mechanisms remain invariant across domains [4]}, [5]}, [6]}. In this light, learning a causal model of the environment enables agents to reason about distribution shifts and to exploit the invariance of learnt causal mechanisms across different environments. Hence, we posit that world models with a causal structure can facilitate modular transfer of knowledge. To date, however, methods for causal discovery [7]}, [8]}, [9]}, [10]}, [11]} require access to abstract causal variables to learn causal models from data. These are not typically available in the context of world model learning, where we wish to operate directly on high-dimensional observations.
| [8] | [
[
767,
770
]
] | https://openalex.org/W3133236490 |
05023560-4e8c-4204-b99c-9fc499d03b57 | Causality plays a central role in understanding distribution changes, which can be modelled as causal interventions [1]}. The Sparse Mechanism Shift hypothesis [1]}, [3]} (SMS) states that naturally occurring shifts in the data distribution can be attributed to sparse and local changes in the causal generative process. This implies that many causal mechanisms remain invariant across domains [4]}, [5]}, [6]}. In this light, learning a causal model of the environment enables agents to reason about distribution shifts and to exploit the invariance of learnt causal mechanisms across different environments. Hence, we posit that world models with a causal structure can facilitate modular transfer of knowledge. To date, however, methods for causal discovery [7]}, [8]}, [9]}, [10]}, [11]} require access to abstract causal variables to learn causal models from data. These are not typically available in the context of world model learning, where we wish to operate directly on high-dimensional observations.
| [9] | [
[
773,
776
]
] | https://openalex.org/W2801890059 |