Zhongchao Wu is currently a retired Professor. He received his Ph.D. degree in Physics from Cambridge University in 1984 under the supervision of Prof. Stephen Hawking, and his doctoral thesis was titled “Cosmological models and the inflationary universe”. His research mainly focuses on cosmology
Using the synchronous coordinates, the creation of a Schwarzschild black hole immersed in a de Sitter spacetime can be viewed as a coherent creation of a collection of timelike geodesics. The previously supposed conical singularities do not exist at the horizons of the constrained instanton. Instead, the unavoidable irregularity is presented as a non-vanishing second fundamental form elsewhere at the quantum transition 3-surface. The same arguments can be applied to charged, topological, or higher dimensional black hole cases.
Graphical Abstract
Schematic diagram of a black hole creation. The spherical symmetry of space is simplified to circumferential symmetry around the central axis here.
Abstract
Using the synchronous coordinates, the creation of a Schwarzschild black hole immersed in a de Sitter spacetime can be viewed as a coherent creation of a collection of timelike geodesics. The previously supposed conical singularities do not exist at the horizons of the constrained instanton. Instead, the unavoidable irregularity is presented as a non-vanishing second fundamental form elsewhere at the quantum transition 3-surface. The same arguments can be applied to charged, topological, or higher dimensional black hole cases.
Public Summary
Using the synchronous coordinates, the creation of a Schwarzschild black hole immersed in a de Sitter spacetime can be viewed as a coherent creation of a collection of timelike geodesics.
The previously supposed conical singularities do not exist at the horizons of the constrained instanton. Instead, the unavoidable irregularity is presented as a non-vanishing second fundamental form elsewhere at the quantum transition 3-surface.
Question generation[1, 2] (QG) is defined as the task of generating fluent, meaningful questions automatically from texts with optional answers, so it can be mainly divided into two streams: answer-aware QG[3] that requires answers, and answer-agnostic QG[4] that does not. QG is the reverse task of question answering (QA), which is a long-standing and valuable task helping computers achieve machine reading comprehension[5, 6], dating back to the 1960s[7]. As with many other supervised learning[8, 9] tasks, QA will also encounter the lack of annotated data in spite of the fact that annotated data sometimes make the most essential part of the whole work.
QG is a popular choice for data augmentation for QA to alleviate insufficient labeled data. With the continuous development of Internet technology, it is becoming increasingly easier to obtain valuable data from the Internet. However, question-answer pairs (as shown in Table 1) are still such expensive corpora that typically require manual annotation by crowdsourcing before being used for supervised learning on QA and QG tasks. To alleviate the high-cost problem of generating question-answer pairs, it is natural to consider answer-agnostic QG, since its only input is raw text.
Table
1.
A case of QA-pairs generated by our QGAE model: the model accepts unannotated texts as input, extracts the highlighted phrase “Lorentz’s law” as an answer, then uses this answer to make question generation.
Input context: Through combining the definition of electric current as the time rate of change of electric charge, a rule of vector multiplication called
Lorentz’s law describes the force on a charge moving in a magnetic field. The connection between electricity and magnetism allows for the description of a unified electromagnetic force that acts on a charge. This force can be written as a sum of the electrostatic force (due to the electric field) and the magnetic force (due to the magnetic field).
Extracted answer: Lorentz’s law
Generated question: What describes the force on a charge moving in a magnetic field?
Although labeled answers are not necessary, answer-agnostic QG is still facing a great challenge. Most previous works focused on providing additional information to their models by leveraging named entity recognition (NER)[10] to obtain extra linguistic features, adding answer position features[11], using knowledge graphs[12], and some other methods to improve the generation effect. These methods effectively improve the fluency and accuracy of generated texts, but answer-agnostic QG still performs worse than answer-aware QG. Thus, answer-aware QG may play an irreplaceable role, and changing answer-agnostic QG to answer-aware QG is a good choice. Apart from this, there is still an obstacle in generating question-answer pairs that answer-agnostic QG can’t generate answers. To address this issue, researchers often add an additional measure for question-answer pair generation: answer extraction. Compared with generating an answer, extracting an exact span in the context is much simpler.
Explicitly extracting candidate answers will not only resolve the demand for the lack of answers but also can transform answer-agnostic QG into answer-aware QG. As shown in Fig. 1, some works such as RGF[13] (retrieve-generate filter) proposed a multi-stage pipeline method to handle the problem. A multi-stage pipeline method is often designed in complexity, including several parts, and each part may need different inputs. Some early RNN-based[14–17] works optimized pipeline methods in an end-to-end way, which makes the overall structure lighter and faster. Though pre-trained language models (PLMs) have occupied dominance in both natural language generation and understanding, there is still no end-to-end work using pre-trained models to generate question-answer pairs. We are sure there is enough potential for PLMs to achieve the task.
Figure
1.
The difference between multi-stage methods and end-to-end models is that a multi-stage method usually has more than one model in the whole workflow. In every stage, a multi-stage method may need to deal with different inputs and outputs, while on the contrary, an end-to-end model only needs a definite kind of input.
In this study, we are motivated by the weak performance of answer-agnostic QG compared to answer-aware QG, inspired by the combination of QG and AE tasks, trying to propose an answer-agnostic question generation model called question generation with answer extractor (QGAE) to alleviate the high demand for large-scale QA pairs. QGAE is a multi-task model that requires only raw texts as input and can achieve the dual tasks: answer extraction and question generation. We design our model based on the PLM model BART[18], which has dual encoders and a decoder to generate questions and extract answers in parallel. In our study, question generation is the main task, which is the most challenging part similar to all other generation tasks for generated texts’ high syntactic diversity and semantic substitutability, so we pay more attention and assign a higher weight to the corresponding module. Therefore answer extraction is considered an auxiliary task. The design not only makes it feasible to turn answer-agnostic question generation into answer-aware question generation but also enables the model to be considered capable of generating question-answer pairs. The contributions of this paper are summarized as follows:
● We are the first to propose a new end-to-end model using PLMs, which is called QGAE for answer-agnostic question generation.
● The QGAE model generates question-answer pairs from unannotated texts without requiring any additional information.
● Our model achieves state-of-the-art performance in generating high-quality question-answer pairs, outperforming existing methods by a significant margin.
The rest of this paper is organized as follows. In Section 2, we review the related works of question generation and answer extraction. In Section 3, we formulate the QG task and AE task. In Section 4, we describe each module of our QGAE model. In Section 5, we introduce our experiment in detail. In the last Section 6, we conclude this work and give a detailed analysis.
2.
Related works
2.1
Question generation
The QG field was devoted great interest by researchers for its great potential benefits; therefore, it has made great progress in application scenarios such as data augmentation[19], chatbots[20], machine reading comprehension[21], and intelligent tutors[22].
In the neural model age, Du et al.[4] proposed the first neural QG model focused on answer-agnostic QG. They investigated the effect of encoding sentence- vs. paragraph-level information by using an attention-based model and found that as the size of the input text increased, the evaluation score of the output decreased. To deal with the rare or unknown word problem, Gulcehre et al.[23] proposed a copy mechanism that was first used in the neural machine translation[24] to solve the out-of-vocabulary problem. This mechanism was absorbed in the QG task and widely used. Following the old experience of rule-based QG[25], Wu et al.[26] suggested two new strategies to deal with this task: question type prediction and a copy loss mechanism. Du et al.[15] combined answer extraction and question generation in an LSTM[27] model including answer feature embedding, denoting answer span with the usual BIO tagging scheme[28].
In the transformer-based[29] PLM era, compared to auto-encoder models, auto-regressive[30] models are widely picked as baselines for the QG task. Laban et al.[20] fine-tuned a GPT2[31] as the base part of a question-driven news chatbot. Wang et al.[32] leveraged BART to propose QAGS (question answering and generation for summarization) to evaluate automatic summarization. Bhambhoria et al.[33] leveraged T5[34] to generate QA pairs for COVID-19 literature. Paranjape et al.[13] developed a retrieve-generate filter (RGF) technique to create counterfactual evaluation and training data with minimal human supervision, which is a multi-stage job.
The traditional works above have motivated us to explicitly infer the candidate answer to transform the answer-agnostic QG into the answer-aware QG. Meanwhile, PLMs with fine-tuning achieved SOTA in many NLP fields, becoming benchmarks hard to bypass. In multi-stage work, researchers will choose different PLMs for different stages in question-answer pair generation, which is effective but heavy. There’s still no end-to-end work to handle the whole task. Therefore, we combine answer extraction and question generation using PLMs and propose an end-to-end model that extracts answers and generates questions in parallel.
2.2
Answer extraction
Information extraction[35, 36] (IE) is basically defined as the task of turning the unstructured information expressed in natural language text into a structured 3-tuple representation as (NE1; R; NE2). Thus, answer extraction can be seen as a sub-field of IE, expecting to pick the most valuable phrase from tuples, regardless of whether it is a named entity, a relation, or their combination: an event. Many IE systems have been proposed for open domains. Yahya et al.[37] describe ReNoun, an open information extraction system that complements previous efforts that rely on big knowledge bases by focusing on nominal attributes and on the long tail. Del Corro and Gemulla[38] proposed ClausIE, a novel, clause-based approach to open information extraction, which extracts relations and their arguments from natural language text. Additionally, some rule-based systems using man-made extraction rules have been proposed, including verb-based[39], semantic role labeling[40], and dependency parse trees[41].
In the era of pre-trained models, auto-encoder[42] models, such as BERT[43] have made great progress in natural language understanding (NLU) tasks. BERT achieves SOTA in the GLUE[44] score which is a multi-task benchmark including named entity recognition. It is a declaration that large PLMs are blossoming in the IE field and will take the place of traditional methods.
3.
Task definition
Answer-agnostic question generation. It aims to generate fluent, meaningful questions Q={q1,q2,⋯,qn} from unlabeled input context C={c1,c2,⋯,cm} without a specific answer. Suppose the length of the question sequence is n while the length of the context sequence is m. During training, this task aims to maximize the conditional probability of Q. All relevant parameters in the model are denoted by θ:
p(Q|C;θ)=n∏t=1p(qt|C,qi<t;θ),
(1)
where the probability of each qt is predicted based on all the words generated previously (i.e., qi<t), and input sentence C.
In our work, we split traditional answer-agnostic question generation into 2 sub-tasks: answer extraction and answer-aware question generation, as in early works.
Answer extraction. It supposes there is at least one question-worthy candidate answer in the input context C={c1,c2,⋯,cm} and then returns its answer A={ai,ai+1,⋯,aj}, where A’s span is limited by C, therefore, 1⩽i⩽j⩽m.
Answer-aware question generation. It is similar to answer-agnostic question generation while it provides an additional answer A={a1,a2,⋯,al}, l is the length of the answer:
p(Q|C,A;θ)=n∏t=1p(qt|C,A,qi<t;θ).
(2)
4.
Model
4.1
Foundation model
We choose BART (bidirectional and auto-regressive transformer) as our foundation model. BART is a sequence-to-sequence model that uses a standard transformer-based encoder-decoder architecture, inheriting its encoder from BERT’s bidirectional encoder and its decoder from GPT’s left-to-right decoder, and is particularly effective for text generation as well as reading comprehension tasks. One limitation of BART is that it cannot simultaneously perform NLU and NLG (natural language generation) tasks. It excels at tasks such as text generation and reading comprehension individually, but integrating these tasks in a single model remains a challenge. However, with its strong foundation, we believe that BART has the potential to be further improved to handle such tasks effectively.
4.2
QGAE
QGAE is a sequence-to-sequence model as shown in Fig. 2 which mainly adopts BART’s architecture while adding an additional encoder, so there are two encoders and a decoder. The model first extracts the phrase with high probability as A and rebuilds input C to A,C. The model will return the rebuild input A,C, and Q.
Figure
2.
The architecture of QGAE consists of two encoders and one decoder, which take raw texts as input and generate question-answer pairs.
Answer extractor encoder is the first encoder inherited from BART similar to BERT and is used to understand the input context and extract the most valuable phrase. We leverage this encoder by appending an extra linear as a classifier to predict the high probability answer span position. Because BART only supports, at most, a pair of sequences as input, we choose the highest score answer of all predictions as the candidate answer. This module will focus on the first task answer extraction (AE).
We select cross entropy to calculate the loss of the AE task. K is the number of classes. In this task, class K is the position of the input paragraph span in the range [0,m−1], and m is the input context length. xi,k indicates that the ith sample is the kth category. p is the probability distribution of annotated data while q is the probability distribution of prediction data:
H(p,q)=−1NN∑i=1K∑k=1p(xi,k)⋅log(q(xi,k)).
(3)
Concretely, we put the specific answer into Eq. (3), and the equation can be changed as:
LAE=ℓ(ˉa,a)=1NN∑i=1Li,
(4)
Li=−K∑k=1ti,k⋅logeˉai,kK∑j=1eˉai,j,
(5)
where a is the labeled answer span as ground-truth, ˉa is the target candidate answer span, and N is the data size. ti,k indicates that the true label of the ith answer is the kth category, which can only take 0 or 1.
4.2.2
Question generation encoder-decoder
Question generation encoder-decoder is mainly derived from BART but adds a unique function leveraging the candidate answer extracted from the first encoder to rebuild input ⟨s⟩C⟨/s⟩ to traditional QG inputs as ⟨s⟩A⟨/s⟩⟨/s⟩C⟨/s⟩. Then, the module uses rebuilt input to generate text as BART does. This module will focus on the second task question generation (QG).
The loss of the QG task is also cross entropy with the only difference being that we use the labeled questions q as ground-truth and prediction questions ˉq, and class K is the vocabulary size of the model:
LQG=ℓ(ˉq,q)=1NN∑i=1Li,
(6)
Li=−K∑k=1ti,k⋅logeˉqi,kK∑j=1eˉqi,j.
(7)
4.2.3
QGAE loss
The QGAE loss is the loss of the multi-task model, in this work, it is the sum of the answer extraction loss and question generation loss:
L=αLAE+(1−α)LQG,
(8)
where α is the weight of the AE task as a hyper-parameter.
5.
Experiments
5.1
Dataset
The Stanford question answering dataset (SQuAD) is the most famous reading comprehension dataset for reversible tasks: question answering and question generation. As the Table 2 shows, it has two versions, SQuAD1.1[45] and SQuAD2.0[46], consisting of questions posed by crowd-workers on a set of Wikipedia articles. Each article has several paragraphs, and each paragraph is asked a set of questions and provided answers, where the answer to every question is a segment of text, or span, from the corresponding reading passage. In SQuAD2.0, because of a percentage of unanswerable questions are added to the dataset, some answers may be null.
Table
2.
Statistics of datasets SQuAD1.1 and SQuAD2.0. No matter in which dataset, an example consists of a context, a question, and an optional answer. The term “negative example” refers to a context passage paired with an unanswerable question, which is intended to help models learn to identify when a question cannot be answered correctly based on the given context.
We implement our models in HuggingFace[47] architecture and fine-tune the model with V100 32 GB GPUs. We first fine-tune BART-base on SQuAD2.0 for 2 epochs to obtain checkpoint BART-base-SQuAD2.0-2 epoch (BbS2). Then we use BbS2 to initialize our QGAE model; more specifically, QGAE’s dual encoder is initialized by the BbS2’s encoder twice and some linear layers that do not exist in BbS2 but in the QGAE will be initialized randomly. We set the batch size to 20, epoch to 3, learning rate to 0.00002, dropout to 0.2, beam search size to 10, max input length to 1024, max question size to 20, and min question size to 3. We perform gradient descent by the Adam optimizer[48]. The coefficient α of task 1 answer extraction is 0.3 while the coefficient of the question generation task is 0.7.
5.3
Evaluation
We report the evaluation results with four metrics: BLEU, METEOR, ROUGE-L, and exact match (EM).
BLEU. BLEU is an algorithm first for evaluating machine-translated text from one natural language to another, later adopted by the text generation task. BLEU compares n-gram words appearing in candidates and references and punishes too-short sentences with a brevity penalty.
ROUGE. ROUGE is a set of metrics including ROUGE-N, ROUGE-L, and ROUGE-W. In this work, we mainly choose ROUGE-L, which is the longest common sub-sequence (LCS)-based statistic. LCS takes into account sentence-level structure similarity naturally and identifies the longest co-occurring in sequence n-grams automatically.
METEOR. METEOR is also a metric based on the harmonic mean of unigram precision and recall, with recall weighted higher than precision.
Exact match. Exact match measures the percentage of predictions that match any one of the ground truth answers exactly.
As each paragraph in the SQuAD dataset may have several question-answer pairs, we use paragraphs as input and compare outputs with a group of question-answer pairs and choose the highest score with BLEU-4 as the main indicator.
6.
Results and discussion
6.1
Results
In Table 3, we compare our proposed end-to-end QGAE model with 3 other types of earlier works: standalone answer extraction task, standalone answer-agnostic question generation, and multi-stage QA-pair generation pipeline. All the data used in the experiments have been replicated from the following paper.
Table
3.
Comparison of method performance in major metrics (including QG metrics and AE metric) on the SQuAD dataset. These methods are divided into four types according to their primary research fields. The first two classifications focus on their own independent fields, while the latter two classifications can accomplish these two tasks at the same time.
Attention LSTM. Attention LSTM was proposed by Du et al.[4] and was the first work to focus on answer-agnostic QG.
Self-attention transformers. Self-attention transformers[50] explore how transformers can be adapted to the task of neural question generation without constraining the model to focus on a specific answer passage.
Question-driven LSTM. Question-driven LSTM[26] proposed two new strategies question type prediction and a copy loss mechanism to address the task.
(Ⅲ) Multi-stage QA-pair generation pipeline
MCF. Wang et al.[51] proposed a multi-stage framework that can extract question-worthy phrases and improve the performance of question generation. We chose this framework as the baseline for the specific task of generating QA pairs and used it to evaluate the performance.
6.2
Discussion
The performance shows that our end-to-end QGAE model not only achieves SOTA in the answer extraction task but also makes a great improvement in the answer-agnostic question generation compared with the traditional encoder-decoder architecture. Even if multi-stage work MCF has a much more complex workflow, has a weaker comprehensive performance than our work. What is more? QGAE is lighter, more convenient, and more portable since it only requires fine-tuning of one pre-trained model, whereas multi-stage methods need at least two models for stage AE and QG.
Although great progress has been made in the EM score, reaching 53.82%, there is still much room for improvement in extraction accuracy. Our model may extract candidate answers that are not ground truth but also meaningful, while extraction accuracy is judged and limited by the labeled data. Specifically, the range of candidate answers is very wide, ranging from named entities to relationships, to events. However, only a small percentage of key phrases are included in the training dataset while others are out of range. Candidate answers beyond the confines of the dataset may make the later question generation task in the wrong direction, performing worse when choosing traditional machine-translation evaluation indicators. Despite all this, prediction sentences not in the ground truth are still valuable and reasonable. The high diversity of generated sentences, to a certain extent, is an advantage that will make our model competitive in different scenes for data augmentation.
Therefore it can be concluded that we have expanded our model’s function not only to generate questions but also to generate QA-pairs compared to the baseline model and better than any previous work, which proved our model is diverse and efficient.
7.
Conclusions
In this paper, our focus is on answer-agnostic question generation, which can be extended to question-answer pair generation. This task can be divided into two sub-tasks: answer extraction and question generation. We proposed an end-to-end model called question generation with answer extractor (QGAE) using raw text without costing any additional information, which can generate question-answer pairs in parallel. Compared to the multi-stage question-answer generation method, QGAE has several advantages. First, QGAE is able to generate question-answer pairs in parallel, whereas the multi-stage method requires multiple rounds of generation and refinement. Second, it is lighter, more convenient, and more portable than multi-stage methods in training, which reduces the complexity of the overall system. Third, our model achieves a better average score and greater diversity. Overall, QGAE is a more efficient and versatile approach to answer-agnostic question generation, with potential applications in various natural language processing tasks.
In further work, we will try to compile more datasets into one ensemble to improve the accuracy of answer extraction. Not only that, we will try to change our main task to information retrieval to optimize our answer extraction, as different weight biases in sub-tasks lead to an imbalance in the model’s focus in the two sub-tasks. All in all, this is still pioneering work in pre-trained language models adapting question-answer pair generation.
Conflict of interest
The author declares that he has no conflict of interest.
Using the synchronous coordinates, the creation of a Schwarzschild black hole immersed in a de Sitter spacetime can be viewed as a coherent creation of a collection of timelike geodesics.
The previously supposed conical singularities do not exist at the horizons of the constrained instanton. Instead, the unavoidable irregularity is presented as a non-vanishing second fundamental form elsewhere at the quantum transition 3-surface.
Hawking S W. The boundary conditions of the universe. In: Bruck H A, Coyne G V, Longair M S, et al., editors. Astrophysical Cosmology. Proceedings on Cosmology and Fundamental Physics. Vatican: The Pontifical Academy of Sciences, 1982 : 563–574.
[2]
Hartle J B, Hawking S W. Wave function of the universe. Physical Review D, 1983, 28: 2960. DOI: 10.1103/physrevd.28.2960
[3]
Gibbons G W, Hawking S W. Cosmological event horizons, thermodynamics, and particle creation. Physical Review D, 1977, 15: 2738. DOI: 10.1103/physrevd.15.2738
[4]
Gibbons G W, Hawking S W. Action integrals and partition functions in quantum gravity. Physical Review D, 1977, 15: 2752. DOI: 10.1103/physrevd.15.2752
[5]
Wu Z C. Entropy of a black hole with distinct surface gravities. General Relativity and Gravitation, 2000, 32: 1823–1833. DOI: 10.1023/A:1001984731021
[6]
Bousso R, Hawking S W. Probability for primordial black holes. Physical Review D, 1995, 52: 5659. DOI: 10.1103/physrevd.52.5659
[7]
Wu Z C. Quantum creation of a black hole. International Journal of Modern Physics D, 1997, 6: 199–210. DOI: 10.1142/s0218271897000121
[8]
Bousso R, Hawking S W. Lorentzian condition in quantum gravity. Physical Review D, 1999, 59: 103501. DOI: 10.1103/physrevd.59.103501
[9]
Wu Z C. Creation of closed or open universe from constrained instanton. General Relativity and Gravitation, 1998, 30: 1639–1643. DOI: 10.1023/A:1026668322706
[10]
Misner C W, Thorne K S, Wheeler J A. Gravitation. Princeton, USA: Princeton University Press, 1973 .
[11]
Teitelboim C. Action and entropy of extreme and nonextreme black holes. Physical Review D, 1995, 51: 4315. DOI: 10.1103/physrevd.51.4315
[12]
Wu Z C. Wave function of a black hole. In: Ruffini R, editor. Proceeding of the Fourth Marcel Grossman Meeting on General Relativity. Amsterdam: North Holland, 1986 : 1203–1210.
[13]
Hawking S W, Ross S F. Duality between electric and magnetic black holes. Physical Review D, 1995, 52: 5865. DOI: 10.1103/physrevd.52.5865
[14]
Mann R B, Ross S F. Cosmological production of charged black hole pairs. Physical Review D, 1995, 52: 2254. DOI: 10.1103/physrevd.52.2254
[15]
Dias Ó J C, Lemos J P S. Pair creation of higher dimensional black holes on a de Sitter background. Physical Review D, 2004, 70: 124023. DOI: 10.1103/physrevd.70.124023
[16]
Myers R C. Myers-Perry black holes. In: Horowitz G T, editor. Black Holes in Higher Dimensions. Cambridge, UK: Cambridge University Press, 2012: 101–133.
Anusree, V.C., Ptok, A., Sobieszczyk, P. et al. Correlation stabilized ferromagnetic MnRuAs with distorted kagome lattice. Physical Review B, 2025, 111(3): 035125.
DOI:10.1103/PhysRevB.111.035125
2.
Chen, Z., Wu, X., Zhou, S. et al. Discovery of a long-ranged charge order with 1/4 Ge1-dimerization in an antiferromagnetic Kagome metal. Nature Communications, 2024, 15(1): 6262.
DOI:10.1038/s41467-024-50661-x
3.
Wang, Y.. Encoding innumerable charge density waves of FeGe into polymorphs of LiFe6Ge6. Science China: Physics, Mechanics and Astronomy, 2024, 67(9): 297011.
DOI:10.1007/s11433-024-2423-2
4.
Wu, Q., Quan, W., Pan, S. et al. Atomically Thin Kagome-Structured Co9Te16 Achieved through Self-Intercalation and Its Flat Band Visualization. Nano Letters, 2024, 24(25): 7672-7680.
DOI:10.1021/acs.nanolett.4c01526
5.
Wang, Y.. Enhanced spin-polarization via partial Ge-dimerization as the driving force of the charge density wave in FeGe. Physical Review Materials, 2023, 7(10): 104006.
DOI:10.1103/PhysRevMaterials.7.104006
Hawking S W. The boundary conditions of the universe. In: Bruck H A, Coyne G V, Longair M S, et al., editors. Astrophysical Cosmology. Proceedings on Cosmology and Fundamental Physics. Vatican: The Pontifical Academy of Sciences, 1982 : 563–574.
[2]
Hartle J B, Hawking S W. Wave function of the universe. Physical Review D, 1983, 28: 2960. DOI: 10.1103/physrevd.28.2960
[3]
Gibbons G W, Hawking S W. Cosmological event horizons, thermodynamics, and particle creation. Physical Review D, 1977, 15: 2738. DOI: 10.1103/physrevd.15.2738
[4]
Gibbons G W, Hawking S W. Action integrals and partition functions in quantum gravity. Physical Review D, 1977, 15: 2752. DOI: 10.1103/physrevd.15.2752
[5]
Wu Z C. Entropy of a black hole with distinct surface gravities. General Relativity and Gravitation, 2000, 32: 1823–1833. DOI: 10.1023/A:1001984731021
[6]
Bousso R, Hawking S W. Probability for primordial black holes. Physical Review D, 1995, 52: 5659. DOI: 10.1103/physrevd.52.5659
[7]
Wu Z C. Quantum creation of a black hole. International Journal of Modern Physics D, 1997, 6: 199–210. DOI: 10.1142/s0218271897000121
[8]
Bousso R, Hawking S W. Lorentzian condition in quantum gravity. Physical Review D, 1999, 59: 103501. DOI: 10.1103/physrevd.59.103501
[9]
Wu Z C. Creation of closed or open universe from constrained instanton. General Relativity and Gravitation, 1998, 30: 1639–1643. DOI: 10.1023/A:1026668322706
[10]
Misner C W, Thorne K S, Wheeler J A. Gravitation. Princeton, USA: Princeton University Press, 1973 .
[11]
Teitelboim C. Action and entropy of extreme and nonextreme black holes. Physical Review D, 1995, 51: 4315. DOI: 10.1103/physrevd.51.4315
[12]
Wu Z C. Wave function of a black hole. In: Ruffini R, editor. Proceeding of the Fourth Marcel Grossman Meeting on General Relativity. Amsterdam: North Holland, 1986 : 1203–1210.
[13]
Hawking S W, Ross S F. Duality between electric and magnetic black holes. Physical Review D, 1995, 52: 5865. DOI: 10.1103/physrevd.52.5865
[14]
Mann R B, Ross S F. Cosmological production of charged black hole pairs. Physical Review D, 1995, 52: 2254. DOI: 10.1103/physrevd.52.2254
[15]
Dias Ó J C, Lemos J P S. Pair creation of higher dimensional black holes on a de Sitter background. Physical Review D, 2004, 70: 124023. DOI: 10.1103/physrevd.70.124023
[16]
Myers R C. Myers-Perry black holes. In: Horowitz G T, editor. Black Holes in Higher Dimensions. Cambridge, UK: Cambridge University Press, 2012: 101–133.
Anusree, V.C., Ptok, A., Sobieszczyk, P. et al. Correlation stabilized ferromagnetic MnRuAs with distorted kagome lattice. Physical Review B, 2025, 111(3): 035125.
DOI:10.1103/PhysRevB.111.035125
2.
Chen, Z., Wu, X., Zhou, S. et al. Discovery of a long-ranged charge order with 1/4 Ge1-dimerization in an antiferromagnetic Kagome metal. Nature Communications, 2024, 15(1): 6262.
DOI:10.1038/s41467-024-50661-x
3.
Wang, Y.. Encoding innumerable charge density waves of FeGe into polymorphs of LiFe6Ge6. Science China: Physics, Mechanics and Astronomy, 2024, 67(9): 297011.
DOI:10.1007/s11433-024-2423-2
4.
Wu, Q., Quan, W., Pan, S. et al. Atomically Thin Kagome-Structured Co9Te16 Achieved through Self-Intercalation and Its Flat Band Visualization. Nano Letters, 2024, 24(25): 7672-7680.
DOI:10.1021/acs.nanolett.4c01526
5.
Wang, Y.. Enhanced spin-polarization via partial Ge-dimerization as the driving force of the charge density wave in FeGe. Physical Review Materials, 2023, 7(10): 104006.
DOI:10.1103/PhysRevMaterials.7.104006
Table
1.
A case of QA-pairs generated by our QGAE model: the model accepts unannotated texts as input, extracts the highlighted phrase “Lorentz’s law” as an answer, then uses this answer to make question generation.
Input context: Through combining the definition of electric current as the time rate of change of electric charge, a rule of vector multiplication called
Lorentz’s law describes the force on a charge moving in a magnetic field. The connection between electricity and magnetism allows for the description of a unified electromagnetic force that acts on a charge. This force can be written as a sum of the electrostatic force (due to the electric field) and the magnetic force (due to the magnetic field).
Extracted answer: Lorentz’s law
Generated question: What describes the force on a charge moving in a magnetic field?
Table
2.
Statistics of datasets SQuAD1.1 and SQuAD2.0. No matter in which dataset, an example consists of a context, a question, and an optional answer. The term “negative example” refers to a context passage paired with an unanswerable question, which is intended to help models learn to identify when a question cannot be answered correctly based on the given context.
Table
3.
Comparison of method performance in major metrics (including QG metrics and AE metric) on the SQuAD dataset. These methods are divided into four types according to their primary research fields. The first two classifications focus on their own independent fields, while the latter two classifications can accomplish these two tasks at the same time.