Xing-Yu Bao received his master degree in Physics from the University of Science and Technology of China in 2023. His research interests mainly focus on ion trap and quantum computing
Jin-Ming Cui works as an Associate Researcher in the University of Science and Technology of China. He got his B.S. degree and Ph.D. degree from USTC in 2008 and 2013, respectively. During his Ph.D.’s study, he focused on experiment works with optical micro-cavities and NV centers. Now his major research focuses on quantum information based on trapped ion system, including quantum simulation, quantum computation, quantum network, and fiber Fabry-Perot microcavity
The main purpose of this paper is to design a novel coupled system of an ion trap and a fiber cavity. This integrated solution is achieved by fabricating a fiber cavity with a metal mask on the side and end faces of the fiber. The fiber cavity with the metal mask can transmit light and electric charges, and the metal mask on the fiber end-face can shield electric charges on the dielectric high-reflection film. This system is designed to trap a single 138Ba+ ion and realize coupling of the fiber cavity to the fluorescence at a 493 nm wavelength of 138Ba+. To efficiently collect fluorescent photons, we perform a theoretical analysis of the overall system to achieve optimal coupling of each individual part. The cavity length is designed to be 250 μm, and the optimized coupling parameters are (g,κ,γ)/2π=(55,105,20) MHz. We also improve the stability and reliability of the system by analyzing the vibration, performance of the ion trap, and thermal stability. The core of the system is composed of materials with similar thermal expansion coefficients to improve thermal stability. The system uses spring connections to isolate vibrations inside and outside the vacuum chamber. We theoretically solve the difficulties of manufacturing the coupled system and have completed the experimental verification of some key technologies. The whole system is expected to be extended into a complex quantum network system to realize quantum computation and communication.
Graphical Abstract
Scheme and parameters of the designed fiber cavity ion trap.
Abstract
The main purpose of this paper is to design a novel coupled system of an ion trap and a fiber cavity. This integrated solution is achieved by fabricating a fiber cavity with a metal mask on the side and end faces of the fiber. The fiber cavity with the metal mask can transmit light and electric charges, and the metal mask on the fiber end-face can shield electric charges on the dielectric high-reflection film. This system is designed to trap a single 138Ba+ ion and realize coupling of the fiber cavity to the fluorescence at a 493 nm wavelength of 138Ba+. To efficiently collect fluorescent photons, we perform a theoretical analysis of the overall system to achieve optimal coupling of each individual part. The cavity length is designed to be 250 μm, and the optimized coupling parameters are (g,κ,γ)/2π=(55,105,20) MHz. We also improve the stability and reliability of the system by analyzing the vibration, performance of the ion trap, and thermal stability. The core of the system is composed of materials with similar thermal expansion coefficients to improve thermal stability. The system uses spring connections to isolate vibrations inside and outside the vacuum chamber. We theoretically solve the difficulties of manufacturing the coupled system and have completed the experimental verification of some key technologies. The whole system is expected to be extended into a complex quantum network system to realize quantum computation and communication.
Public Summary
This work designs a high-efficiency and high-photon-rate quantum node based on a cold trapped ion.
The optimized small-mode-volume fiber cavity serves as a resonant cavity to efficiently collect ion fluorescence photons.
The end face of the fiber is coated with a metal mask, which can effectively reduce the charging effect of the fiber, contributing to stable ion trapping in the cavity mode.
Question generation[1, 2] (QG) is defined as the task of generating fluent, meaningful questions automatically from texts with optional answers, so it can be mainly divided into two streams: answer-aware QG[3] that requires answers, and answer-agnostic QG[4] that does not. QG is the reverse task of question answering (QA), which is a long-standing and valuable task helping computers achieve machine reading comprehension[5, 6], dating back to the 1960s[7]. As with many other supervised learning[8, 9] tasks, QA will also encounter the lack of annotated data in spite of the fact that annotated data sometimes make the most essential part of the whole work.
QG is a popular choice for data augmentation for QA to alleviate insufficient labeled data. With the continuous development of Internet technology, it is becoming increasingly easier to obtain valuable data from the Internet. However, question-answer pairs (as shown in Table 1) are still such expensive corpora that typically require manual annotation by crowdsourcing before being used for supervised learning on QA and QG tasks. To alleviate the high-cost problem of generating question-answer pairs, it is natural to consider answer-agnostic QG, since its only input is raw text.
Table
1.
A case of QA-pairs generated by our QGAE model: the model accepts unannotated texts as input, extracts the highlighted phrase “Lorentz’s law” as an answer, then uses this answer to make question generation.
Input context: Through combining the definition of electric current as the time rate of change of electric charge, a rule of vector multiplication called
Lorentz’s law describes the force on a charge moving in a magnetic field. The connection between electricity and magnetism allows for the description of a unified electromagnetic force that acts on a charge. This force can be written as a sum of the electrostatic force (due to the electric field) and the magnetic force (due to the magnetic field).
Extracted answer: Lorentz’s law
Generated question: What describes the force on a charge moving in a magnetic field?
Although labeled answers are not necessary, answer-agnostic QG is still facing a great challenge. Most previous works focused on providing additional information to their models by leveraging named entity recognition (NER)[10] to obtain extra linguistic features, adding answer position features[11], using knowledge graphs[12], and some other methods to improve the generation effect. These methods effectively improve the fluency and accuracy of generated texts, but answer-agnostic QG still performs worse than answer-aware QG. Thus, answer-aware QG may play an irreplaceable role, and changing answer-agnostic QG to answer-aware QG is a good choice. Apart from this, there is still an obstacle in generating question-answer pairs that answer-agnostic QG can’t generate answers. To address this issue, researchers often add an additional measure for question-answer pair generation: answer extraction. Compared with generating an answer, extracting an exact span in the context is much simpler.
Explicitly extracting candidate answers will not only resolve the demand for the lack of answers but also can transform answer-agnostic QG into answer-aware QG. As shown in Fig. 1, some works such as RGF[13] (retrieve-generate filter) proposed a multi-stage pipeline method to handle the problem. A multi-stage pipeline method is often designed in complexity, including several parts, and each part may need different inputs. Some early RNN-based[14–17] works optimized pipeline methods in an end-to-end way, which makes the overall structure lighter and faster. Though pre-trained language models (PLMs) have occupied dominance in both natural language generation and understanding, there is still no end-to-end work using pre-trained models to generate question-answer pairs. We are sure there is enough potential for PLMs to achieve the task.
Figure
1.
The difference between multi-stage methods and end-to-end models is that a multi-stage method usually has more than one model in the whole workflow. In every stage, a multi-stage method may need to deal with different inputs and outputs, while on the contrary, an end-to-end model only needs a definite kind of input.
In this study, we are motivated by the weak performance of answer-agnostic QG compared to answer-aware QG, inspired by the combination of QG and AE tasks, trying to propose an answer-agnostic question generation model called question generation with answer extractor (QGAE) to alleviate the high demand for large-scale QA pairs. QGAE is a multi-task model that requires only raw texts as input and can achieve the dual tasks: answer extraction and question generation. We design our model based on the PLM model BART[18], which has dual encoders and a decoder to generate questions and extract answers in parallel. In our study, question generation is the main task, which is the most challenging part similar to all other generation tasks for generated texts’ high syntactic diversity and semantic substitutability, so we pay more attention and assign a higher weight to the corresponding module. Therefore answer extraction is considered an auxiliary task. The design not only makes it feasible to turn answer-agnostic question generation into answer-aware question generation but also enables the model to be considered capable of generating question-answer pairs. The contributions of this paper are summarized as follows:
● We are the first to propose a new end-to-end model using PLMs, which is called QGAE for answer-agnostic question generation.
● The QGAE model generates question-answer pairs from unannotated texts without requiring any additional information.
● Our model achieves state-of-the-art performance in generating high-quality question-answer pairs, outperforming existing methods by a significant margin.
The rest of this paper is organized as follows. In Section 2, we review the related works of question generation and answer extraction. In Section 3, we formulate the QG task and AE task. In Section 4, we describe each module of our QGAE model. In Section 5, we introduce our experiment in detail. In the last Section 6, we conclude this work and give a detailed analysis.
2.
Related works
2.1
Question generation
The QG field was devoted great interest by researchers for its great potential benefits; therefore, it has made great progress in application scenarios such as data augmentation[19], chatbots[20], machine reading comprehension[21], and intelligent tutors[22].
In the neural model age, Du et al.[4] proposed the first neural QG model focused on answer-agnostic QG. They investigated the effect of encoding sentence- vs. paragraph-level information by using an attention-based model and found that as the size of the input text increased, the evaluation score of the output decreased. To deal with the rare or unknown word problem, Gulcehre et al.[23] proposed a copy mechanism that was first used in the neural machine translation[24] to solve the out-of-vocabulary problem. This mechanism was absorbed in the QG task and widely used. Following the old experience of rule-based QG[25], Wu et al.[26] suggested two new strategies to deal with this task: question type prediction and a copy loss mechanism. Du et al.[15] combined answer extraction and question generation in an LSTM[27] model including answer feature embedding, denoting answer span with the usual BIO tagging scheme[28].
In the transformer-based[29] PLM era, compared to auto-encoder models, auto-regressive[30] models are widely picked as baselines for the QG task. Laban et al.[20] fine-tuned a GPT2[31] as the base part of a question-driven news chatbot. Wang et al.[32] leveraged BART to propose QAGS (question answering and generation for summarization) to evaluate automatic summarization. Bhambhoria et al.[33] leveraged T5[34] to generate QA pairs for COVID-19 literature. Paranjape et al.[13] developed a retrieve-generate filter (RGF) technique to create counterfactual evaluation and training data with minimal human supervision, which is a multi-stage job.
The traditional works above have motivated us to explicitly infer the candidate answer to transform the answer-agnostic QG into the answer-aware QG. Meanwhile, PLMs with fine-tuning achieved SOTA in many NLP fields, becoming benchmarks hard to bypass. In multi-stage work, researchers will choose different PLMs for different stages in question-answer pair generation, which is effective but heavy. There’s still no end-to-end work to handle the whole task. Therefore, we combine answer extraction and question generation using PLMs and propose an end-to-end model that extracts answers and generates questions in parallel.
2.2
Answer extraction
Information extraction[35, 36] (IE) is basically defined as the task of turning the unstructured information expressed in natural language text into a structured 3-tuple representation as (NE1; R; NE2). Thus, answer extraction can be seen as a sub-field of IE, expecting to pick the most valuable phrase from tuples, regardless of whether it is a named entity, a relation, or their combination: an event. Many IE systems have been proposed for open domains. Yahya et al.[37] describe ReNoun, an open information extraction system that complements previous efforts that rely on big knowledge bases by focusing on nominal attributes and on the long tail. Del Corro and Gemulla[38] proposed ClausIE, a novel, clause-based approach to open information extraction, which extracts relations and their arguments from natural language text. Additionally, some rule-based systems using man-made extraction rules have been proposed, including verb-based[39], semantic role labeling[40], and dependency parse trees[41].
In the era of pre-trained models, auto-encoder[42] models, such as BERT[43] have made great progress in natural language understanding (NLU) tasks. BERT achieves SOTA in the GLUE[44] score which is a multi-task benchmark including named entity recognition. It is a declaration that large PLMs are blossoming in the IE field and will take the place of traditional methods.
3.
Task definition
Answer-agnostic question generation. It aims to generate fluent, meaningful questions Q={q1,q2,⋯,qn} from unlabeled input context C={c1,c2,⋯,cm} without a specific answer. Suppose the length of the question sequence is n while the length of the context sequence is m. During training, this task aims to maximize the conditional probability of Q. All relevant parameters in the model are denoted by θ:
p(Q|C;θ)=n∏t=1p(qt|C,qi<t;θ),
(1)
where the probability of each qt is predicted based on all the words generated previously (i.e., qi<t), and input sentence C.
In our work, we split traditional answer-agnostic question generation into 2 sub-tasks: answer extraction and answer-aware question generation, as in early works.
Answer extraction. It supposes there is at least one question-worthy candidate answer in the input context C={c1,c2,⋯,cm} and then returns its answer A={ai,ai+1,⋯,aj}, where A’s span is limited by C, therefore, 1⩽i⩽j⩽m.
Answer-aware question generation. It is similar to answer-agnostic question generation while it provides an additional answer A={a1,a2,⋯,al}, l is the length of the answer:
p(Q|C,A;θ)=n∏t=1p(qt|C,A,qi<t;θ).
(2)
4.
Model
4.1
Foundation model
We choose BART (bidirectional and auto-regressive transformer) as our foundation model. BART is a sequence-to-sequence model that uses a standard transformer-based encoder-decoder architecture, inheriting its encoder from BERT’s bidirectional encoder and its decoder from GPT’s left-to-right decoder, and is particularly effective for text generation as well as reading comprehension tasks. One limitation of BART is that it cannot simultaneously perform NLU and NLG (natural language generation) tasks. It excels at tasks such as text generation and reading comprehension individually, but integrating these tasks in a single model remains a challenge. However, with its strong foundation, we believe that BART has the potential to be further improved to handle such tasks effectively.
4.2
QGAE
QGAE is a sequence-to-sequence model as shown in Fig. 2 which mainly adopts BART’s architecture while adding an additional encoder, so there are two encoders and a decoder. The model first extracts the phrase with high probability as A and rebuilds input C to A,C. The model will return the rebuild input A,C, and Q.
Figure
2.
The architecture of QGAE consists of two encoders and one decoder, which take raw texts as input and generate question-answer pairs.
Answer extractor encoder is the first encoder inherited from BART similar to BERT and is used to understand the input context and extract the most valuable phrase. We leverage this encoder by appending an extra linear as a classifier to predict the high probability answer span position. Because BART only supports, at most, a pair of sequences as input, we choose the highest score answer of all predictions as the candidate answer. This module will focus on the first task answer extraction (AE).
We select cross entropy to calculate the loss of the AE task. K is the number of classes. In this task, class K is the position of the input paragraph span in the range [0,m−1], and m is the input context length. xi,k indicates that the ith sample is the kth category. p is the probability distribution of annotated data while q is the probability distribution of prediction data:
H(p,q)=−1NN∑i=1K∑k=1p(xi,k)⋅log(q(xi,k)).
(3)
Concretely, we put the specific answer into Eq. (3), and the equation can be changed as:
LAE=ℓ(ˉa,a)=1NN∑i=1Li,
(4)
Li=−K∑k=1ti,k⋅logeˉai,kK∑j=1eˉai,j,
(5)
where a is the labeled answer span as ground-truth, ˉa is the target candidate answer span, and N is the data size. ti,k indicates that the true label of the ith answer is the kth category, which can only take 0 or 1.
4.2.2
Question generation encoder-decoder
Question generation encoder-decoder is mainly derived from BART but adds a unique function leveraging the candidate answer extracted from the first encoder to rebuild input ⟨s⟩C⟨/s⟩ to traditional QG inputs as ⟨s⟩A⟨/s⟩⟨/s⟩C⟨/s⟩. Then, the module uses rebuilt input to generate text as BART does. This module will focus on the second task question generation (QG).
The loss of the QG task is also cross entropy with the only difference being that we use the labeled questions q as ground-truth and prediction questions ˉq, and class K is the vocabulary size of the model:
LQG=ℓ(ˉq,q)=1NN∑i=1Li,
(6)
Li=−K∑k=1ti,k⋅logeˉqi,kK∑j=1eˉqi,j.
(7)
4.2.3
QGAE loss
The QGAE loss is the loss of the multi-task model, in this work, it is the sum of the answer extraction loss and question generation loss:
L=αLAE+(1−α)LQG,
(8)
where α is the weight of the AE task as a hyper-parameter.
5.
Experiments
5.1
Dataset
The Stanford question answering dataset (SQuAD) is the most famous reading comprehension dataset for reversible tasks: question answering and question generation. As the Table 2 shows, it has two versions, SQuAD1.1[45] and SQuAD2.0[46], consisting of questions posed by crowd-workers on a set of Wikipedia articles. Each article has several paragraphs, and each paragraph is asked a set of questions and provided answers, where the answer to every question is a segment of text, or span, from the corresponding reading passage. In SQuAD2.0, because of a percentage of unanswerable questions are added to the dataset, some answers may be null.
Table
2.
Statistics of datasets SQuAD1.1 and SQuAD2.0. No matter in which dataset, an example consists of a context, a question, and an optional answer. The term “negative example” refers to a context passage paired with an unanswerable question, which is intended to help models learn to identify when a question cannot be answered correctly based on the given context.
We implement our models in HuggingFace[47] architecture and fine-tune the model with V100 32 GB GPUs. We first fine-tune BART-base on SQuAD2.0 for 2 epochs to obtain checkpoint BART-base-SQuAD2.0-2 epoch (BbS2). Then we use BbS2 to initialize our QGAE model; more specifically, QGAE’s dual encoder is initialized by the BbS2’s encoder twice and some linear layers that do not exist in BbS2 but in the QGAE will be initialized randomly. We set the batch size to 20, epoch to 3, learning rate to 0.00002, dropout to 0.2, beam search size to 10, max input length to 1024, max question size to 20, and min question size to 3. We perform gradient descent by the Adam optimizer[48]. The coefficient α of task 1 answer extraction is 0.3 while the coefficient of the question generation task is 0.7.
5.3
Evaluation
We report the evaluation results with four metrics: BLEU, METEOR, ROUGE-L, and exact match (EM).
BLEU. BLEU is an algorithm first for evaluating machine-translated text from one natural language to another, later adopted by the text generation task. BLEU compares n-gram words appearing in candidates and references and punishes too-short sentences with a brevity penalty.
ROUGE. ROUGE is a set of metrics including ROUGE-N, ROUGE-L, and ROUGE-W. In this work, we mainly choose ROUGE-L, which is the longest common sub-sequence (LCS)-based statistic. LCS takes into account sentence-level structure similarity naturally and identifies the longest co-occurring in sequence n-grams automatically.
METEOR. METEOR is also a metric based on the harmonic mean of unigram precision and recall, with recall weighted higher than precision.
Exact match. Exact match measures the percentage of predictions that match any one of the ground truth answers exactly.
As each paragraph in the SQuAD dataset may have several question-answer pairs, we use paragraphs as input and compare outputs with a group of question-answer pairs and choose the highest score with BLEU-4 as the main indicator.
6.
Results and discussion
6.1
Results
In Table 3, we compare our proposed end-to-end QGAE model with 3 other types of earlier works: standalone answer extraction task, standalone answer-agnostic question generation, and multi-stage QA-pair generation pipeline. All the data used in the experiments have been replicated from the following paper.
Table
3.
Comparison of method performance in major metrics (including QG metrics and AE metric) on the SQuAD dataset. These methods are divided into four types according to their primary research fields. The first two classifications focus on their own independent fields, while the latter two classifications can accomplish these two tasks at the same time.
Attention LSTM. Attention LSTM was proposed by Du et al.[4] and was the first work to focus on answer-agnostic QG.
Self-attention transformers. Self-attention transformers[50] explore how transformers can be adapted to the task of neural question generation without constraining the model to focus on a specific answer passage.
Question-driven LSTM. Question-driven LSTM[26] proposed two new strategies question type prediction and a copy loss mechanism to address the task.
(Ⅲ) Multi-stage QA-pair generation pipeline
MCF. Wang et al.[51] proposed a multi-stage framework that can extract question-worthy phrases and improve the performance of question generation. We chose this framework as the baseline for the specific task of generating QA pairs and used it to evaluate the performance.
6.2
Discussion
The performance shows that our end-to-end QGAE model not only achieves SOTA in the answer extraction task but also makes a great improvement in the answer-agnostic question generation compared with the traditional encoder-decoder architecture. Even if multi-stage work MCF has a much more complex workflow, has a weaker comprehensive performance than our work. What is more? QGAE is lighter, more convenient, and more portable since it only requires fine-tuning of one pre-trained model, whereas multi-stage methods need at least two models for stage AE and QG.
Although great progress has been made in the EM score, reaching 53.82%, there is still much room for improvement in extraction accuracy. Our model may extract candidate answers that are not ground truth but also meaningful, while extraction accuracy is judged and limited by the labeled data. Specifically, the range of candidate answers is very wide, ranging from named entities to relationships, to events. However, only a small percentage of key phrases are included in the training dataset while others are out of range. Candidate answers beyond the confines of the dataset may make the later question generation task in the wrong direction, performing worse when choosing traditional machine-translation evaluation indicators. Despite all this, prediction sentences not in the ground truth are still valuable and reasonable. The high diversity of generated sentences, to a certain extent, is an advantage that will make our model competitive in different scenes for data augmentation.
Therefore it can be concluded that we have expanded our model’s function not only to generate questions but also to generate QA-pairs compared to the baseline model and better than any previous work, which proved our model is diverse and efficient.
7.
Conclusions
In this paper, our focus is on answer-agnostic question generation, which can be extended to question-answer pair generation. This task can be divided into two sub-tasks: answer extraction and question generation. We proposed an end-to-end model called question generation with answer extractor (QGAE) using raw text without costing any additional information, which can generate question-answer pairs in parallel. Compared to the multi-stage question-answer generation method, QGAE has several advantages. First, QGAE is able to generate question-answer pairs in parallel, whereas the multi-stage method requires multiple rounds of generation and refinement. Second, it is lighter, more convenient, and more portable than multi-stage methods in training, which reduces the complexity of the overall system. Third, our model achieves a better average score and greater diversity. Overall, QGAE is a more efficient and versatile approach to answer-agnostic question generation, with potential applications in various natural language processing tasks.
In further work, we will try to compile more datasets into one ensemble to improve the accuracy of answer extraction. Not only that, we will try to change our main task to information retrieval to optimize our answer extraction, as different weight biases in sub-tasks lead to an imbalance in the model’s focus in the two sub-tasks. All in all, this is still pioneering work in pre-trained language models adapting question-answer pair generation.
Acknowledgements
We thank the USTC Center for Micro and Nanoscale Research and Fabrication for the technical suggestions on the fabrication process of the metal mask on the fiber facet. This work was supported by the National Natural Science Foundation of China (11821404, 11804330), the Key Research Program of Frontier Sciences, CAS (QYZDY-SSW-SLH003), the Science Foundation of the CAS (ZDRW-XH2019-1), the Fundamental Research Funds for the Central Universities (WK2470000026, WK2470000027, WK2470000028, WK2470000038), the Anhui Initiative in Quantum Information Technologies (AHY020100), and the National Program for Support of Topnotch Young Professionals (BB2470000005).
Conflict of interest
The authors declare that they have no conflict of interest.
This work designs a high-efficiency and high-photon-rate quantum node based on a cold trapped ion.
The optimized small-mode-volume fiber cavity serves as a resonant cavity to efficiently collect ion fluorescence photons.
The end face of the fiber is coated with a metal mask, which can effectively reduce the charging effect of the fiber, contributing to stable ion trapping in the cavity mode.
Duan L M, Monroe C. Colloquium: Quantum networks with trapped ions. Reviews of Modern Physics,2010, 82: 1209–1224. DOI: 10.1103/RevModPhys.82.1209
[2]
Ritter S, Nölleke C, Hahn C, et al. An elementary quantum network of single atoms in optical cavities. Nature,2012, 484: 195–200. DOI: 10.1038/nature11023
[3]
Reiserer A, Kalb N, Rempe G, et al. A quantum gate between a flying optical photon and a single trapped atom. Nature,2014, 508: 237–240. DOI: 10.1038/nature13177
[4]
Uphoff M, Brekenfeld M, Rempe G, et al. An integrated quantum repeater at telecom wavelength with single atoms in optical fiber cavities. Applied Physics B,2016, 122: 46. DOI: https://doi.org/10.1007/s00340-015-6299-2
Paul W. Electromagnetic traps for charged and neutral particles. Reviews of Modern Physics,1990, 62: 531–540. DOI: 10.1103/RevModPhys.62.531
[7]
Wang C X, He R, Li R R, et al. Advances in the study of ion trap structures in quantum computation and simulation. Acta Physica Sinica,2022, 71: 133701. DOI: 10.7498/aps.71.20220224
[8]
Prestage J D, Dick G J, Maleki L. New ion trap for frequency standard applications. Journal of Applied Physics,1989, 66: 1013–1017. DOI: 10.1063/1.343486
[9]
Schmidt-Kaler F, Häffner H, Gulde S, et al. How to realize a universal quantum gate with trapped ions. Applied Physics B,2003, 77: 789–796. DOI: https://doi.org/10.1007/s00340-003-1346-9
[10]
He R, Cui J M, Li R R, et al. An ion trap apparatus with high optical access in multiple directions. Review of Scientific Instruments,2021, 92: 073201. DOI: 10.1063/5.0043985
[11]
Chiaverini J, Blakestad R B, Britton J, et al. Surface-electrode architecture for ion-trap quantum information processing. Quantum Information and Computation,2005, 5: 419–439. DOI: 10.26421/QIC5.6-1
[12]
David Romaszko Z, Hong S, Siegele M, et al. Engineering of microfabricated ion traps and integration of advanced on-chip features. Nature Reviews Physics,2020, 2: 285–299. DOI: 10.1038/s42254-020-0182-8
[13]
Leibfried D, Blatt R, Monroe C, et al. Quantum dynamics of single trapped ions. Reviews of Modern Physics,2003, 75: 281–324. DOI: 10.1103/RevModPhys.75.281
[14]
Mehta K K, Zhang C, Malinowski M, et al. Integrated optical multi-ion quantum logic. Nature,2020, 586: 533–537. DOI: 10.1038/s41586-020-2823-6
[15]
Chou C K, Auchter C, Lilieholm J, et al. Note: Single ion imaging and fluorescence collection with a parabolic mirror trap. Review of Scientific Instruments,2017, 88: 086101. DOI: 10.1063/1.4996506
[16]
Wang Z, Wang B R, Ma Q L, et al. Design of a novel monolithic parabolic-mirror ion-trap to precisely align the RF null point with the optical focus. arXiv: 2004.08845, 2020.
[17]
Law C K, Kimble H J. Deterministic generation of a bit-stream of single-photon pulses. Journal of Modern Optics,1997, 44: 2067–2074. DOI: 10.1080/09500349708231869
[18]
Hunger D, Steinmetz T, Colombe Y, et al. A fiber Fabry-Perot cavity with high finesse. New Journal of Physics,2010, 12: 065038. DOI: 10.1088/1367-2630/12/6/065038
[19]
Schupp J, Krcmarsky V, Krutyanskiy V, et al. Interface between trapped-ion qubits and traveling photons with close-to-optimal efficiency. PRX Quantum,2021, 2: 020331. DOI: 10.1103/PRXQuantum.2.020331
[20]
Krutyanskiy V, Galli M, Krcmarsky V, et al. Entanglement of trapped-ion qubits separated by 230 meters. Physical Review Letters,2023, 130: 050803. DOI: 10.1103/PhysRevLett.130.050803
[21]
Wilk T, Webster S C, Kuhn A, et al. Single-atom single-photon quantum interface. Science,2007, 317: 488–490. DOI: 10.1126/science.1143835
[22]
Daiss S, Langenfeld S, Welte S, et al. A quantum-logic gate between distant quantum-network modules. Science,2021, 371: 614–617. DOI: 10.1126/science.abe3150
[23]
Thomas P, Ruscio L, Morin O, et al. Efficient generation of entangled multiphoton graph states from a single atom. Nature,2022, 608: 677–681. DOI: 10.1038/s41586-022-04987-5
[24]
Brandstätter B, McClung A, Schüppert K, et al. Integrated fiber-mirror ion trap for strong ion-cavity coupling. The Review of Scientific Instruments,2013, 84: 123104. DOI: 10.1063/1.4838696
[25]
Ballance T G, Meyer H M, Kobel P, et al. Cavity-induced backaction in Purcell-enhanced photon emission of a single ion in an ultraviolet fiber cavity. Physical Review A,2017, 95: 033812. DOI: 10.1103/PhysRevA.95.033812
[26]
Lee M, Lee M, Hong S, et al. Microelectromechanical-system-based design of a high-finesse fiber cavity integrated with an ion trap. Physical Review Applied,2019, 12: 044052. DOI: 10.1103/PhysRevApplied.12.044052
[27]
Takahashi H, Kassa E, Christoforou C, et al. Strong coupling of a single ion to an optical cavity. Physical Review Letters,2020, 124: 013602. DOI: 10.1103/PhysRevLett.124.013602
[28]
Teller M, Messerer V, Schüppert K, et al. Integrating a fiber cavity into a wheel trap for strong ion-cavity coupling. AVS Quantum Science,2023, 5: 012001. DOI: 10.1116/5.0121534
[29]
Kumph M, Henkel C, Rabl P, et al. Electric-field noise above a thin dielectric layer on metal electrodes. New Journal of Physics,2016, 18: 023020. DOI: 10.1088/1367-2630/18/2/023020
[30]
Teller M, Fioretto D A, Holz P C, et al. Heating of a trapped ion induced by dielectric materials. Physical Review Letters,2021, 126: 230505. DOI: 10.1103/PhysRevLett.126.230505
[31]
Sterk J D, Luo L, Manning T A, et al. Photon collection from a trapped ion-cavity system. Physical Review A,2012, 85: 062308. DOI: 10.1103/PhysRevA.85.062308
Anusree, V.C., Ptok, A., Sobieszczyk, P. et al. Correlation stabilized ferromagnetic MnRuAs with distorted kagome lattice. Physical Review B, 2025, 111(3): 035125.
DOI:10.1103/PhysRevB.111.035125
2.
Chen, Z., Wu, X., Zhou, S. et al. Discovery of a long-ranged charge order with 1/4 Ge1-dimerization in an antiferromagnetic Kagome metal. Nature Communications, 2024, 15(1): 6262.
DOI:10.1038/s41467-024-50661-x
3.
Wang, Y.. Encoding innumerable charge density waves of FeGe into polymorphs of LiFe6Ge6. Science China: Physics, Mechanics and Astronomy, 2024, 67(9): 297011.
DOI:10.1007/s11433-024-2423-2
4.
Wu, Q., Quan, W., Pan, S. et al. Atomically Thin Kagome-Structured Co9Te16 Achieved through Self-Intercalation and Its Flat Band Visualization. Nano Letters, 2024, 24(25): 7672-7680.
DOI:10.1021/acs.nanolett.4c01526
5.
Wang, Y.. Enhanced spin-polarization via partial Ge-dimerization as the driving force of the charge density wave in FeGe. Physical Review Materials, 2023, 7(10): 104006.
DOI:10.1103/PhysRevMaterials.7.104006
Figure
1.
(a) Schematic diagram of the fiber cavity ion trap system. The fiber cavity ion trap is placed in the middle of the vacuum chamber. A single 138Ba+ ion is trapped in the center. The red and blue lines are the modulation lasers. (b) The basic atomic energy level for photon generation.
Figure
2.
Diagram of the fiber cavity ion trap system in vacuum. The fiber electrodes are mounted on a fiber mounting base. This system uses a ceramic circuit board to energize the ion trap. The stainless steel base of the device is composed of two spring-connected parts, and the lower part of the stainless steel base is directly connected to the vacuum chamber through grippers.
Figure
3.
Flow chart for fabricating fiber electrodes. The cut bare fiber undergoes a complete process to form a central light-permeable electrode structure with a gold layer around the cavity surface to shield the charges from the high-reflection film.
Figure
4.
Arrangement of the fiber electrodes in the system. The four side fiber electrodes are grounded by default, and the middle fiber electrodes are applied with the RF electric field.
Figure
5.
Schematic diagram of the interaction between the fiber cavity and the ion. In the presence of an external electric field Escos(ωLt), the single ion interacts with the cavity. In cavity quantum electrodynamics, this interaction can be described using (g,κ,γ). There is only a fundamental mode in the cavity and the distribution is affected by the structure of the cavity. Here, ws,i\)(\(s=f,m\)and\(i=1,2) is the radius of the mode field at each location, w0 is the radius of the waist, Ri\)(\(i=1,2) is the radius of curvature (ROC), ni\)(\(i=1,2) is the refractive index, and L is the cavity length. At the fiber end face, only the mode field within diameter Di\)(\(i=1,2) can be reflected by the fiber.
Figure
6.
Variation curves of each coupling with the radius of curvature (ROC) of the left cavity surface R1 when reflectivity R1=99.8%, R2=99.98%, and the type of fiber is LMA-10 photonic crystal fiber. P is the overall efficiency in the figure.
Figure
7.
Simulated potential energy. A drive voltage of 100 V is applied to the RF fiber electrodes. The origin of these figures is the center of the fiber cavity.The curves in (a)–(c) show the potential energy along the x, y, and z directions through the origin, and each red dashed line represents the result of the fit of the quadratic function. The images in (d)–(f) show the potential energy in terms of different cross-section.
Figure
8.
Schematic diagram of the connection of the stainless steel base. The figure shows the spring arrangement as seen from the side. The stainless steel base uses a total of ten springs, four of which are used to achieve the connection in the vertical direction, and the remaining six springs are used to achieve the connection in the side.
Figure
9.
Simulated resonant modes of the upper part of the system. The solid blue lines in the figure are the original position of the upper part. The darker the color, the greater the displacement of the upper part.
References
[1]
Duan L M, Monroe C. Colloquium: Quantum networks with trapped ions. Reviews of Modern Physics,2010, 82: 1209–1224. DOI: 10.1103/RevModPhys.82.1209
[2]
Ritter S, Nölleke C, Hahn C, et al. An elementary quantum network of single atoms in optical cavities. Nature,2012, 484: 195–200. DOI: 10.1038/nature11023
[3]
Reiserer A, Kalb N, Rempe G, et al. A quantum gate between a flying optical photon and a single trapped atom. Nature,2014, 508: 237–240. DOI: 10.1038/nature13177
[4]
Uphoff M, Brekenfeld M, Rempe G, et al. An integrated quantum repeater at telecom wavelength with single atoms in optical fiber cavities. Applied Physics B,2016, 122: 46. DOI: https://doi.org/10.1007/s00340-015-6299-2
Paul W. Electromagnetic traps for charged and neutral particles. Reviews of Modern Physics,1990, 62: 531–540. DOI: 10.1103/RevModPhys.62.531
[7]
Wang C X, He R, Li R R, et al. Advances in the study of ion trap structures in quantum computation and simulation. Acta Physica Sinica,2022, 71: 133701. DOI: 10.7498/aps.71.20220224
[8]
Prestage J D, Dick G J, Maleki L. New ion trap for frequency standard applications. Journal of Applied Physics,1989, 66: 1013–1017. DOI: 10.1063/1.343486
[9]
Schmidt-Kaler F, Häffner H, Gulde S, et al. How to realize a universal quantum gate with trapped ions. Applied Physics B,2003, 77: 789–796. DOI: https://doi.org/10.1007/s00340-003-1346-9
[10]
He R, Cui J M, Li R R, et al. An ion trap apparatus with high optical access in multiple directions. Review of Scientific Instruments,2021, 92: 073201. DOI: 10.1063/5.0043985
[11]
Chiaverini J, Blakestad R B, Britton J, et al. Surface-electrode architecture for ion-trap quantum information processing. Quantum Information and Computation,2005, 5: 419–439. DOI: 10.26421/QIC5.6-1
[12]
David Romaszko Z, Hong S, Siegele M, et al. Engineering of microfabricated ion traps and integration of advanced on-chip features. Nature Reviews Physics,2020, 2: 285–299. DOI: 10.1038/s42254-020-0182-8
[13]
Leibfried D, Blatt R, Monroe C, et al. Quantum dynamics of single trapped ions. Reviews of Modern Physics,2003, 75: 281–324. DOI: 10.1103/RevModPhys.75.281
[14]
Mehta K K, Zhang C, Malinowski M, et al. Integrated optical multi-ion quantum logic. Nature,2020, 586: 533–537. DOI: 10.1038/s41586-020-2823-6
[15]
Chou C K, Auchter C, Lilieholm J, et al. Note: Single ion imaging and fluorescence collection with a parabolic mirror trap. Review of Scientific Instruments,2017, 88: 086101. DOI: 10.1063/1.4996506
[16]
Wang Z, Wang B R, Ma Q L, et al. Design of a novel monolithic parabolic-mirror ion-trap to precisely align the RF null point with the optical focus. arXiv: 2004.08845, 2020.
[17]
Law C K, Kimble H J. Deterministic generation of a bit-stream of single-photon pulses. Journal of Modern Optics,1997, 44: 2067–2074. DOI: 10.1080/09500349708231869
[18]
Hunger D, Steinmetz T, Colombe Y, et al. A fiber Fabry-Perot cavity with high finesse. New Journal of Physics,2010, 12: 065038. DOI: 10.1088/1367-2630/12/6/065038
[19]
Schupp J, Krcmarsky V, Krutyanskiy V, et al. Interface between trapped-ion qubits and traveling photons with close-to-optimal efficiency. PRX Quantum,2021, 2: 020331. DOI: 10.1103/PRXQuantum.2.020331
[20]
Krutyanskiy V, Galli M, Krcmarsky V, et al. Entanglement of trapped-ion qubits separated by 230 meters. Physical Review Letters,2023, 130: 050803. DOI: 10.1103/PhysRevLett.130.050803
[21]
Wilk T, Webster S C, Kuhn A, et al. Single-atom single-photon quantum interface. Science,2007, 317: 488–490. DOI: 10.1126/science.1143835
[22]
Daiss S, Langenfeld S, Welte S, et al. A quantum-logic gate between distant quantum-network modules. Science,2021, 371: 614–617. DOI: 10.1126/science.abe3150
[23]
Thomas P, Ruscio L, Morin O, et al. Efficient generation of entangled multiphoton graph states from a single atom. Nature,2022, 608: 677–681. DOI: 10.1038/s41586-022-04987-5
[24]
Brandstätter B, McClung A, Schüppert K, et al. Integrated fiber-mirror ion trap for strong ion-cavity coupling. The Review of Scientific Instruments,2013, 84: 123104. DOI: 10.1063/1.4838696
[25]
Ballance T G, Meyer H M, Kobel P, et al. Cavity-induced backaction in Purcell-enhanced photon emission of a single ion in an ultraviolet fiber cavity. Physical Review A,2017, 95: 033812. DOI: 10.1103/PhysRevA.95.033812
[26]
Lee M, Lee M, Hong S, et al. Microelectromechanical-system-based design of a high-finesse fiber cavity integrated with an ion trap. Physical Review Applied,2019, 12: 044052. DOI: 10.1103/PhysRevApplied.12.044052
[27]
Takahashi H, Kassa E, Christoforou C, et al. Strong coupling of a single ion to an optical cavity. Physical Review Letters,2020, 124: 013602. DOI: 10.1103/PhysRevLett.124.013602
[28]
Teller M, Messerer V, Schüppert K, et al. Integrating a fiber cavity into a wheel trap for strong ion-cavity coupling. AVS Quantum Science,2023, 5: 012001. DOI: 10.1116/5.0121534
[29]
Kumph M, Henkel C, Rabl P, et al. Electric-field noise above a thin dielectric layer on metal electrodes. New Journal of Physics,2016, 18: 023020. DOI: 10.1088/1367-2630/18/2/023020
[30]
Teller M, Fioretto D A, Holz P C, et al. Heating of a trapped ion induced by dielectric materials. Physical Review Letters,2021, 126: 230505. DOI: 10.1103/PhysRevLett.126.230505
[31]
Sterk J D, Luo L, Manning T A, et al. Photon collection from a trapped ion-cavity system. Physical Review A,2012, 85: 062308. DOI: 10.1103/PhysRevA.85.062308
Anusree, V.C., Ptok, A., Sobieszczyk, P. et al. Correlation stabilized ferromagnetic MnRuAs with distorted kagome lattice. Physical Review B, 2025, 111(3): 035125.
DOI:10.1103/PhysRevB.111.035125
2.
Chen, Z., Wu, X., Zhou, S. et al. Discovery of a long-ranged charge order with 1/4 Ge1-dimerization in an antiferromagnetic Kagome metal. Nature Communications, 2024, 15(1): 6262.
DOI:10.1038/s41467-024-50661-x
3.
Wang, Y.. Encoding innumerable charge density waves of FeGe into polymorphs of LiFe6Ge6. Science China: Physics, Mechanics and Astronomy, 2024, 67(9): 297011.
DOI:10.1007/s11433-024-2423-2
4.
Wu, Q., Quan, W., Pan, S. et al. Atomically Thin Kagome-Structured Co9Te16 Achieved through Self-Intercalation and Its Flat Band Visualization. Nano Letters, 2024, 24(25): 7672-7680.
DOI:10.1021/acs.nanolett.4c01526
5.
Wang, Y.. Enhanced spin-polarization via partial Ge-dimerization as the driving force of the charge density wave in FeGe. Physical Review Materials, 2023, 7(10): 104006.
DOI:10.1103/PhysRevMaterials.7.104006
Table
1.
A case of QA-pairs generated by our QGAE model: the model accepts unannotated texts as input, extracts the highlighted phrase “Lorentz’s law” as an answer, then uses this answer to make question generation.
Input context: Through combining the definition of electric current as the time rate of change of electric charge, a rule of vector multiplication called
Lorentz’s law describes the force on a charge moving in a magnetic field. The connection between electricity and magnetism allows for the description of a unified electromagnetic force that acts on a charge. This force can be written as a sum of the electrostatic force (due to the electric field) and the magnetic force (due to the magnetic field).
Extracted answer: Lorentz’s law
Generated question: What describes the force on a charge moving in a magnetic field?
Table
2.
Statistics of datasets SQuAD1.1 and SQuAD2.0. No matter in which dataset, an example consists of a context, a question, and an optional answer. The term “negative example” refers to a context passage paired with an unanswerable question, which is intended to help models learn to identify when a question cannot be answered correctly based on the given context.
Table
3.
Comparison of method performance in major metrics (including QG metrics and AE metric) on the SQuAD dataset. These methods are divided into four types according to their primary research fields. The first two classifications focus on their own independent fields, while the latter two classifications can accomplish these two tasks at the same time.