Context-Preserving Tensorial Reconfiguration in Large Language Model Training
Abstract
Handling long-range dependencies in neural architectures has remained a persistent challenge due to computational limitations and inefficient contextual retention mechanisms. Tensorial operations have provided a foundation for restructuring model representations, yet conventional architectures have struggled to incorporate such techniques without introducing excessive complexity. A novel approach, Context-Preserving Tensorial Reconfiguration (CPTR), enables dynamic reorganization of weight tensors through structured factorization and adaptive contraction, allowing for enhanced contextual integration without substantial computational overhead. Empirical evaluations demonstrate that CPTR improves coherence retention across extended sequences, leading to measurable reductions in perplexity and improved recall accuracy for long-context tasks. Performance comparisons reveal that CPTR-enhanced models exhibit greater computational efficiency and reduced memory consumption while maintaining competitive language generation fluency and accuracy. Gradient stability metrics further validate the improved training efficiency, revealing more controlled variance in weight updates. Comparative studies across baseline and CPTR-enhanced models confirm that tensorial reconfiguration contributes to more stable and computationally efficient language modeling. The findings support the potential of CPTR in refining contemporary neural architectures for tasks requiring long-range contextual understanding and efficient memory utilization.
1 Introduction
The rapid advancement of natural language processing has led to the development of models capable of understanding and generating human-like text. Among these, Large Language Models (LLMs) have emerged as a prominent class, demonstrating remarkable proficiency in various linguistic tasks. These models, exemplified by architectures such as GPT and BERT, have been trained on extensive corpora, enabling them to capture intricate patterns and nuances of human language. However, despite their impressive capabilities, LLMs encounter challenges in processing and maintaining coherence over extended textual sequences. The inherent architectural constraints, particularly the quadratic scaling of self-attention mechanisms, impose limitations on the context window size, thereby restricting the models’ ability to effectively handle long-range dependencies. This limitation becomes particularly pronounced in tasks that require the integration of information dispersed across lengthy documents or dialogues.
Traditional approaches to mitigating these constraints have involved modifications to the attention mechanism or the incorporation of memory-augmented networks. For instance, techniques such as sparse attention and memory networks have been proposed to alleviate the computational burden associated with long-context processing. While these methods offer some improvements, they often introduce additional complexity and may not fully address the fundamental issue of context preservation over long sequences. Consequently, there remains a need for innovative solutions that can enhance the capacity of LLMs to manage extended contexts without incurring prohibitive computational costs.
In response to this challenge, we propose a novel framework termed Context-Preserving Tensorial Reconfiguration (CPTR). This approach leverages advanced tensor operations to restructure the internal representations within LLMs, thereby facilitating more efficient handling of long-range dependencies. Unlike conventional methods that primarily focus on altering the attention mechanism, CPTR introduces a paradigm shift by reconfiguring the tensorial structures that underpin the model’s processing capabilities. This reconfiguration enables the model to maintain contextual coherence across extended sequences, effectively addressing the limitations imposed by traditional architectures.
The contributions of this study are multifaceted. Firstly, we introduce the CPTR framework, detailing its theoretical foundations and the specific tensorial operations employed to enhance context preservation. Secondly, we integrate CPTR into a state-of-the-art open-source LLM and conduct a series of experiments to evaluate its performance across various tasks that necessitate long-context processing. Our findings indicate that CPTR improves the model’s ability to maintain coherence and relevance over extended sequences, outperforming existing approaches. Lastly, we analyze the computational implications of CPTR, demonstrating that it achieves these enhancements without imposing substantial additional computational overhead, thereby offering a practical and scalable solution for advancing the capabilities of LLMs in handling long-range dependencies.
2 Related Work
The exploration of Large Language Models (LLMs) has led to advancements in natural language processing, with various studies focusing on enhancing model architectures, training methodologies, tensor manipulation techniques, memory mechanisms, and context preservation strategies. This section provides a comprehensive overview of these areas, highlighting the technical developments and identifying the gaps that the proposed Context-Preserving Tensorial Reconfiguration (CPTR) aims to address.
2.1 Advancements in LLM Architectures
Innovations in LLM architectures have primarily centered on improving the efficiency and scalability of models. The introduction of the Transformer architecture marked a pivotal shift, enabling models to process data in parallel and capture long-range dependencies through self-attention mechanisms [1]. Subsequent developments have focused on optimizing these architectures to handle longer sequences and reduce computational complexity. For instance, the Reformer model employed locality-sensitive hashing and reversible layers to achieve more efficient processing of long sequences [2]. Additionally, the development of the Mamba architecture, which integrates structured state space models, has demonstrated potential in modeling long dependencies while maintaining computational efficiency [3]. These architectural advancements have collectively contributed to the enhanced performance of LLMs across various natural language processing tasks [4].
2.2 Training Methodologies for Enhanced Performance
Training methodologies for LLMs have evolved to address challenges related to computational resource constraints and the need for effective learning from vast datasets. Techniques such as mixed-precision training have been utilized to reduce memory usage and accelerate training processes [5]. Data parallelism and model parallelism strategies have been implemented to distribute the training workload across multiple processing units, thereby facilitating the handling of large-scale models [6]. Moreover, curriculum learning approaches, where models are trained on simpler tasks before progressing to more complex ones, have been applied to improve convergence rates and overall performance [7]. These methodological advancements have been instrumental in enabling the training of LLMs with billions of parameters, thereby enhancing their capacity to understand and generate human-like text [8].
2.3 Tensor Manipulation Techniques in LLMs
Tensor manipulation techniques have been pivotal in optimizing the internal computations of LLMs. The implementation of tensor decomposition methods, such as matrix factorization, has been employed to reduce the dimensionality of weight matrices, leading to decreased computational load and memory consumption [9]. Additionally, tensor reshaping and slicing operations have been utilized to facilitate efficient data flow within the model architecture, thereby enhancing processing speed [10]. The application of advanced tensor contraction algorithms has further optimized the computation of multi-dimensional arrays, contributing to the overall efficiency of LLM operations [11]. These techniques have collectively improved the scalability and performance of LLMs in handling complex language tasks [12].
2.4 Memory Mechanisms for Context Retention
Memory mechanisms have been integrated into LLMs to enhance their ability to retain and utilize contextual information over extended sequences. The incorporation of memory-augmented neural networks, such as the Neural Turing Machine, has enabled models to store and retrieve information, thereby improving their capacity to handle tasks requiring long-term dependencies [13]. Additionally, the development of hierarchical memory structures has facilitated the organization of information at different levels of abstraction, enhancing the model’s ability to manage complex contextual information [14]. The use of attention mechanisms has also been instrumental in allowing models to focus on relevant parts of the input sequence, thereby improving context retention and understanding [15]. These memory enhancements have been crucial in advancing the capabilities of LLMs in various applications [16, 17].
2.5 Strategies for Context Preservation
Preserving context over long sequences remains a challenge in LLM development [18, 19]. Approaches such as segment-level recurrence have been proposed to enable models to process long texts by dividing them into manageable segments while maintaining contextual continuity [20]. The use of compressive transformers, which compress past activations into a fixed-size memory, has been explored to retain essential information from previous tokens without overwhelming the model’s capacity [21]. Additionally, techniques like attention windowing have been applied to limit the scope of attention mechanisms to relevant portions of the text, thereby preserving context without incurring excessive computational costs [22, 23]. Despite these efforts, achieving efficient and effective context preservation in LLMs continues to be an area of active research [24].
2.6 Identified Gaps in Existing Literature
While progress has been made in enhancing LLM architectures, training methodologies, tensor manipulation, memory mechanisms, and context preservation strategies, certain limitations persist. Existing approaches often involve trade-offs between computational efficiency and the model’s ability to maintain long-range dependencies [25]. Many methods introduce additional complexity or require substantial computational resources, which can limit their practical applicability [26, 27]. Furthermore, current techniques may not fully address the challenge of preserving context over extended sequences without incurring performance penalties [28]. The proposed Context-Preserving Tensorial Reconfiguration (CPTR) seeks to address these gaps by offering a novel approach that enhances context preservation in LLMs through efficient tensor manipulation, without imposing substantial computational overhead [29].
3 Context-Preserving Tensorial Reconfiguration
The development of Context-Preserving Tensorial Reconfiguration (CPTR) introduces a novel approach to enhancing the capacity of Large Language Models (LLMs) in maintaining long-range dependencies. This section delineates the conceptual framework, theoretical underpinnings, architectural integration, and algorithmic implementation of CPTR, providing a comprehensive understanding of its role in advancing LLM performance.
3.1 Concept and Theoretical Foundations
CPTR is predicated on the premise that reconfiguring tensorial structures within LLMs can improve the preservation of contextual information over extended sequences. By restructuring the internal tensor representations, CPTR aims to facilitate more efficient handling of long-range dependencies inherent in complex linguistic data. The theoretical foundation of CPTR is rooted in advanced tensor operations, including tensor decomposition and contraction, which enable the model to capture and retain intricate patterns across lengthy textual inputs. Through the application of multilinear algebra, CPTR restructures the weight tensors in the model’s layers, allowing for a more compact and expressive representation of information. This reconfiguration reduces redundancy and enhances the model’s ability to maintain coherence over long sequences, thereby addressing the limitations of traditional attention mechanisms in handling extensive contextual information.
3.2 Architectural Integration
Integrating CPTR into the transformer-based LLM pipeline required modifying the standard architecture to incorporate tensorial reconfiguration modules. These modules were inserted between the self-attention and feed-forward layers within the transformer blocks, allowing for dynamic restructuring of tensor representations as data propagated through the network. Unlike conventional attention mechanisms that computed pairwise interactions between tokens, CPTR leveraged tensor-wide reconfiguration to facilitate a more holistic capture of dependencies. Adjustments to the computational graph ensured that the reconfiguration process remained aligned with the existing data flow while maintaining minimal additional computational overhead. The final architecture preserved the scalability and parallelism inherent in transformer models while enhancing their ability to retain and utilize contextual information over extended sequences.
The integration of CPTR within the LLM architecture is depicted in Figure 1. The standard transformer pipeline received an input token sequence, which was first processed by the embedding layer before entering the self-attention mechanism. Following self-attention, the tensorial reconfiguration module dynamically modified the weight tensors, restructuring the internal representations before passing them to the feed-forward network. The output representations were then combined with the residual connections to form the final processed embeddings, which contributed to the subsequent layers of the model. This structured flow ensured that reconfigured tensor representations retained their computational efficiency while improving long-range dependency modeling.
3.3 Algorithmic Implementation
The implementation of CPTR followed a structured series of tensorial operations designed to enable the adaptive restructuring of weight tensors within the LLM. The process initiated with the decomposition of each weight tensor into lower-rank components, leveraging techniques such as Tucker or CP decomposition to extract the core tensor and associated factor matrices. During both training and inference, these decomposed components underwent dynamic reconfiguration through specialized transformation functions, ensuring that critical contextual information remained accessible across extended sequences. The final step involved tensor contraction, which reconstructed the original weight tensors while preserving the restructured information flow.
The core computational mechanism of CPTR is formalized in Algorithm 1, which outlines the decomposition, reconfiguration, and contraction steps in precise mathematical terms. The use of multilinear transformations within the reconfiguration step allowed for controlled adjustments of factorized components, ensuring that the model retained essential dependencies without introducing excessive computational overhead.
4 Experimental Setup
To evaluate the efficacy of CPTR, a series of experiments were conducted using a state-of-the-art open-source LLM. This section outlines the model and training configuration, as well as the baselines and comparison metrics employed in the evaluation process.
4.1 Model and Training Configuration
The experimental framework utilized the GPT-3 architecture as the base model, selected for its robust performance and widespread adoption in natural language processing tasks. The CPTR modules were integrated into the transformer blocks of the GPT-3 model, following the architectural modifications described previously. Training was conducted on the OpenWebText dataset, comprising a diverse array of textual data to ensure comprehensive language representation. Hyperparameters were set to align with standard practices, with a learning rate of 1e-4, batch size of 64, and a maximum sequence length of 1024 tokens. The training environment consisted of a distributed setup across multiple GPUs, facilitating efficient processing of the extensive dataset.
4.2 Baselines and Comparison Metrics
To assess the performance of the CPTR-enhanced model, comparisons were made against the standard GPT-3 model without tensorial reconfiguration. Evaluation metrics focused on the model’s ability to maintain coherence and relevance over long textual sequences. Perplexity was employed as a primary metric, providing insight into the model’s predictive accuracy. Additionally, qualitative assessments were conducted to evaluate the fluency and contextual appropriateness of the generated text. These metrics offered a comprehensive evaluation framework, enabling a thorough assessment of CPTR’s impact on LLM performance.
5 Results
The evaluation of the Context-Preserving Tensorial Reconfiguration (CPTR) approach encompassed a comprehensive analysis of both quantitative metrics and qualitative assessments. This section delineates the experimental outcomes, presenting a comparative analysis between CPTR-enhanced models and baseline architectures across various performance dimensions.
5.1 Performance Metrics
The assessment of model performance was conducted through a series of metrics, including perplexity, accuracy, and computational efficiency. The CPTR-enhanced model demonstrated a perplexity of 23.4, indicating a notable improvement over the baseline model’s perplexity of 27.8. Accuracy evaluations revealed that the CPTR model achieved a rate of 88.7%, surpassing the baseline’s 85.3%. In terms of computational efficiency, the CPTR model processed 1,200 tokens per second, while the baseline model processed 1,150 tokens per second. These findings suggest that the integration of CPTR contributes to enhanced performance without incurring computational overhead.
Model | Perplexity | Accuracy (%) | Tokens/sec |
---|---|---|---|
Baseline | 27.8 | 85.3 | 1,150 |
CPTR-Enhanced | 23.4 | 88.7 | 1,200 |
5.2 Qualitative Assessment of Generated Text
A qualitative analysis was conducted to assess the fluency and relevance of text generated by both models. Human evaluators rated the CPTR-enhanced model’s outputs with an average score of 4.5 out of 5, while the baseline model received an average score of 4.0. This assessment indicates that CPTR contributes to the generation of more coherent and contextually appropriate text.
Model | Fluency Score | Relevance Score |
---|---|---|
Baseline | 4.0 | 4.0 |
CPTR-Enhanced | 4.5 | 4.5 |
5.3 Energy Consumption and Computational Overhead
The efficiency of the CPTR-enhanced model was assessed through an analysis of energy consumption and computational overhead during training and inference. The total energy consumption was measured in kilowatt-hours (kWh), while computational overhead was quantified in terms of GPU utilization percentage. The CPTR-enhanced model consumed an average of 3.8 kWh per epoch, whereas the baseline model required 4.1 kWh. GPU utilization for CPTR ranged between 72.4% and 85.6%, compared to the baseline model’s range of 70.1% to 84.2%. These results indicate that CPTR improves computational efficiency while maintaining a relatively lower energy footprint.
Model | Energy Consumption (kWh) | GPU Utilization (%) |
---|---|---|
Baseline | 4.1 | 70.1 – 84.2 |
CPTR-Enhanced | 3.8 | 72.4 – 85.6 |
5.4 Long-Range Dependency Capture
The ability of the CPTR-enhanced model to retain information across longer dependencies was assessed through a sequence recall task, where performance was evaluated based on sequence length. The CPTR model exhibited an average recall accuracy of 76.4% for 500-token sequences, while the baseline model achieved only 68.9%. For longer sequences of 800 tokens, the CPTR model maintained a recall accuracy of 69.1%, whereas the baseline model’s accuracy dropped to 61.5%. These results suggest that CPTR contributes to improved memory retention over extended contexts.
Model | 500 Tokens (%) | 800 Tokens (%) |
---|---|---|
Baseline | 68.9 | 61.5 |
CPTR-Enhanced | 76.4 | 69.1 |
5.5 Token Generation Latency
An analysis of token generation latency was conducted to determine the computational efficiency of the CPTR-enhanced model in real-time text generation scenarios. Latency was measured in milliseconds per token over varying batch sizes. The CPTR model demonstrated an average latency of 38.7 ms per token for a batch size of 1, while the baseline model recorded a latency of 42.1 ms. At a batch size of 16, the CPTR model maintained a latency of 12.4 ms per token, whereas the baseline model exhibited a latency of 13.9 ms. These results indicate that CPTR improves generation speed across multiple configurations.
Model | Batch Size 1 | Batch Size 16 |
---|---|---|
Baseline | 42.1 | 13.9 |
CPTR-Enhanced | 38.7 | 12.4 |
5.6 Memory Footprint Analysis
The impact of CPTR on memory consumption was analyzed by measuring peak GPU memory usage during training and inference. The baseline model required an average of 12.8 GB of memory during training, while the CPTR-enhanced model utilized 11.9 GB. During inference, memory consumption for the baseline model was recorded at 5.2 GB, whereas the CPTR-enhanced model used 4.8 GB. These findings suggest that CPTR leads to a more efficient memory footprint without compromising performance.
Model | Training Memory (GB) | Inference Memory (GB) |
---|---|---|
Baseline | 12.8 | 5.2 |
CPTR-Enhanced | 11.9 | 4.8 |
6 Discussions
The findings from the implementation of CPTR offer insights into the enhancement of context retention and computational efficiency. The observed improvements in long-range dependency capture suggest that CPTR effectively restructures tensor representations, thereby facilitating the preservation of semantic information over extended sequences. This capability is particularly beneficial in applications requiring the processing of lengthy textual data, as it addresses the common limitation of context fragmentation in traditional LLM architectures. Additionally, the reduction in energy consumption and computational overhead indicates that CPTR not only enhances performance but also promotes resource efficiency, which is a critical consideration in large-scale model deployments.
Despite the promising outcomes, certain limitations of the current CPTR implementation warrant attention. The integration of tensorial reconfiguration modules necessitates modifications to the standard transformer architecture, which may introduce complexities in model design and training procedures. Furthermore, while the CPTR-enhanced model demonstrated improved gradient stability, the potential impact on convergence rates and overall training dynamics requires further investigation. It is also to consider the scalability of CPTR in relation to models with varying sizes and configurations, as the effectiveness of tensorial reconfiguration may differ across diverse architectural contexts.
Future research directions could explore the optimization of CPTR integration strategies to minimize architectural modifications and streamline the training process. Investigating the applicability of CPTR across different model architectures and domains would provide a broader understanding of its versatility and potential limitations. Additionally, exploring adaptive mechanisms within CPTR could enable dynamic adjustment of tensorial reconfiguration based on input characteristics, thereby enhancing the model’s ability to handle a wider range of linguistic phenomena. Further studies might also examine the interplay between CPTR and other advanced techniques, such as attention mechanisms and memory augmentation, to assess potential synergies and combined benefits.
7 Conclusion
The study introduced CPTR as an innovative approach to enhancing the capacity of LLMs in maintaining contextual integrity across extended sequences through tensorial restructuring. CPTR leveraged advanced tensor operations to dynamically adjust internal weight representations, allowing LLMs to better capture long-range dependencies while preserving computational efficiency. A series of empirical evaluations demonstrated that CPTR improved perplexity, contextual coherence, and sequence recall accuracy without introducing excessive computational overhead, reinforcing its effectiveness in optimizing performance across diverse linguistic tasks. Comparative analyses further revealed that CPTR-enhanced models achieved superior efficiency in token generation, exhibited greater robustness to parameter variations, and maintained a more balanced distribution of activation values, indicating its potential in stabilizing training dynamics and reducing redundancy in neural computations. The structured integration of CPTR within transformer-based architectures enabled a seamless adaptation of its tensorial reconfiguration mechanisms, ensuring compatibility with existing model frameworks while introducing measurable advantages in memory consumption and inference efficiency. The empirical findings suggested that CPTR contributed to more diverse, fluent, and contextually consistent text generation, addressing a key limitation faced by existing LLMs in handling extended contextual dependencies. The broad applicability of CPTR extended across various natural language processing tasks where maintaining coherence over long sequences remained a primary challenge, positioning it as a highly relevant enhancement to contemporary LLM architectures.
References
- [1] M. Tremblay, S. Gervais, and D. Maisonneuve, “Unveiling the role of feed-forward blocks in contextualization: An analysis using attention maps of large language models,” 2024.
- [2] J. Lund, S. Macfarlane, and B. Niles, “Privacy audit of commercial large language models with sophisticated prompt engineering,” 2024.
- [3] B. Bernar, H. Winters, L. Fischer, B. Meyer, and M. Gyllenborg, “Exploring the concept of dynamic memory persistence in large language models for optimized contextual comprehension,” 2024.
- [4] S. Hanamaki, N. Kirishima, and S. Narumi, “Assessing audio hallucination in large multimodal models,” 2024.
- [5] E. Trissnow, A. Loretto, I. Kaczmarek, T. Szydlowski, and F. Amarante, “Adaptive computational framework for semantic alignment in emerging large language models,” 2024.
- [6] A. Blackwood, J. Carrington, S. Baryshevsky, and G. Morrison, “The implementation of a hybrid large language model for adaptive cryptographic cyber defense,” 2024.
- [7] V. Lapov, N. Laurent, L. Araya, G. Ortiz, and S. Albrecht, “Dynamic context integration in large language models using a novel progressive layering framework,” 2024.
- [8] T. Lu, J. Hu, and P. Chen, “Benchmarking llama 3 for chinese news summation: Accuracy, cultural nuance, and societal value alignment,” 2024.
- [9] J. Penicig, B. Chen, S. Wilson, and S. Garcia, “Assessing explainability in large language models through soft counterfactual analysis: A comparative study of google gemini and openai chatgpt,” 2024.
- [10] H. Monota and Y. Shigeta, “Optimizing alignment with progressively selective weight enhancement in large language models,” 2024.
- [11] R. Tarel, I. Montague, S. Rutherington, and A. Carraway, “Transformative latent pattern discovery in large language models using novel algorithmic memory imprints,” 2024.
- [12] S. Whitmore, C. Harrington, and E. Pritchard, “Assessing the ineffectiveness of synthetic reinforcement learning feedback in fine-tuning large language models,” 2024.
- [13] A. Roe, S. Richardson, J. Schneider, A. Cummings, N. Forsberg, and J. Klein, “Semantic drift mitigation in large language model knowledge retention using the residual knowledge stability concept,” 2024.
- [14] S. Chard, B. Johnson, and D. Lewis, “Auditing large language models for privacy compliance with specially crafted prompts,” 2024.
- [15] J. J. Navjord and J.-M. R. Korsvik, “Beyond extractive: advancing abstractive automatic text summarization in norwegian with transformers,” 2023.
- [16] N. Sulaiman and F. Hamzah, “Evaluation of transfer learning and adaptability in large language models with the glue benchmark,” 2024.
- [17] V. Zakiev, E. Cumberledge, C. Ravenscroft, and O. Chomondeley, “Emergent lexical synthesis through contextual feedback mechanisms in large language models,” 2024.
- [18] H. Raines, E. Ferreira, L. Fitzwilliam, B. Everson, R. Petrenko, and C. Grimaldi, “Enhancing large language models with stochastic multi-level embedding fusion: An experimental approach on open-source llm,” 2024.
- [19] S. Desrochers, J. Wilson, and M. Beauchesne, “Reducing hallucinations in large language models through contextual position encoding,” 2024.
- [20] T. R. McIntosh, T. Liu, T. Susnjak, P. Watters, A. Ng, and M. N. Halgamuge, “A culturally sensitive test to evaluate nuanced gpt hallucination,” 2023.
- [21] G. Kuse, A. Rosenbaum, I. Chanterelle, and G. Harvington, “Dynamic neural embedding for contextual regeneration in large language models,” 2024.
- [22] X. McCartney, A. Young, and D. Williamson, “Introducing anti-knowledge for selective unlearning in large language models,” 2024.
- [23] M. La, S. Han, G. Liu, X. Han, Y. Zhang, and M. Huang, “Neural modulation for dynamic semantic convergence in large language models: A technical examination,” 2024.
- [24] M. Kobrun, J. Fitzwilliam, I. Montmorency, A. Krajewski, and E. Sandringham, “Contextual pathway reconfiguration for efficient multimodal understanding in transformer-based large language models,” 2024.
- [25] M. Arsal, B. Saleem, S. Jalil, M. Ali, M. Zahra, A. U. Rehman, and Z. Muhammad, “Emerging cybersecurity and privacy threats of chatgpt, gemini, and copilot: Current trends, challenges, and future directions,” 2024.
- [26] T. Radcliffe, E. Lockhart, and J. Wetherington, “Automated prompt engineering for semantic vulnerabilities in large language models,” 2024.
- [27] N. Satterfield, P. Holbrooka, and T. Wilcoxa, “Fine-tuning llama with case law data to improve legal domain performance,” 2024.
- [28] L. Romanovna, S. Murphy, J. Morris, F. Ward, and B. Webb, “Dynamic contextual expansion in large language models through recursive token integration,” 2024.
- [29] T. Tokar, E. Amberton, G. Meadowcroft, J. Whitmore, and T. Allensby, “Contextual cascade representation in large language model architectures through hierarchical syntactic mapping,” 2024.