Export Citations
Save this search
Please login to be able to save your searches and receive alerts for new content matching your search criteria.
- articleNovember 2024
Advancing 3D point cloud understanding through deep transfer learning: A comprehensive survey
- Shahab Saquib Sohail,
- Yassine Himeur,
- Hamza Kheddar,
- Abbes Amira,
- Fodil Fadli,
- Shadi Atalla,
- Abigail Copiaco,
- Wathiq Mansoor
AbstractThe 3D point cloud (3DPC) has significantly evolved and benefited from the advances of deep learning (DL). However, the latter faces various issues, including the lack of data or annotated data, the existence of a significant gap between training ...
Highlights- Deep transfer learning (DTL) and domain adaptation (DA) enhance 3D point cloud data.
- Taxonomy of DTL methods and knowledge transfer strategies for 3DPC.
- DTL’s applications in 3DPC, i.e., object detection, semantic labeling, and ...
- research-articleNovember 2024
LLMs-based machine translation for E-commerce
- Dehong Gao,
- Kaidi Chen,
- Ben Chen,
- Huangyu Dai,
- Linbo Jin,
- Wen Jiang,
- Wei Ning,
- Shanqing Yu,
- Qi Xuan,
- Xiaoyan Cai,
- Libin Yang,
- Zhen Wang
Expert Systems with Applications: An International Journal (EXWA), Volume 258, Issue Chttps://doi.org/10.1016/j.eswa.2024.125087AbstractLarge language models(LLMs) have shown promising performance for various downstream tasks, especially machine translation. However, LLMs and Specialized Translation Models (STMs) are designed to handle general translation needs, they are not well-...
Highlights- Existing universal translation models perform poorly in the e-commerce field.
- Translation models based on LLMs improve translation capabilities more efficiently.
- Resources based on e-commerce texts are conducive to e-commerce ...
- research-articleNovember 2024
Long Exposure: Accelerating Parameter-Efficient Fine-Tuning for LLMs under Shadowy Sparsity
SC '24: Proceedings of the International Conference for High Performance Computing, Networking, Storage, and AnalysisArticle No.: 75, Pages 1–18https://doi.org/10.1109/SC41406.2024.00081The adaptation of pre-trained large language models (LLMs) to diverse downstream tasks via fine-tuning is critical for numerous applications. However, the inefficiency of parameter-efficient fine-tuning (PEFT) techniques presents significant challenges ...
- ArticleNovember 2024
AQLoRA: An Adaptive Quantization-Based Efficient Fine-Tuning Method for LLMs
Natural Language Processing and Chinese ComputingPages 268–280https://doi.org/10.1007/978-981-97-9434-8_21AbstractLarge language models (LLMs) have shown exceptional performance in the domain of composite artificial intelligence tasks, offering a preliminary insight into the potential of general artificial intelligence. The fine-tuning process for LLMs ...
- ArticleNovember 2024
ConFit: Contrastive Fine-Tuning of Text-to-Text Transformer for Relation Classification
Natural Language Processing and Chinese ComputingPages 16–29https://doi.org/10.1007/978-981-97-9434-8_2AbstractRelation classification (RC) is commonly the second step in a relation extraction pipeline, which asserts the relation of two identified entities based on their context. The latest trend for dealing with the task resorts to pre-trained language ...
-
- research-articleNovember 2024
FINEST: Stabilizing Recommendations by Rank-Preserving Fine-Tuning
ACM Transactions on Knowledge Discovery from Data (TKDD), Volume 18, Issue 9Article No.: 237, Pages 1–22https://doi.org/10.1145/3695256Modern recommender systems may output considerably different recommendations due to small perturbations in the training data. Changes in the data from a single user will alter the recommendations as well as the recommendations of other users. In ...
- ArticleNovember 2024
Semantic-Guided Robustness Tuning for Few-Shot Transfer Across Extreme Domain Shift
AbstractIn this work, we focus on the cross-domain few-shot classification (CDFSC), which is mostly challenged by the low-data problem as well as extreme domain shift between base and novel target classes. Current methods always employ a lightweight ...
- ArticleOctober 2024
Collaborative Vision-Text Representation Optimizing for Open-Vocabulary Segmentation
AbstractPre-trained vision-language models, e.g. CLIP, have been increasingly used to address the challenging Open-Vocabulary Segmentation (OVS) task, benefiting from their well-aligned vision-text embedding space. Typical solutions involve either ...
- research-articleNovember 2024
LLM-Commentator: Novel fine-tuning strategies of large language models for automatic commentary generation using football event data
AbstractReal-time commentary on football matches is a challenging task that requires precise and coherent descriptions of events as they unfold. Traditional methods often fall short in providing timely and accurate insights into the game. This study aims ...
- research-articleSeptember 2024
Advanced fine-tuning procedures to enhance DNN robustness in visual coding for machines
Journal on Image and Video Processing (JIVP), Volume 2024, Issue 1https://doi.org/10.1186/s13640-024-00650-3AbstractVideo Coding for Machines (VCM) is gaining momentum in applications like autonomous driving, industry manufacturing, and surveillance, where the robustness of machine learning algorithms against coding artifacts is one of the key success factors. ...
- ArticleSeptember 2024
Transferability of Non-contrastive Self-supervised Learning to Chronic Wound Image Recognition
Artificial Neural Networks and Machine Learning – ICANN 2024Pages 427–444https://doi.org/10.1007/978-3-031-72353-7_31AbstractChronic wounds pose significant challenges in medical practice, necessitating effective treatment approaches and reduced burden on healthcare staff. Computer-aided diagnosis (CAD) systems offer promising solutions to enhance treatment outcomes. ...
- ArticleSeptember 2024
CSAFT: Continuous Semantic Augmentation Fine-Tuning for Legal Large Language Models
Artificial Neural Networks and Machine Learning – ICANN 2024Pages 293–307https://doi.org/10.1007/978-3-031-72344-5_20AbstractPrevious researches have demonstrated that fine-tuning with legal data can significantly enhance the performance of Large Language Models (LLMs) in legal question-answer (Q&A), affirming that data augmentation is an effective strategy. However, ...
- ArticleSeptember 2024
Effects of Training Strategies and the Amount of Speech Data on the Quality of Speech Synthesis
AbstractDuring the development of a speech synthesizer, we often face a lack of training data. This paper describes how the amount of data used to train a speech synthesizer affects the quality of the final synthetic speech. To answer this question, we ...
- research-articleOctober 2024
Improving unbalanced image classification through fine-tuning method of reinforcement learning
AbstractImage classification, especially unbalanced image classification, holds considerable promise for practical applications. Existing research focuses mainly on enhancing the effectiveness of classifiers through approaches such as data resampling and ...
Highlights- Implemented the unbalanced image classification task using reinforcement learning fine-tuning method.
- This method relies on the reward function model to maximize cumulative rewards.
- The fine-tuning process should control the ...
- research-articleSeptember 2024
Lv-Adapter: Adapting Vision Transformers for Visual Classification with Linear-layers and Vectors
Computer Vision and Image Understanding (CVIU), Volume 246, Issue Chttps://doi.org/10.1016/j.cviu.2024.104049AbstractLarge pre-trained models based on Vision Transformers (ViTs) contain nearly billions of parameters, demanding substantial computational resources and storage space. This restricts their transferability across different tasks. Recent approaches ...
Highlights- Proposed a Lv-Adapter fine-tuning module composed of Linear-layers and Vectors.
- Lv-Adapter can be plug-and-play in transformer-based models.
- When pretrained models are transferred to downstream tasks, Lv-Adapter can improve model ...
- ArticleAugust 2024
A Pre-trained Knowledge Tracing Model with Limited Data
AbstractOnline education systems have gained increasing popularity due to their capability to fully preserve users’ learning data. This advantage enables researchers to assess learners’ mastery through their learning trajectories, thereby facilitating ...
- ArticleAugust 2024
A Comparative Study of Different Pre-trained Language Models for Sentiment Analysis of Human-Computer Negotiation Dialogue
Knowledge Science, Engineering and ManagementPages 301–317https://doi.org/10.1007/978-981-97-5501-1_23AbstractThis paper offers a comprehensive comparative study of various pre-trained language models for sentiment analysis in human-computer negotiation dialogues. It examines numerous state-of-the-art PLMs, including GPT-3.5, BERT, and its variants, along ...
- research-articleAugust 2024
Online Scheduling and Pricing for Multi-LoRA Fine-Tuning Tasks
ICPP '24: Proceedings of the 53rd International Conference on Parallel ProcessingPages 357–366https://doi.org/10.1145/3673038.3673083Fine-tuning pre-trained models with task-specific data can produce customized models effective for downstream tasks. However, operating large-scale such fine-tuning tasks in real time in the data center faces non-trivial challenges, including ...
- research-articleAugust 2024
Selective privacy-preserving framework for large language models fine-tuning
Information Sciences: an International Journal (ISCI), Volume 678, Issue Chttps://doi.org/10.1016/j.ins.2024.121000AbstractFine-tuning pre-trained large language models (LLMs) helps various downstream tasks, but brings serious privacy leaks when relying on large amounts of data for training. Differentially private stochastic gradient descent (DPSGD) has been designed ...