Tag

pretrained models

0 views collected around this technical thread.

Rare Earth Juejin Tech Community
Rare Earth Juejin Tech Community
Apr 8, 2024 · Artificial Intelligence

PreFLMR: Scaling Up Fine-Grained Late-Interaction Multi-modal Retrievers

The article introduces PreFLMR, an open‑source, general‑purpose pre‑trained multimodal retriever that leverages fine‑grained late‑interaction to boost retrieval‑augmented generation for knowledge‑intensive visual tasks, describes its M2KR benchmark, training stages, and strong experimental results across multiple tasks.

AIFLMRRAG
0 likes · 11 min read
PreFLMR: Scaling Up Fine-Grained Late-Interaction Multi-modal Retrievers
AntTech
AntTech
Dec 13, 2023 · Artificial Intelligence

IEEE ICDM 2023 Graph Learning Challenge: Community Detection and Fraud Group Mining

The IEEE ICDM 2023 Graph Learning Challenge, co‑hosted by Ant Group and Zhejiang University, showcased deep graph learning approaches for community detection and fraud‑group mining, highlighting the winning team's Risk‑DCRN method and emphasizing the importance of pretrained models in large‑scale network analysis.

Deep LearningICDMcommunity-detection
0 likes · 5 min read
IEEE ICDM 2023 Graph Learning Challenge: Community Detection and Fraud Group Mining
DataFunSummit
DataFunSummit
Jun 24, 2023 · Artificial Intelligence

From Model to Service: Alibaba Cloud Machine Learning PAI One‑Stop Model Development and Deployment Practice

This article presents an end‑to‑end overview of Alibaba Cloud’s Machine Learning PAI platform, detailing the three‑stage ML workflow, challenges in model development, the role of pre‑trained and open‑source models, PAI’s architecture, a hands‑on demo, and MLOps best practices for efficient model deployment.

AI PlatformAlibaba CloudModel Deployment
0 likes · 11 min read
From Model to Service: Alibaba Cloud Machine Learning PAI One‑Stop Model Development and Deployment Practice
DataFunTalk
DataFunTalk
Jun 17, 2023 · Artificial Intelligence

Research on Text Generation for Structured Data

This article reviews the rapidly evolving field of structured‑data text generation, covering AI development stages, core concepts, model architectures from pipeline to pretrained transformers, key challenges such as content selection, numeric representation, reasoning and style control, and outlines recent research directions and Q&A insights.

AInumeric reasoningpretrained models
0 likes · 21 min read
Research on Text Generation for Structured Data
DataFunTalk
DataFunTalk
Jun 10, 2023 · Artificial Intelligence

Financial Event Analysis and Applications Based on Pre-trained Models

This article introduces the tasks, techniques, and frameworks for financial event analysis using pre‑trained language models, covering unstructured data parsing, event semantics, graph construction, detection, extraction, and prediction, and presents the TDE‑GTEE model that achieves state‑of‑the‑art performance even in few‑shot scenarios.

AIEvent Extractionevent graph
0 likes · 18 min read
Financial Event Analysis and Applications Based on Pre-trained Models
NetEase LeiHuo Testing Center
NetEase LeiHuo Testing Center
Jun 2, 2023 · Artificial Intelligence

AI Techniques for a Global Search Platform: Word Segmentation, Text Similarity, Image Retrieval, and Multimodal Models

This article shares the development of a global search platform that leverages AI technologies such as Chinese word segmentation, part‑of‑speech tagging, text similarity via Simhash and Synonyms, image similarity using histogram, Hamming distance and ResNet‑50, and multimodal CLIP‑based models to improve search efficiency and accuracy.

AIMultimodalNLP
0 likes · 12 min read
AI Techniques for a Global Search Platform: Word Segmentation, Text Similarity, Image Retrieval, and Multimodal Models
DataFunTalk
DataFunTalk
May 18, 2023 · Artificial Intelligence

Query Intent Recognition in Enterprise Search: Knowledge‑Enhanced and Pretrained Model Approaches

This article explains how Alibaba's enterprise search system tackles query intent recognition by combining knowledge‑enhanced techniques, short‑text classification, and pretrained language models such as StructBERT and prompt‑learning, and it shares two real‑world case studies, experimental results, and future research directions.

NLPenterprise searchknowledge enhancement
0 likes · 19 min read
Query Intent Recognition in Enterprise Search: Knowledge‑Enhanced and Pretrained Model Approaches
Sohu Tech Products
Sohu Tech Products
Mar 22, 2023 · Artificial Intelligence

An Overview of Prompt Learning in Natural Language Processing

This article reviews the evolution of NLP training paradigms, explains why prompt learning is needed, defines its core concepts, and surveys major hard‑template and soft‑template methods such as PET, LM‑BFF, P‑tuning, and Prefix‑tuning, highlighting their advantages for few‑shot and zero‑shot scenarios.

NLPfew-shotpretrained models
0 likes · 10 min read
An Overview of Prompt Learning in Natural Language Processing
Xiaohongshu Tech REDtech
Xiaohongshu Tech REDtech
Feb 16, 2023 · Artificial Intelligence

Intelligent Creative Generation and Optimization for Xiaohongshu Advertising

Xiaohongshu’s end‑to‑end intelligent creative platform extracts high‑quality images, generates diverse titles with RED‑pretrained GPT‑2/T5 models, and selects the best ads using a UCB‑based multi‑armed bandit that balances CTR uplift, revenue and user‑experience, while employing position‑corrected metrics and a scalable dual‑tower DNN to boost long‑tail performance and overall revenue.

AINLPUCB
0 likes · 18 min read
Intelligent Creative Generation and Optimization for Xiaohongshu Advertising
DataFunTalk
DataFunTalk
Jan 28, 2023 · Artificial Intelligence

Industry Search: Background, Technologies, and Real‑World Applications

This article presents a comprehensive overview of industry search, covering its background, core retrieval and ranking technologies—including sparse and dense retrieval, pre‑trained language models, tokenization, NER, adaptive multi‑task training, and re‑ranking models—followed by detailed case studies such as address analysis, family‑ID unification, emergency call handling, education photo‑search, and power‑knowledge‑base integration.

MultimodalNLPaddress analysis
0 likes · 13 min read
Industry Search: Background, Technologies, and Real‑World Applications
Rare Earth Juejin Tech Community
Rare Earth Juejin Tech Community
Oct 18, 2022 · Artificial Intelligence

Practical Implementation of Vision Transformer (ViT) for Image Classification in PyTorch

This article walks readers through building, training, and evaluating a Vision Transformer (ViT) model for a five‑class flower classification task, providing detailed code snippets, model architecture explanations, training script adjustments, and experimental results that highlight the importance of pre‑trained weights.

Deep LearningPyTorchViT
0 likes · 13 min read
Practical Implementation of Vision Transformer (ViT) for Image Classification in PyTorch
DataFunTalk
DataFunTalk
Aug 28, 2022 · Artificial Intelligence

Emerging Paths Toward General AI: Trends in Large‑Scale Pretrained Models

The article reviews how the Transformer breakthrough, the rapid scaling of large language models such as GPT‑3, Switch Transformer, and Alibaba's AliceMind and M6, together with multimodal research, are shaping the next phase of artificial intelligence toward more general, collaborative, and open AI systems.

AI TrendsArtificial IntelligenceDeep Learning
0 likes · 5 min read
Emerging Paths Toward General AI: Trends in Large‑Scale Pretrained Models
AntTech
AntTech
Jul 7, 2022 · Artificial Intelligence

Ant Group Insurance Technology Wins First Place in Fine‑Grained Dialogue Social Bias Detection at NLPCC 2023

Ant Group's insurance technology team secured the top spot in the fine‑grained dialogue social bias detection task at the 11th CCF NLPCC conference, showcasing their AI‑driven bias‑mitigation methods, a proprietary pre‑trained model AntInsBert, and a claim‑automation system that boosts insurance service fairness and efficiency.

AntInsBertInsurance AINLP
0 likes · 3 min read
Ant Group Insurance Technology Wins First Place in Fine‑Grained Dialogue Social Bias Detection at NLPCC 2023
DataFunTalk
DataFunTalk
May 22, 2022 · Artificial Intelligence

Advances in Information‑Flow Recommendation: Pre‑trained Models and Multimodal User‑Interface Modeling

This article reviews Huawei Noah's Ark Lab's work on modern information‑flow recommendation, covering the evolution from collaborative filtering to deep learning, the application of BERT‑based pre‑training for news ranking, multimodal user‑interface modeling, practical deployment challenges, and future research directions.

AIBERTHuawei
0 likes · 19 min read
Advances in Information‑Flow Recommendation: Pre‑trained Models and Multimodal User‑Interface Modeling
DataFunTalk
DataFunTalk
May 5, 2022 · Artificial Intelligence

NLP Evolution: Symbolic Deep Parsing vs Neural Pre‑trained Models, Low‑Code Trends, and Semi‑Automated Applications

The article reviews the history and current state of NLP, compares symbolic deep‑parsing and neural pre‑trained approaches, discusses the knowledge‑bottleneck and low‑code trend, and illustrates semi‑automated, low‑code NLP deployment in the financial domain while pondering future integration of symbolic and neural methods.

NLPSemi-AutomatedSymbolic AI
0 likes · 23 min read
NLP Evolution: Symbolic Deep Parsing vs Neural Pre‑trained Models, Low‑Code Trends, and Semi‑Automated Applications
DataFunTalk
DataFunTalk
Mar 17, 2022 · Artificial Intelligence

A Survey of Text Classification and Intent Recognition: Industrial and Research Perspectives

This article reviews recent developments in text classification and intent recognition, comparing industrial practices such as business‑coupled feature engineering with research trends like pretrained language models, and provides references and practical insights for building effective NLP solutions.

Industry ApplicationsNLPText Classification
0 likes · 13 min read
A Survey of Text Classification and Intent Recognition: Industrial and Research Perspectives
DataFunSummit
DataFunSummit
Feb 28, 2022 · Artificial Intelligence

UGC Sentiment Analysis Solutions and Applications in Taobao

This article presents a comprehensive overview of Taobao's user‑generated content (UGC) sentiment analysis pipeline, covering background, task definition, challenges, model architecture—including RoBERTa‑based extraction, sentiment‑knowledge pre‑training, and graph augmentation—personalized impression ranking, business impact cases, and future research directions.

UGCaspect extractione-commerce
0 likes · 16 min read
UGC Sentiment Analysis Solutions and Applications in Taobao
DataFunTalk
DataFunTalk
Feb 2, 2022 · Artificial Intelligence

UGC Sentiment Analysis Solutions and Applications in Taobao

This article presents a comprehensive overview of Taobao's user‑generated content sentiment analysis pipeline, covering task definition, challenges, model architecture with RoBERTa‑based extraction, sentiment‑knowledge pre‑training, graph augmentation, personalized ranking, business impact metrics, and future research directions.

Deep LearningUGCe-commerce
0 likes · 16 min read
UGC Sentiment Analysis Solutions and Applications in Taobao
DataFunSummit
DataFunSummit
Jan 13, 2022 · Artificial Intelligence

DeltaLM: A Multilingual Pretrained Encoder‑Decoder Model for Neural Machine Translation

DeltaLM is a multilingual pretrained encoder‑decoder model that leverages cross‑lingual transfer from a pretrained encoder and novel decoder architecture, employs span‑corruption and translation‑pair pretraining tasks, and uses a two‑stage fine‑tuning strategy to achieve strong zero‑shot and supervised translation performance across over 100 languages.

DeltaLMZero-shotcross-lingual transfer
0 likes · 12 min read
DeltaLM: A Multilingual Pretrained Encoder‑Decoder Model for Neural Machine Translation
Baidu Geek Talk
Baidu Geek Talk
Nov 29, 2021 · Artificial Intelligence

Pretrained Models for First-Stage Information Retrieval: A Comprehensive Review

This comprehensive review by Dr. Fan Yixing surveys how pretrained language models have transformed first‑stage information retrieval, tracing the shift from traditional term‑based methods to neural sparse, dense, and hybrid approaches, and discussing key challenges such as hard‑negative mining, joint indexing‑representation learning, and generative‑discriminative training.

Hybrid RetrievalNeural IRSparse Retrieval
0 likes · 15 min read
Pretrained Models for First-Stage Information Retrieval: A Comprehensive Review