L

Luke Large

Developed by studio-ousia
LUKE is a Transformer-based pre-trained model specifically designed for words and entities, providing deep contextual representations through entity-aware self-attention mechanisms.
Downloads 1,040
Release Time : 3/2/2022

Model Overview

LUKE is an innovative pre-trained contextual representation method that treats words and entities in text as independent tokens and outputs their context-dependent representations. The model employs an entity-aware self-attention mechanism, extending the traditional Transformer's self-attention by considering token types (word or entity) when computing attention scores.

Model Features

Entity-aware Self-attention Mechanism
Extends the traditional Transformer's self-attention mechanism by considering token types (word or entity) when computing attention scores.
Joint Representation of Words and Entities
Treats words and entities in text as independent tokens and outputs their context-dependent representations.
Outstanding Multi-task Performance
Achieves state-of-the-art results on five mainstream natural language processing benchmarks.

Model Capabilities

Named Entity Recognition
Entity Typing
Relation Classification
Extractive Question Answering
Cloze-style Question Answering

Use Cases

Information Extraction
Named Entity Recognition
Identify and classify named entities (e.g., person names, locations, organizations) from text
Achieves 94.3 F1 score on the CoNLL-2003 dataset
Relation Classification
Identify relationship types between entities
Achieves 72.7 F1 score on the TACRED dataset
Question Answering
Extractive Question Answering
Extract answers from given text to answer natural language questions
Achieves 90.2 EM/95.4 F1 on the SQuAD v1.1 dataset
Cloze-style Question Answering
Fill in blanks in sentences by understanding context
Achieves 90.6 EM/91.2 F1 on the ReCoRD dataset
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase