32.3734, Books: Pretrained Transformers for Text Ranking: Lin, Nogueira, Yates

The LINGUIST List linguist at listserv.linguistlist.org
Tue Nov 30 17:20:28 UTC 2021


LINGUIST List: Vol-32-3734. Tue Nov 30 2021. ISSN: 1069 - 4875.

Subject: 32.3734, Books: Pretrained Transformers for Text Ranking: Lin, Nogueira, Yates

Moderator: Malgorzata E. Cavar (linguist at linguistlist.org)
Student Moderator: Jeremy Coburn, Lauren Perkins
Managing Editor: Becca Morris
Team: Helen Aristar-Dry, Everett Green, Sarah Robinson, Nils Hjortnaes, Joshua Sims, Billy Dickson
Jobs: jobs at linguistlist.org | Conferences: callconf at linguistlist.org | Pubs: pubs at linguistlist.org

Homepage: http://linguistlist.org

Please support the LL editors and operation with a donation at:
           https://funddrive.linguistlist.org/donate/

Editor for this issue: Billy Dickson <billyd at linguistlist.org>
================================================================


Date: Tue, 30 Nov 2021 12:20:03
From: Brent Beckley [beckley at morganclaypool.com]
Subject: Pretrained Transformers for Text Ranking: Lin, Nogueira, Yates

 


Title: Pretrained Transformers for Text Ranking 
Subtitle: BERT and Beyond 
Series Title: Synthesis Lectures on Human Language Technologies  

Publication Year: 2021 
Publisher: Morgan & Claypool Publishers
	   http://www.morganclaypool.com
	

Book URL: https://www.morganclaypoolpublishers.com/catalog_Orig/product_info.php?products_id=1681 


Author: Jimmy Lin
Author: Rodrigo Nogueira
Author: Andrew Yates

Paperback: ISBN:  9781636392288 Pages: 325 Price: ----  89.95


Abstract:

The goal of text ranking is to generate an ordered list of texts retrieved
from a corpus in response to a query. Although the most common formulation of
text ranking is search, instances of the task can also be found in many
natural language processing (NLP) applications. This book provides an overview
of text ranking with neural network architectures known as transformers, of
which BERT (Bidirectional Encoder Representations from Transformers) is the
best-known example. The combination of transformers and self-supervised
pretraining has been responsible for a paradigm shift in NLP, information
retrieval (IR), and beyond.

This book provides a synthesis of existing work as a single point of entry for
practitioners who wish to gain a better understanding of how to apply
transformers to text ranking problems and researchers who wish to pursue work
in this area. It covers a wide range of modern techniques, grouped into two
high-level categories: transformer models that perform reranking in
multi-stage architectures and dense retrieval techniques that perform ranking
directly. Two themes pervade the book: techniques for handling long documents,
beyond typical sentence-by-sentence processing in NLP, and techniques for
addressing the tradeoff between effectiveness (i.e., result quality) and
efficiency (e.g., query latency, model and index size). Although transformer
architectures and pretraining techniques are recent innovations, many aspects
of how they are applied to text ranking are relatively well understood and
represent mature techniques. However, there remain many open research
questions, and thus in addition to laying out the foundations of pretrained
transformers for text ranking, this book also attempts to prognosticate where
the field is heading.
 



Linguistic Field(s): Computational Linguistics


Written In: English  (eng)

See this book announcement on our website: 
http://linguistlist.org/pubs/books/get-book.cfm?BookID=158133




------------------------------------------------------------------------------

***************************    LINGUIST List Support    ***************************
 The 2020 Fund Drive is under way! Please visit https://funddrive.linguistlist.org
  to find out how to donate and check how your university, country or discipline
     ranks in the fund drive challenges. Or go directly to the donation site:
                   https://crowdfunding.iu.edu/the-linguist-list

                        Let's make this a short fund drive!
                Please feel free to share the link to our campaign:
                    https://funddrive.linguistlist.org/donate/
 


----------------------------------------------------------
LINGUIST List: Vol-32-3734	
----------------------------------------------------------






More information about the LINGUIST mailing list