Articoli correlati a Pretrained Transformers for Text Ranking: BERT and...

Pretrained Transformers for Text Ranking: BERT and Beyond - Brossura

 
9783031010538: Pretrained Transformers for Text Ranking: BERT and Beyond

Sinossi

The goal of text ranking is to generate an ordered list of texts retrieved from a corpus in response to a query. Although the most common formulation of text ranking is search, instances of the task can also be found in many natural language processing (NLP) applications.This book provides an overview of text ranking with neural network architectures known as transformers, of which BERT (Bidirectional Encoder Representations from Transformers) is the best-known example. The combination of transformers and self-supervised pretraining has been responsible for a paradigm shift in NLP, information retrieval (IR), and beyond. This book provides a synthesis of existing work as a single point of entry for practitioners who wish to gain a better understanding of how to apply transformers to text ranking problems and researchers who wish to pursue work in this area. It covers a wide range of modern techniques, grouped into two high-level categories: transformer models that perform reranking inmulti-stage architectures and dense retrieval techniques that perform ranking directly. Two themes pervade the book: techniques for handling long documents, beyond typical sentence-by-sentence processing in NLP, and techniques for addressing the tradeoff between effectiveness (i.e., result quality) and efficiency (e.g., query latency, model and index size). Although transformer architectures and pretraining techniques are recent innovations, many aspects of how they are applied to text ranking are relatively well understood and represent mature techniques. However, there remain many open research questions, and thus in addition to laying out the foundations of pretrained transformers for text ranking, this book also attempts to prognosticate where the field is heading.

Le informazioni nella sezione "Riassunto" possono far riferimento a edizioni diverse di questo titolo.

Informazioni sull?autore

Jimmy Lin holds the David R. Cheriton Chair in the David R. Cheriton School of Computer Science at the University of Waterloo. Prior to 2015, he was a faculty at the University of Maryland, College Park. Lin received his Ph.D. in Electrical Engineering and Computer Science from the Massachusetts Institute of Technology in 2004.Rodrigo Nogueira is a post-doctoral researcher at the University of Waterloo, an adjunct professor at the University of Campinas (UNICAMP), and a senior research scientist at NeuralMind, a startup focused on applying deep learning to document and image analysis. Nogueira received his Ph.D. in Computer Science from the New York University in 2019.Andrew Yates is an assistant professor in the Informatics Institute at the University of Amsterdam. Prior to 2021, he was a post-doctoral researcher and then senior researcher at the Max Planck Institute for Informatics. Yates received his Ph.D. in Computer Science from Georgetown University in 2016.

Le informazioni nella sezione "Su questo libro" possono far riferimento a edizioni diverse di questo titolo.

EUR 9,70 per la spedizione da Germania a Italia

Destinazione, tempi e costi

Altre edizioni note dello stesso titolo

9781636392288: Pretrained Transformers for Text Ranking: Bert and Beyond

Edizione in evidenza

ISBN 10:  1636392288 ISBN 13:  9781636392288
Casa editrice: Morgan & Claypool, 2021
Brossura

Risultati della ricerca per Pretrained Transformers for Text Ranking: BERT and...

Immagini fornite dal venditore

Lin, Jimmy|Nogueira, Rodrigo|Yates, Andrew
ISBN 10: 3031010531 ISBN 13: 9783031010538
Nuovo Brossura
Print on Demand

Da: moluna, Greven, Germania

Valutazione del venditore 5 su 5 stelle 5 stelle, Maggiori informazioni sulle valutazioni dei venditori

Condizione: New. Dieser Artikel ist ein Print on Demand Artikel und wird nach Ihrer Bestellung fuer Sie gedruckt. The goal of text ranking is to generate an ordered list of texts retrieved from a corpus in response to a query. Although the most common formulation of text ranking is search, instances of the task can also be found in many natural language processing (NLP. Codice articolo 608129300

Contatta il venditore

Compra nuovo

EUR 72,89
Convertire valuta
Spese di spedizione: EUR 9,70
Da: Germania a: Italia
Destinazione, tempi e costi

Quantità: Più di 20 disponibili

Aggiungi al carrello

Foto dell'editore

Lin, Jimmy; Nogueira, Rodrigo; Yates, Andrew
Editore: Springer, 2021
ISBN 10: 3031010531 ISBN 13: 9783031010538
Nuovo Brossura

Da: Ria Christie Collections, Uxbridge, Regno Unito

Valutazione del venditore 5 su 5 stelle 5 stelle, Maggiori informazioni sulle valutazioni dei venditori

Condizione: New. In English. Codice articolo ria9783031010538_new

Contatta il venditore

Compra nuovo

EUR 84,86
Convertire valuta
Spese di spedizione: EUR 10,34
Da: Regno Unito a: Italia
Destinazione, tempi e costi

Quantità: Più di 20 disponibili

Aggiungi al carrello

Immagini fornite dal venditore

Jimmy Lin
ISBN 10: 3031010531 ISBN 13: 9783031010538
Nuovo Taschenbuch
Print on Demand

Da: BuchWeltWeit Ludwig Meier e.K., Bergisch Gladbach, Germania

Valutazione del venditore 5 su 5 stelle 5 stelle, Maggiori informazioni sulle valutazioni dei venditori

Taschenbuch. Condizione: Neu. This item is printed on demand - it takes 3-4 days longer - Neuware -The goal of text ranking is to generate an ordered list of texts retrieved from a corpus in response to a query. Although the most common formulation of text ranking is search, instances of the task can also be found in many natural language processing (NLP) applications.This book provides an overview of text ranking with neural network architectures known as transformers, of which BERT (Bidirectional Encoder Representations from Transformers) is the best-known example. The combination of transformers and self-supervised pretraining has been responsible for a paradigm shift in NLP, information retrieval (IR), and beyond. This book provides a synthesis of existing work as a single point of entry for practitioners who wish to gain a better understanding of how to apply transformers to text ranking problems and researchers who wish to pursue work in this area. It covers a wide range of modern techniques, grouped into two high-level categories: transformer models that perform reranking inmulti-stage architectures and dense retrieval techniques that perform ranking directly. Two themes pervade the book: techniques for handling long documents, beyond typical sentence-by-sentence processing in NLP, and techniques for addressing the tradeoff between effectiveness (i.e., result quality) and efficiency (e.g., query latency, model and index size). Although transformer architectures and pretraining techniques are recent innovations, many aspects of how they are applied to text ranking are relatively well understood and represent mature techniques. However, there remain many open research questions, and thus in addition to laying out the foundations of pretrained transformers for text ranking, this book also attempts to prognosticate where the field is heading. 328 pp. Englisch. Codice articolo 9783031010538

Contatta il venditore

Compra nuovo

EUR 85,59
Convertire valuta
Spese di spedizione: EUR 11,00
Da: Germania a: Italia
Destinazione, tempi e costi

Quantità: 2 disponibili

Aggiungi al carrello

Immagini fornite dal venditore

Jimmy Lin
ISBN 10: 3031010531 ISBN 13: 9783031010538
Nuovo Taschenbuch

Da: AHA-BUCH GmbH, Einbeck, Germania

Valutazione del venditore 5 su 5 stelle 5 stelle, Maggiori informazioni sulle valutazioni dei venditori

Taschenbuch. Condizione: Neu. Druck auf Anfrage Neuware - Printed after ordering - The goal of text ranking is to generate an ordered list of texts retrieved from a corpus in response to a query. Although the most common formulation of text ranking is search, instances of the task can also be found in many natural language processing (NLP) applications.This book provides an overview of text ranking with neural network architectures known as transformers, of which BERT (Bidirectional Encoder Representations from Transformers) is the best-known example. The combination of transformers and self-supervised pretraining has been responsible for a paradigm shift in NLP, information retrieval (IR), and beyond. This book provides a synthesis of existing work as a single point of entry for practitioners who wish to gain a better understanding of how to apply transformers to text ranking problems and researchers who wish to pursue work in this area. It covers a wide range of modern techniques, grouped into two high-level categories: transformer models that perform reranking inmulti-stage architectures and dense retrieval techniques that perform ranking directly. Two themes pervade the book: techniques for handling long documents, beyond typical sentence-by-sentence processing in NLP, and techniques for addressing the tradeoff between effectiveness (i.e., result quality) and efficiency (e.g., query latency, model and index size). Although transformer architectures and pretraining techniques are recent innovations, many aspects of how they are applied to text ranking are relatively well understood and represent mature techniques. However, there remain many open research questions, and thus in addition to laying out the foundations of pretrained transformers for text ranking, this book also attempts to prognosticate where the field is heading. Codice articolo 9783031010538

Contatta il venditore

Compra nuovo

EUR 85,59
Convertire valuta
Spese di spedizione: EUR 14,99
Da: Germania a: Italia
Destinazione, tempi e costi

Quantità: 1 disponibili

Aggiungi al carrello

Immagini fornite dal venditore

Jimmy Lin
ISBN 10: 3031010531 ISBN 13: 9783031010538
Nuovo Taschenbuch

Da: buchversandmimpf2000, Emtmannsberg, BAYE, Germania

Valutazione del venditore 5 su 5 stelle 5 stelle, Maggiori informazioni sulle valutazioni dei venditori

Taschenbuch. Condizione: Neu. Neuware -The goal of text ranking is to generate an ordered list of texts retrieved from a corpus in response to a query. Although the most common formulation of text ranking is search, instances of the task can also be found in many natural language processing (NLP) applications.This book provides an overview of text ranking with neural network architectures known as transformers, of which BERT (Bidirectional Encoder Representations from Transformers) is the best-known example. The combination of transformers and self-supervised pretraining has been responsible for a paradigm shift in NLP, information retrieval (IR), and beyond. This book provides a synthesis of existing work as a single point of entry for practitioners who wish to gain a better understanding of how to apply transformers to text ranking problems and researchers who wish to pursue work in this area. It covers a wide range of modern techniques, grouped into two high-level categories: transformer models that perform reranking inmulti-stage architectures and dense retrieval techniques that perform ranking directly. Two themes pervade the book: techniques for handling long documents, beyond typical sentence-by-sentence processing in NLP, and techniques for addressing the tradeoff between effectiveness (i.e., result quality) and efficiency (e.g., query latency, model and index size). Although transformer architectures and pretraining techniques are recent innovations, many aspects of how they are applied to text ranking are relatively well understood and represent mature techniques. However, there remain many open research questions, and thus in addition to laying out the foundations of pretrained transformers for text ranking, this book also attempts to prognosticate where the field is heading.Springer Verlag GmbH, Tiergartenstr. 17, 69121 Heidelberg 328 pp. Englisch. Codice articolo 9783031010538

Contatta il venditore

Compra nuovo

EUR 85,59
Convertire valuta
Spese di spedizione: EUR 15,00
Da: Germania a: Italia
Destinazione, tempi e costi

Quantità: 2 disponibili

Aggiungi al carrello

Foto dell'editore

Lin, Jimmy
Editore: Springer 2021-10, 2021
ISBN 10: 3031010531 ISBN 13: 9783031010538
Nuovo PF

Da: Chiron Media, Wallingford, Regno Unito

Valutazione del venditore 4 su 5 stelle 4 stelle, Maggiori informazioni sulle valutazioni dei venditori

PF. Condizione: New. Codice articolo 6666-IUK-9783031010538

Contatta il venditore

Compra nuovo

EUR 81,47
Convertire valuta
Spese di spedizione: EUR 23,00
Da: Regno Unito a: Italia
Destinazione, tempi e costi

Quantità: 10 disponibili

Aggiungi al carrello

Foto dell'editore

Lin, Jimmy; Nogueira, Rodrigo; Yates, Andrew
Editore: Springer, 2021
ISBN 10: 3031010531 ISBN 13: 9783031010538
Nuovo Brossura

Da: Best Price, Torrance, CA, U.S.A.

Valutazione del venditore 5 su 5 stelle 5 stelle, Maggiori informazioni sulle valutazioni dei venditori

Condizione: New. SUPER FAST SHIPPING. Codice articolo 9783031010538

Contatta il venditore

Compra nuovo

EUR 79,37
Convertire valuta
Spese di spedizione: EUR 25,77
Da: U.S.A. a: Italia
Destinazione, tempi e costi

Quantità: 2 disponibili

Aggiungi al carrello

Foto dell'editore

Lin, Jimmy; Nogueira, Rodrigo; Yates, Andrew
Editore: Springer, 2021
ISBN 10: 3031010531 ISBN 13: 9783031010538
Nuovo Brossura

Da: California Books, Miami, FL, U.S.A.

Valutazione del venditore 5 su 5 stelle 5 stelle, Maggiori informazioni sulle valutazioni dei venditori

Condizione: New. Codice articolo I-9783031010538

Contatta il venditore

Compra nuovo

EUR 104,47
Convertire valuta
Spese di spedizione: EUR 7,74
Da: U.S.A. a: Italia
Destinazione, tempi e costi

Quantità: Più di 20 disponibili

Aggiungi al carrello

Foto dell'editore

Lin, Jimmy; Nogueira, Rodrigo; Yates, Andrew
Editore: Springer, 2021
ISBN 10: 3031010531 ISBN 13: 9783031010538
Nuovo Brossura

Da: Books Puddle, New York, NY, U.S.A.

Valutazione del venditore 4 su 5 stelle 4 stelle, Maggiori informazioni sulle valutazioni dei venditori

Condizione: New. 1st edition NO-PA16APR2015-KAP. Codice articolo 26394734880

Contatta il venditore

Compra nuovo

EUR 105,47
Convertire valuta
Spese di spedizione: EUR 7,74
Da: U.S.A. a: Italia
Destinazione, tempi e costi

Quantità: 4 disponibili

Aggiungi al carrello

Foto dell'editore

Lin, Jimmy; Nogueira, Rodrigo; Yates, Andrew
Editore: Springer, 2021
ISBN 10: 3031010531 ISBN 13: 9783031010538
Nuovo Brossura
Print on Demand

Da: Majestic Books, Hounslow, Regno Unito

Valutazione del venditore 5 su 5 stelle 5 stelle, Maggiori informazioni sulle valutazioni dei venditori

Condizione: New. Print on Demand. Codice articolo 401642239

Contatta il venditore

Compra nuovo

EUR 107,82
Convertire valuta
Spese di spedizione: EUR 10,18
Da: Regno Unito a: Italia
Destinazione, tempi e costi

Quantità: 4 disponibili

Aggiungi al carrello

Vedi altre 2 copie di questo libro

Vedi tutti i risultati per questo libro