Skip to main content

Lund University Publications

LUND UNIVERSITY LIBRARIES

Python for Natural Language Processing : Programming with NumPy, scikit-learn, Keras, and PyTorch: Third Edition

Nugues, Pierre M. LU orcid (2024) In Cognitive Technologies Part F3170.
Abstract

Since the last edition of this book (2014), progress has been astonishing in all areas of Natural Language Processing, with recent achievements in Text Generation that spurred a media interest going beyond the traditional academic circles. Text Processing has meanwhile become a mainstream industrial tool that is used, to various extents, by countless companies. As such, a revision of this book was deemed necessary to catch up with the recent breakthroughs, and the author discusses models and architectures that have been instrumental in the recent progress of Natural Language Processing. As in the first two editions, the intention is to expose the reader to the theories used in Natural Language Processing, and to programming examples... (More)

Since the last edition of this book (2014), progress has been astonishing in all areas of Natural Language Processing, with recent achievements in Text Generation that spurred a media interest going beyond the traditional academic circles. Text Processing has meanwhile become a mainstream industrial tool that is used, to various extents, by countless companies. As such, a revision of this book was deemed necessary to catch up with the recent breakthroughs, and the author discusses models and architectures that have been instrumental in the recent progress of Natural Language Processing. As in the first two editions, the intention is to expose the reader to the theories used in Natural Language Processing, and to programming examples that are essential for a deep understanding of the concepts. Although present in the previous two editions, Machine Learning is now even more pregnant, having replaced many of the earlier techniques to process text. Many new techniques build on the availability of text. Using Python notebooks, the reader will be able to load small corpora, format text, apply the models through executing pieces of code, gradually discover the theoretical parts by possibly modifying the code or the parameters, and traverse theories and concrete problems through a constant interaction between the user and the machine. The data sizes and hardware requirements are kept to a reasonable minimum so that a user can see instantly, or at least quickly, the results of most experiments on most machines. The book does not assume a deep knowledge of Python, and an introduction to this language aimed at Text Processing is given in Ch. 2, which will enable the reader to touch all the programming concepts, including NumPy arrays and PyTorch tensors as fundamental structures to represent and process numerical data in Python, or Keras for training Neural Networks to classify texts. Covering topics like Word Segmentation and Part-of-Speech and Sequence Annotation, the textbook also gives an in-depth overview of Transformers (for instance, BERT), Self-Attention and Sequence-to-Sequence Architectures.

(Less)
Please use this url to cite or link to this publication:
author
organization
publishing date
type
Book/Report
publication status
published
subject
keywords
Annotation Schemes, BERT, Information Theory, Keras, Machine Learning, Machine Translation, Named Entity Recognition, Natural Language Processing, Neural Networks, NumPy, Part-of-Speech Tagging, Python, PyTorch, scikit-learn, Text Segmentation, Tokenization, Transformer, Word Embeddings, word2vec
in
Cognitive Technologies
volume
Part F3170
edition
3
pages
512 pages
publisher
Springer Science and Business Media B.V.
external identifiers
  • scopus:85207659007
ISSN
2197-6635
1611-2482
ISBN
978-3-031-57549-5
978-3-031-57548-8
DOI
10.1007/978-3-031-57549-5
language
English
LU publication?
yes
id
c2279413-357f-4c3b-9d9c-91b2eb11bea1
date added to LUP
2024-12-18 11:40:08
date last changed
2025-05-07 21:50:46
@book{c2279413-357f-4c3b-9d9c-91b2eb11bea1,
  abstract     = {{<p>Since the last edition of this book (2014), progress has been astonishing in all areas of Natural Language Processing, with recent achievements in Text Generation that spurred a media interest going beyond the traditional academic circles. Text Processing has meanwhile become a mainstream industrial tool that is used, to various extents, by countless companies. As such, a revision of this book was deemed necessary to catch up with the recent breakthroughs, and the author discusses models and architectures that have been instrumental in the recent progress of Natural Language Processing. As in the first two editions, the intention is to expose the reader to the theories used in Natural Language Processing, and to programming examples that are essential for a deep understanding of the concepts. Although present in the previous two editions, Machine Learning is now even more pregnant, having replaced many of the earlier techniques to process text. Many new techniques build on the availability of text. Using Python notebooks, the reader will be able to load small corpora, format text, apply the models through executing pieces of code, gradually discover the theoretical parts by possibly modifying the code or the parameters, and traverse theories and concrete problems through a constant interaction between the user and the machine. The data sizes and hardware requirements are kept to a reasonable minimum so that a user can see instantly, or at least quickly, the results of most experiments on most machines. The book does not assume a deep knowledge of Python, and an introduction to this language aimed at Text Processing is given in Ch. 2, which will enable the reader to touch all the programming concepts, including NumPy arrays and PyTorch tensors as fundamental structures to represent and process numerical data in Python, or Keras for training Neural Networks to classify texts. Covering topics like Word Segmentation and Part-of-Speech and Sequence Annotation, the textbook also gives an in-depth overview of Transformers (for instance, BERT), Self-Attention and Sequence-to-Sequence Architectures.</p>}},
  author       = {{Nugues, Pierre M.}},
  isbn         = {{978-3-031-57549-5}},
  issn         = {{2197-6635}},
  keywords     = {{Annotation Schemes; BERT; Information Theory; Keras; Machine Learning; Machine Translation; Named Entity Recognition; Natural Language Processing; Neural Networks; NumPy; Part-of-Speech Tagging; Python; PyTorch; scikit-learn; Text Segmentation; Tokenization; Transformer; Word Embeddings; word2vec}},
  language     = {{eng}},
  publisher    = {{Springer Science and Business Media B.V.}},
  series       = {{Cognitive Technologies}},
  title        = {{Python for Natural Language Processing : Programming with NumPy, scikit-learn, Keras, and PyTorch: Third Edition}},
  url          = {{http://dx.doi.org/10.1007/978-3-031-57549-5}},
  doi          = {{10.1007/978-3-031-57549-5}},
  volume       = {{Part F3170}},
  year         = {{2024}},
}