Index
As this ebook edition doesn't have fixed pagination, the page numbers below are hyperlinked for reference only, based on the printed edition of this book.
B
bag of words
documents, putting into 55-59
BART model 224
BERTopic
reference link 192
topics, visualizing from 187-191
used, for topic modeling 157-162
Best Match 25 (bm25) 217
Bidirectional Encoder Representations from Transformer (BERT)
used, for K-Means topic modeling 153-157
using, instead of word embeddings 73-76
bigram model 59
C
Central Processing Unit (CPU) 199
character n-grams 64
chatbot
creating, with LLM 257-262
classification (CLS) tokens 197
classifier-invariant approach
explainability, enhancing via 228-232
Closed Domain Question Answering (CDQA) 214
CNN DailyMail dataset 224
code generation
with LLM 263-269
community detection clustering
with SBERT 150-152
Compute Unified Device Architecture (CUDA) 198
...