1z0-931-23 Exam QuestionsBrowse all questions from this exam

1z0-931-23 Exam - Question 44


Which stage of the indexing pipeline divides text into tokens?

Show Answer
Correct Answer: AB

The stage of the indexing pipeline that divides text into tokens is the tokenizer. Tokenization is the process of breaking down text into smaller units, typically words or terms, which can then be processed for indexing. The lexer, on the other hand, is typically used in programming language interpreters and compilers, where it divides a stream of text into tokens based on syntax rules. In the context of text indexing, the correct terminology for dividing text into tokens is tokenization, hence the correct answer is the tokenizer.

Discussion

1 comment
Sign in to comment
piipoOption: A
Jun 13, 2024

TOKENIZER