Professional Machine Learning Engineer Exam QuestionsBrowse all questions from this exam

Professional Machine Learning Engineer Exam - Question 269


You are developing a model to help your company create more targeted online advertising campaigns. You need to create a dataset that you will use to train the model. You want to avoid creating or reinforcing unfair bias in the model. What should you do? (Choose two.)

Show Answer
Correct Answer: DE

To avoid creating or reinforcing unfair bias in a model, it is essential to properly design the dataset and validate the model's fairness after training. Collecting a stratified sample of production traffic helps ensure that all relevant demographic groups are proportionately represented in the dataset, reducing the risk of underrepresentation and bias (Option D). Additionally, conducting fairness tests across sensitive categories and demographics on the trained model is crucial to identify and address any disparities or biases that the model may exhibit (Option E). These steps help in building and deploying a fair and unbiased model.

Discussion

11 comments
Sign in to comment
b1a8faeOption: D
Jan 22, 2024

D E. ChatGPT explanation below (but I think makes quite a lot of sense) Collect a Stratified Sample (Option D): Stratified sampling involves dividing the population into subgroups (strata) and then randomly sampling from each subgroup. This ensures that the training dataset represents the diversity of the population, helping to avoid biases. By collecting a stratified sample of production traffic, you are more likely to have a balanced representation of different demographic groups, reducing the risk of biased model outcomes. Conduct Fairness Tests (Option E): After training the model, it's crucial to conduct fairness tests to evaluate its performance across different sensitive categories and demographics. This involves measuring the model's predictions and outcomes for various groups to identify any disparities. Fairness tests help you assess and address biases that may have been inadvertently introduced during the training process.

omermahgoubOption: D
Apr 13, 2024

D. Stratified sampling to ensure the different demographic groups or categories are proportionally represented in the training data. This helps mitigate bias that might arise if certain groups are under-represented. E. Fairness tests can reveal disparities in how the model treats different populations, allowing you to identify and address potential biases.

CHARLIE2108Option: D
Feb 18, 2024

I went D, E

MultiCloudIronManOption: D
Apr 6, 2024

D and E is the two answers. Two selections are required

pinimichele01
Apr 9, 2024

why not D and A?

pikachu007Option: D
Jan 13, 2024

D. Stratified Sampling: Randomly sampling your data might not accurately represent the diversity of your target audience, potentially introducing bias by over- or under-representing certain demographics. Stratified sampling ensures your training dataset reflects the distribution of sensitive features (e.g., age, gender, income) observed in your production traffic, helping mitigate bias during model training. E. Fairness Testing: Simply collecting unbiased data isn't enough. Regularly testing your trained model for fairness across sensitive categories is crucial. This involves measuring and analyzing metrics like accuracy, precision, recall, and F1 score for different demographic groups. Identifying disparities in performance can trigger further investigation and potential re-training to address bias.

shadz10Option: C
Jan 18, 2024

C, D - Conducting fairness tests across sensitive categories and demographics on the trained model is indeed important. However, this option focuses on post-training analysis rather than dataset creation. While it's a crucial step for ensuring fairness, it doesn't directly address how to create a training dataset to avoid bias. Hence C,D

tavva_prudhvi
Feb 13, 2024

Check b1a8fae comment on why D is better than C!

daidai75Option: D
Jan 23, 2024

I go for D & E: A stratified sample ensures that the training data represents the distribution of the target population across relevant demographics or other sensitive categories. This helps mitigate bias arising from underrepresented groups in the data. Regularly testing the model for fairness across sensitive categories helps identify and address potential bias issues before deploying the model in production. This can involve metrics like precision, recall, and F1 score for different demographic groups.

guilhermebutzkeOption: D
Feb 17, 2024

DE D. Collect a stratified sample of production traffic to build the training dataset: This ensures that the training data represents the diverse demographics that will be targeted by the advertising campaigns. Random sampling might unintentionally underrepresent certain groups, leading to biased model outputs. E. Conduct fairness tests across sensitive categories and demographics on the trained model: This allows you to identify and address any potential biases that may have emerged during the training process. Evaluating the model's performance on different groups helps ensure fair and responsible deployment.

dija123Option: D
Jul 2, 2024

Agree with D and E

AzureDP900Options: DE
Jul 5, 2024

D and E is right answer, question asks us to select 2 right answers • To avoid creating or reinforcing unfair bias in the model, you should collect a representative and diverse dataset (option D) that includes a stratified sample of production traffic. This ensures that your training data is inclusive and accurately represents the diversity of your target audience. • Once you have collected your training dataset, you should conduct fairness tests across sensitive categories and demographics on the trained model (option E). This involves evaluating whether the model treats different demographic groups fairly and without bias. If biases are detected, you can take steps to mitigate them and ensure that your model is fair and accurate.

AzureDP900Options: DE
Jul 5, 2024

D and E is right answer, question asks us to select 2 right answers • To avoid creating or reinforcing unfair bias in the model, you should collect a representative and diverse dataset (option D) that includes a stratified sample of production traffic. This ensures that your training data is inclusive and accurately represents the diversity of your target audience. • Once you have collected your training dataset, you should conduct fairness tests across sensitive categories and demographics on the trained model (option E). This involves evaluating whether the model treats different demographic groups fairly and without bias. If biases are detected, you can take steps to mitigate them and ensure that your model is fair and accurate.