Which three security measures could be applied in different ML workflow stages to defend them against malicious activities? (Select three.)
Answer(s): B,D,F
Security measures can be applied in different ML workflow stages to defend them against malicious activities, such as data theft, model tampering, or adversarial attacks. Some of the security measures are:Launch ML Instances In a virtual private cloud (VPC): A VPC is a logically isolated section of a cloud provider's network that allows users to launch and control their own resources. By launching ML instances in a VPC, users can enhance the security and privacy of their data and models, as well as restrict the access and traffic to and from the instances. Use data encryption: Data encryption is the process of transforming data into an unreadable format using a secret key or algorithm. Data encryption can protect the confidentiality, integrity, and availability of data at rest (stored in databases or files) or in transit (transferred over networks). Data encryption can prevent unauthorized access, modification, or leakage of sensitive data. Use Secrets Manager to protect credentials: Secrets Manager is a service that helps users securely store, manage, and retrieve secrets, such as passwords, API keys, tokens, or certificates. Secrets Manager can help users protect their credentials from unauthorized access or exposure, as well as rotate them automatically to comply with security policies.
A healthcare company experiences a cyberattack, where the hackers were able to reverse-engineer a dataset to break confidentiality.Which of the following is TRUE regarding the dataset parameters?
Answer(s): B
Overfitting is a problem that occurs when a model learns too much from the training data and fails to generalize well to new or unseen data. Overfitting can result from using a low quantity of training data, a high complexity of the model, or a lack of regularization. Overfitting can also increase the risk of reverse-engineering a dataset from a model's outputs, as the model may reveal too much information about the specific features or patterns of the training data. This can break the confidentiality of the data and expose sensitive information about the individuals in the dataset.
When working with textual data and trying to classify text into different languages, which approach to representing features makes the most sense?
A bag of bigrams (2 letter pairs) is an approach to representing features for textual data that involves counting the frequency of each pair of adjacent letters in a text. For example, the word "hello" would be represented as {"he": 1, "el": 1, "ll": 1, "lo": 1}. A bag of bigrams can capture some information about the spelling and structure of words, which can be useful for identifying the language of a text. For example, some languages have more common bigrams than others, such as "th" in English or "ch" in German .
Which of the following options is a correct approach for scheduling model retraining in a weather prediction application?
Answer(s): C
The input format is the way that the data is structured, organized, and presented to the model. For example, the input format could be a CSV file, an image file, or a JSON object. The input format can affect how the model interprets and processes the data, and therefore how it makes predictions.When the input format changes, it may require retraining the model to adapt to the new format and ensure its accuracy and reliability. For example, if the weather prediction application switches from using numerical values to categorical values for some features, such as wind direction or cloud cover, it may need to retrain the model to handle these changes .
Which of the following tools would you use to create a natural language processing application?
Answer(s): D
NLTK (Natural Language Toolkit) is a Python library that provides a set of tools and resources for natural language processing (NLP). NLP is a branch of AI that deals with analyzing, understanding, and generating natural language texts or speech. NLTK offers modules for various NLP tasks, such as tokenization, stemming, lemmatization, parsing, tagging, chunking, sentiment analysis, named entity recognition, machine translation, text summarization, and more .
Post your Comments and Discuss CertNexus AIP-210 exam dumps with other Community members:
ElastiCache for Redis
S3 Object Lock
S3
SFTP
AWS Transfer Family
Amazon SQS
API Gateway
Lambda
usage plan
AWS WAF
Amazon ECS
Application Load Balancer
AWS Global Accelerator
Network Load Balancer
EC2
Auto Scaling group
CloudFront
ALB
AWS PrivateLink
CRR
SSE-S3
Athena
SSE-KMS
RDS Custom for Oracle
s3:GetObject
Amazon OpenSearch Service
CloudWatch Logs
Kinesis Data Firehose
Kinesis
S3 bucket
SQS
AWS Lambda
AWS Secrets Manager
AWS Systems Manager OpsCenter
secretsmanager:GetSecretValue
seq
for h in {1..254}
for h in $(seq 1 254); do
Kinesis Data Streams
Amazon Redshift
secrets:GetSecretValue
aws:PrincipalOrgID
"aws:PrincipalOrgID": "o-1234567890"
Azure Bot Service
Microsoft.Network/applicationSecurityGroups
Microsoft.Network/bastions
Microsoft.Network
COPY INTO
SELECT
COPY INTO @stage/path/file.csv FROM (SELECT col1, col2 FROM my_table WHERE date >= '2024-01-01') FILE_FORMAT=(TYPE=CSV);
Users
External collaboration settings