Generative AI E1 | Competency Id 6618 | Quiz Answers
Updated Question Set: click here
Q1: What happens if the token limit is exceeded?
β The model refuses the request
β The model processes the request partially
β The model ignores the token limit
β The model crashes
Answer: The model refuses the request
Q2: What is a one shot prompt?
β A prompt without any examples
β A prompt with multiple examples
β A prompt with only one example
β A prompt with negative examples
Answer: A prompt with only one example
Q3: How can you influence GitHub Copilot's suggestions?
β By providing extensive comments
β By using different function class names
β By adding types to your code
β All of the above
Answer: All of the above
Q4: What is chunking?
β Breaking up a large piece of text into smaller chunks
β Combining multiple texts into a single chunk
β Removing unnecessary tokens from a text
β Rearranging the tokens in a text
Answer: Breaking up a large piece of text into smaller chunks
Q5: How can token limits be avoided?
β By restructuring the initial prompt
β By using chunking techniques
β By summarizing parts of the text
β All of the above
Answer: All of the above
Q6: What is the current credit amount given for api keys?
β 3
β 18
β 20
β 5
Answer: 20
Q7: What is the purpose of retry logic in prompt engineering?
β To prevent hallucinations in the Al's response
β To ensure consistent formatting of the response
β To handle cases where the Al fails to provide the expected format
β To improve the reliability of the Al's responses
Answer: To handle cases where the Al fails to provide the expected format
Q8: How can you access GitHub Copilot?
β Through a web browser
β Through Visual Studio Code
β Through GitHub's website
β Through a mobile app
Answer: Through Visual Studio Code
Q9: What is the role of human evaluation in prompt engineering?
β To determine the reliability of the prompt
β To fine-tune the model for specific use cases
β To eliminate the need for prompt engineering
β To prevent prompt injection attacks
Answer: To determine the reliability of the prompt
Q10: How does prompt engineering contribute to the reliability of Al models?
β It decreases the reliability of Al output
β It has no impact on the reliability of Al output
β It increases the reliability of Al output
β It depends on the complexity of the prompt
Answer: It increases the reliability of Al output
Q11: What is the importance of context length?
β It affects the cost of the request
β It determines the model's output
β It affects the token limit
β It determines the model's accuracy
Answer: It affects the token limit
Q12: What can be specified to change the style of an image generated by the AI?
β The type of style needed
β The format of the response
β The number of retry attempts
β The prompt engineering technique
Answer: The type of style needed
Q13: How can you toggle through, multiple suggestions in GitHub Copilot?
β By pressing the Tab key
β By pressing the Enter key
β By pressing the Esc key
β By pressing the Shift key
Answer: By pressing the Tab key
Q14: What does copilot do when you define a complete function?
β It generates the documentation for the function
β It copies some of the original code from the if statement
β It understands and learns your coding patterns
β It scans your entire project directory
Answer: It understands and learns your coding patterns
Q15: When should prompt engineering be used?
β When creating a product name for personal use
β When rigorously testing a prompt for development and subsequently production
β When evaluating image models
β When using toxic words in a prompt
Answer: When rigorously testing a prompt for development and subsequently production
Q16: What is the recommended approach when providing examples in prompt engineering?
β Give two similar examples to constrain the creative space
β Give multiple examples for better results
β Give negative examples to limit Al output
β Give no examples for better results
Answer: Give two similar examples to constrain the creative space
Q17: How can hallucinations be avoided when using language models?
β By providing factual answers
β By adding extra prompts
β By training the models on complete information
β By avoiding biases in the training data
Answer: By adding extra prompts
Q18: What is the correct order of steps to follow when making a request to the Chatgpt API?
β Select post method, paste URL, select body, select raw, select JSON, provide model and messages
β Select get method, paste URL, select body, select raw, select JSON, provide model and messages
β Select post method, paste URL, select body, select raw, select text, provide model and messages
β Select get method, paste URL, select body, select raw, select text, provide model and messages
Answer: Select post method, paste URL, select body, select raw, select JSON, provide model and messages
Q19: What can you adjust in chat mode to increase variability?
β Maximum length
β Model parameters
β Temperature
β Instructions
Answer: Temperature
Q20: What are neural networks?
β Algorithms that analyze big data
β Algorithms that simulate like the human brain
β Algorithms that create new datasets
β Algorithms that solve complex equations
Answer: Algorithms that simulate like the human brain
Post a comment
Get your FREE PDF on "100 Ways to Try ChatGPT Today"
Generating link, please wait for: 60 seconds
Comments
Join the conversation and share your thoughts! Leave the first comment.