일 | 월 | 화 | 수 | 목 | 금 | 토 |
---|---|---|---|---|---|---|
1 | 2 | |||||
3 | 4 | 5 | 6 | 7 | 8 | 9 |
10 | 11 | 12 | 13 | 14 | 15 | 16 |
17 | 18 | 19 | 20 | 21 | 22 | 23 |
24 | 25 | 26 | 27 | 28 | 29 | 30 |
Tags
- Microsoft
- 데이터사이언스
- 생성형AI
- GenAI
- nlp
- 코드업
- C#
- 파이썬알고리즘
- 알고리즘
- attention
- Python
- 파이썬기초
- 한빛미디어
- 구글퀵랩
- GenerativeAI
- 코드업파이썬
- 파이썬기초100제
- DataScience
- 자연어처리
- 코드업100제
- 클라우드
- 블레이저
- gcp
- 머신러닝
- 파이썬
- 데이터분석
- 빅데이터
- Azure
- codeup
- Blazor
Archives
- Today
- Total
Tech for good
[Google Cloud Skills Boost(Qwiklabs)] Introduction to Generative AI Learning Path - 10. Introduction to Generative AI Studio 본문
IT/Cloud
[Google Cloud Skills Boost(Qwiklabs)] Introduction to Generative AI Learning Path - 10. Introduction to Generative AI Studio
Diana Kang 2023. 9. 9. 22:23https://www.youtube.com/playlist?list=PLIivdWyY5sqIlLF9JHbyiqzZbib9pFt4x
- If you are an app developer or data scientist and want to build an application, you can use Generative AI Studio to quickly prototype and customize generative AI models with no code or low code.
- If you are a data scientist or ML developer who wants to build and automate a generative AI model, you can start from Model Garden to discover and interact with Google's foundation and third-party open-source models and has built-in MLOps tools to automate the ML pipeline.
- Ask to classify instead of generate.
- i.e. Instead of asking what programming language to learn, ask if Python, Java, or C is a better fit for a beginner in programming.
- A simple strategy might be to select the most likely word at every timestep.
Back to the model parameters, temperature is a number used to tune the degree of randomness.
- Low temperature: Select the words that are highly possible and more predictable.
- Best tasks: Q&A, Summarization
- High temperature: Select the words that have low possibility and more unusual.
- Best tasks: Creative works, ...
- In addition, Top K lets the model randomly return a word from the top K number of words in terms of possibility.
- i.e. top 2 => Get a random word from the top 2 possible words (including flowers and trees).
- This approach allows the other high-scoring word a chance of being selected.
- However, if the probability distribution of the words is highly skewed(편향된) and you have one word that is very likely and everything else is very unlikely, this approach can result in some strange responses.
- Top P allows the model to randomly return a word from the Top P probability of words.
- With Top P, you choose from a set of words with the sum of the likelihoods not exceeding P.
- i.e. P = 0.75 -> Sample from a set of words that have a cumulative probability greater than 0.75.
- In this case, it includes three words: flowers, trees, and herbs.
- This way, the size of the set of words can dynamically increase and decrease according to the probability distribution of the next word on the list.
- But producing prompts can be tricky. Small changes in wording or word order can affect the model results in ways that are not totally predictable.
- And you can not really fit all that many examples into a prompt. Even when you do discover a good prompt for your use case, you might notice the quality of model responses is not totally consistent.
- In this scenario, we take a model that was pretrained on a generic dataset.
- We make a copy of this model. Then using those learned weights as a starting point, we re-train the model on a new domain-specfic dataset.
- This technique has been pretty effective for lots of different use cases.
- But when we try to fine tune LLMs, we run into some challenges.
- Compound all of that computation with the hassle and cost of now having to serve this giant model.
- And as a result, fine-tuning a large language model might not be the best option for you.
- These parameters might be a subset of the existing model parameters or they could be an entirely new set of parameters.
- i.e. Adding some additional layers to the model or an extra embedding to the prompt.
Scaling Down to Scale Up: A Guide to Parameter-Efficient Fine-Tuning
- Model name
- the local or Cloud Storage location of your training data.
Parameter-efficient tuning is ideally suited for scenarios
where you have "modest" amounts of training data(100~1000 training examples).
- Your training data should be structured as a supervised training dataset in a text-to-text format.
- Each record or row in the data will contain the input text. In other words, the prompt is followed by the expected output of the model.
- This means that the model can be tuned for a task that can be modeled as a text-to-text problem.
- When the tuning job completes, you'll see the tuned model in the Vertex AI Model Registry and you can deploy it to an endpoint for serving.
Course Review: Reflection Cards
https://storage.googleapis.com/cloud-training/cls-html5-courses/T-GENAISTUDIO-B/index.html#/
* 같이 읽으면 좋은 자료: https://devocean.sk.com/blog/techBoardDetail.do?ID=164779&boardType=techBlog
https://www.thedatahunt.com/trend-insight/what-is-prompt#peurompeuteu-enjinieoring-vt-pain-tyuning