Points to be aware of when setting up the GPU and points to be aware of when verifying performance are summarized based on the reference link (https://hiwony.tistory.com/3).
The size of deep learning models is getting bigger and bigger, and the model operating environment is limited by a narrow infrastructure.
What should be considered in order to make a deep learning model a service?
After the deep learning model is created, it is a presentation on what direction we should operate and maintain.
The size of deep learning models is getting bigger and bigger, and the model operating environment is limited by a narrow infrastructure.
What should be considered in order to make a deep learning model a service?
After the deep learning model is created, it is a presentation on what direction we should operate and maintain.
PR-383: Solving ImageNet: a Unified Scheme for Training any Backbone to Top R...Sunghoon Joo
油
Tensorflow KR PR-12 season4 slide
PR-383: Solving ImageNet: a Unified Scheme for Training any Backbone to Top Results Reviewer: Sunghoon Joo (VUNO Inc.)
Paper link: https://arxiv.org/abs/2204.03475
YouTube link: https://youtu.be/WeYuLO1nTmE
Pycon korea 2018 kaggle tutorial(kaggle break)Yeonmin Kim
油
Pycon korea 2018 覩語 磯Μ 貂蠍 tutorial 襭.
https://www.pycon.kr/2018/program/tutorial/13
譯 伎 Coursera How to Win a Data Science Competition:Learn from Top Kagglers .
https://www.coursera.org/learn/competitive-data-science
DeepSeek梶 釈 Trend (Faculty Tae Young Lee)Tae Young Lee
油
The document titled "Trends Observed Through DeepSeek" explores advancements in AI and reinforcement learning through the lens of DeepSeek's latest developments. It is structured into three main sections:
DeepSeek-V3
Focuses on context length extension, initially supporting 32,000 characters and later expanding to 128,000 characters.
Introduces Mixture of Experts (MoE) architecture, optimizing computational efficiency using a novel Auxiliary-Loss-Free Load Balancing strategy.
Multi-Head Latent Attention (MLA) reduces memory consumption while maintaining performance, enhancing large-scale model efficiency.
DeepSeek-R1-Zero
Explores advancements in reinforcement learning algorithms, transitioning from RLHF to GRPO (Group Relative Policy Optimization) for cost-effective optimization.
Direct Preference Optimization (DPO) enhances learning by leveraging preference-based optimization instead of traditional reward functions.
DeepSeek-R1 and Data Attribution
Discusses a Cold Start approach using high-quality data (SFT) to ensure stable initial training.
Incorporates reasoning-focused reinforcement learning, balancing logical accuracy with multilingual consistency.
Utilizes rejection sampling and data augmentation to refine AI-generated outputs for enhanced usability and safety.
The document provides a detailed analysis of these methodologies, positioning DeepSeek as a key player in AI model development and reinforcement learning.
Transitioning from the Era of Big Data to LLMs_Deriving InsightsTae Young Lee
油
Transitioning from the Era of Big Data to LLMs: Deriving Insights
Table of Contents
Big Data and LLMs: Evolution Over Time
Definition and role of Big Data
The emergence and advancements of LLMs (Large Language Models)
Differences and connections between Big Data and LLMs
Challenges of Big Data and the Introduction of LLMs
The initial hype around Big Data and infrastructure expansion
Limitations caused by insufficient data utilization
New possibilities unlocked by the development of LLMs
Current State and Limitations of LLMs
Service innovations brought by LLMs
Gaps between expectations and reality
Data privacy and ethical challenges
Complexity in technology management
A Successful Transition from Big Data to LLMs
Creating business value through data
Shifting from domain-focused to process-oriented thinking
Developing new business models through service innovation
Future Directions for Insight Derivation
Integrating AI and data utilization
Effective approaches to derive actionable insights
Establishing real-time decision-making systems powered by LLMs
Key Messages
Limitations of Big Data: Despite the expansion of data infrastructure, many organizations struggled to translate it into actionable services or insights.
Opportunities with LLMs: LLMs have shown remarkable success in natural language processing and leveraging large-scale data, moving beyond infrastructure to create tangible business value.
Challenges Ahead: Leveraging LLMs requires addressing technical complexity, ethical considerations, and operational costs.
Path to Success: Rather than a technology-centric approach, adopting a problem-solving mindset and developing innovative processes are crucial for success.
Conclusion
The transition from Big Data to LLMs represents a paradigm shift in how data is utilized. Overcoming the challenges of LLM adoption and building a business-driven strategy will pave the way for greater insights and value creation.
This presentation explores these topics with practical examples, offering strategies for using data and AI to shape the future of business.
Facebook Meta's technical direction in Large Language Models (LLMs)Tae Young Lee
油
LLaMA (Large Language Model Meta AI) is a series of large language models developed by Meta (formerly Facebook), designed for natural language processing (NLP) tasks. These models are based on transformer architecture and are trained on extensive datasets, covering a wide range of topics and styles. LLaMA models come in various sizes, catering to tasks from lightweight operations to complex language understanding and generation. Meta emphasizes ethical considerations in developing LLaMA, focusing on reducing biases, ensuring safety, and enhancing transparency. These models can be applied to various NLP tasks such as text completion, question answering, and summarization, and can be fine-tuned for specific industries or needs.
FAISS (Facebook AI Similarity Search) is an open-source library developed by Meta for efficient similarity search and clustering of dense vectors. It is widely used in machine learning and AI applications requiring large-scale data processing and retrieval. FAISS is optimized for both CPU and GPU, enabling rapid processing of large datasets. It supports various indexing methods, including flat indexing, inverted indexing, and product quantization, allowing users to balance accuracy and computational efficiency. The library can scale to billions of vectors, making it suitable for extensive applications, and offers both exact and approximate search methods to trade off between speed and precision based on user needs.
FAISS is commonly used in image and text retrieval, efficiently finding similar items within large datasets, and in recommendation systems to identify similar users or products. It provides a Python API for ease of use and can be integrated with other tools and frameworks, such as PyTorch.
Both LLaMA and FAISS represent Meta's efforts to advance AI technology and its wide range of applications. LLaMA focuses on language understanding and generation, while FAISS is centered on efficient data retrieval and similarity search.
MultiModal Embedding integrates various data types, like images, text, and au...Tae Young Lee
油
MultiModal Embedding refers to a technique used to integrate and process different types of data. "Modality" refers to the type or form of data, such as images, text, audio, etc. MultiModal Embedding maps these different modalities into a common space, allowing for the integration and correlation of diverse types of data.
Key Concepts
Integration of Different Modalities:
It transforms data from various types, such as images, text, and audio, into a common vector space. In this space, each piece of data is represented as a vector, enabling the understanding and analysis of relationships between different modalities.
Common Embedding Space:
It maps data from different modalities into a shared embedding space, allowing for comparison or combination of data across modalities. This process helps capture the features of the data effectively and understand interactions between multiple modalities.
Training and Application:
MultiModal Embedding models are typically trained on large datasets that incorporate various modalities, helping the model learn from a richer set of information. These trained models can be used in applications such as search, recommendation systems, and question-answering.
Applications
Image and Text Integration:
For tasks such as generating descriptions for images or comparing the similarity between images and text.
Multimodal Search:
For performing image searches based on text queries or extracting textual information from images.
Automatic Translation:
For performing speech recognition and translation simultaneously by integrating text and audio.
Enhanced Model Understanding:
Helps models learn more comprehensive and diverse information by leveraging various modalities.
Examples
CLIP (Contrastive Language-Image Pretraining): Developed by OpenAI, this model understands and correlates both images and text, allowing for matching tasks between the two modalities.
DALL-E: An image generation model that creates images from textual descriptions. It operates by converting text and images into a shared embedding space.
MultiModal Embedding enables the integration of diverse data types, contributing to the development of more sophisticated and useful models.
A future that integrates LLMs and LAMs (Symposium)Tae Young Lee
油
Presentation material from the IT graduate school joint event
- Korea University Graduate School of Computer Information and Communication
- Sogang University Graduate School of Information and Communication
- Sungkyunkwan University Graduate School of Information and Communication
- Yonsei University Graduate School of Engineering
- Hanyang University Graduate School of Artificial Intelligence Convergence
Exploring Deep Learning Acceleration Technology Embedded in LLMsTae Young Lee
油
Lab's research presentation
I am a doctoral student at Seoul National University of Science and Technology and am currently the head of the Applying LLMs to Various Industry (AL2VI) Lab.
Course Overview:
This course offers a comprehensive exploration of recommender systems, focusing on both theoretical foundations and practical applications. Through a combination of lectures, hands-on exercises, and real-world case studies, you will gain a deep understanding of the key principles, methodologies, and evaluation techniques that drive effective recommendation algorithms.
Course Objectives:
Acquire a solid understanding of recommender systems, including their significance and impact in various domains.
Explore different types of recommendation algorithms, such as collaborative filtering, content-based filtering, and hybrid approaches.
Study cutting-edge techniques, including deep learning, matrix factorization, and graph-based methods, for enhanced recommendation accuracy.
Gain hands-on experience with popular recommendation frameworks and libraries, and learn how to implement and evaluate recommendation models.
Investigate advanced topics in recommender systems, such as fairness, diversity, and explainability, and their ethical implications.
Analyze and discuss real-world case studies and research papers to gain insights into the challenges and future directions of recommender systems.
Course Structure:
Introduction to Recommender Systems
Collaborative Filtering Techniques
Content-Based Filtering and Hybrid Approaches
Matrix Factorization Methods
Deep Learning for Recommender Systems
Graph-Based Recommendation Approaches
Evaluation Metrics and Experimental Design
Ethical Considerations in Recommender Systems
Fairness, Diversity, and Explainability in Recommendations
Case Studies and Research Trends
Course Delivery:
The course will be delivered through a combination of lectures, interactive discussions, hands-on coding exercises, and group projects. You will have access to state-of-the-art resources, including relevant research papers, datasets, and software tools, to enhance your learning experience.
ChatGPT is a natural language processing technology developed by OpenAI. This model is based on the GPT-3 architecture and can be applied to various language tasks by training on large-scale datasets. When applied to a search engine, ChatGPT enables the implementation of an AI-based conversational system that understands user questions or queries and provides relevant information.
ChatGPT takes user questions as input and generates appropriate responses based on them. Since this model considers the context of previous conversations, it can provide more natural dialogue. Moreover, ChatGPT has been trained on diverse information from the internet, allowing it to provide practical and accurate answers to user questions.
When applying ChatGPT to a search engine, the system searches for relevant information based on the user's search query and uses ChatGPT to generate answers to present along with the search results. To do this, the search engine provides an interface that connects with ChatGPT, allowing the user's questions to be passed to the model and the answers generated by the model to be presented alongside the search results.
Deep Learning Through Various ProcessesTae Young Lee
油
2019 AICON Seminar
This paper examines the effectiveness of applying deep learning through a variety of process (eg, medical and manufacturing steel mills, semiconductor processes, general chatbots, and financial industries) data.
The most important thing in deep learning is the legacy architecture of the system to be applied, and in the manufacturing industry, understanding the process is the most important.
This is because deep learning models applied to processes that improve efficiency and that can be directly linked to productivity gains can generate enormous cost savings. I don't know much about the steel industry I used to, but I compare the processes of the semiconductor industry and consider where deep learning algorithms might be applied.
In addition, in the medical industry, security is important above all, and Federated Learning should be used, and in the case of chatbots, an overall overview of the importance of architectural design according to the Intent Scope and the transfer learning will come out.
2. NAS
Model
Model
Memory Limitation
Inference Speed
Worse Performance
Training Model
Model
Data
Data
Data
1)一危 覲 2)覈 覲
GPU
蠏 伎 螳 一危一 覈碁 れ企り
焔レ レ讌
( 一危一 れ 覲伎 )
Serving Layer Model
旧 gradient 覈語 蠍一 觜襦蠍 覓語
覿 旧 牛 襯 襴朱,
覈語 貉れ 磯 旧 覲企 襷 螳
Training Speed
覈 蠍郁 讀螳覃伎 豢襦 蟇碁Μ 螳
企蠍 覓語 覓語螳
覈語 貉れ覃伎 螳 襾殊 覓語螳
蟆 覃覈襴 伎
Memory
Enhanced AI Platform
覲旧 GPU襯 旧 牛 覓語襯 願屋
讌襷 覿 旧朱 覈語 牛朱 覓語
願屋讌 . ( GPU 蟲襷る NVidia襷 譬 (?) )
GPU
螳讌豺蠍 (Pruning)
螳譴豺 覿 (Weight Factorization)
讌 讀襯 (Knowledge Distillation)
螳譴豺 螻旧 (Weight Sharing)
(Quantization)
Pre-train vs. Downstream
Model 豢 蠍一
3. Model Training
Data Labeling
Model Evaluation
Data Versioning
Model Service
Model Prediction
Model Deployment Model Versioning
Serving Architecture
Legacy Interface
Scaling Hardware
Model Life Cycle 譴
覈 覲 Training Image 覦
Workspace Training
Training Shared Memory れ
Training Multi GPU れ
Model Serving Monitoring
Inference 螻 Gateway
POD Monitoring
Rancher 螳 れ襷朱 Node, Pod
襯 覈磯 蠍磯レ 螻
覲企 誤 襯 蠍 伎
Prometheus, Grafana 煙 覈磯
Data Selection Data Cleaning
Data Pre-Processing
Model MetaData 蟯襴
Model Validation
ML-Metadata
Model Registry
ML-Metadata Model Registry