The document discusses the rights of data subjects under the EU GDPR, particularly regarding automated decision-making and profiling. It outlines conditions under which such decisions can be made, emphasizing the need for measures that protect the data subjects' rights and freedoms. Additionally, it includes references to various machine learning and artificial intelligence interpretability frameworks and studies.
【DL輪読会】Efficiently Modeling Long Sequences with Structured State SpacesDeep Learning JP
?
This document summarizes a research paper on modeling long-range dependencies in sequence data using structured state space models and deep learning. The proposed S4 model (1) derives recurrent and convolutional representations of state space models, (2) improves long-term memory using HiPPO matrices, and (3) efficiently computes state space model convolution kernels. Experiments show S4 outperforms existing methods on various long-range dependency tasks, achieves fast and memory-efficient computation comparable to efficient Transformers, and performs competitively as a general sequence model.
* Satoshi Hara and Kohei Hayashi. Making Tree Ensembles Interpretable: A Bayesian Model Selection Approach. AISTATS'18 (to appear).
arXiv ver.: https://arxiv.org/abs/1606.09066#
* GitHub
https://github.com/sato9hara/defragTrees
Amazon SageMaker is a fully managed service that enables developers and data scientists to build, train, and deploy machine learning (ML) models quickly. It provides algorithms, notebooks, APIs and scalable infrastructure for building ML models. Some key features of SageMaker include algorithms for common ML tasks, notebooks for developing models, APIs for training and deployment, and scalable infrastructure for training and hosting models. It also integrates with other AWS services like S3, EC2 and VPC.
The document discusses the Optuna hyperparameter optimization framework, highlighting its features like define-by-run, pruning, and distributed optimization. It provides examples of successful applications in competitions and introduces the use of LightGBM hyperparameter tuning. Additionally, it outlines the installation procedure, key components of Optuna, and the introduction of the lightgbmtuner for automated optimization.
【DL輪読会】Efficiently Modeling Long Sequences with Structured State SpacesDeep Learning JP
?
This document summarizes a research paper on modeling long-range dependencies in sequence data using structured state space models and deep learning. The proposed S4 model (1) derives recurrent and convolutional representations of state space models, (2) improves long-term memory using HiPPO matrices, and (3) efficiently computes state space model convolution kernels. Experiments show S4 outperforms existing methods on various long-range dependency tasks, achieves fast and memory-efficient computation comparable to efficient Transformers, and performs competitively as a general sequence model.
* Satoshi Hara and Kohei Hayashi. Making Tree Ensembles Interpretable: A Bayesian Model Selection Approach. AISTATS'18 (to appear).
arXiv ver.: https://arxiv.org/abs/1606.09066#
* GitHub
https://github.com/sato9hara/defragTrees
Amazon SageMaker is a fully managed service that enables developers and data scientists to build, train, and deploy machine learning (ML) models quickly. It provides algorithms, notebooks, APIs and scalable infrastructure for building ML models. Some key features of SageMaker include algorithms for common ML tasks, notebooks for developing models, APIs for training and deployment, and scalable infrastructure for training and hosting models. It also integrates with other AWS services like S3, EC2 and VPC.
The document discusses the Optuna hyperparameter optimization framework, highlighting its features like define-by-run, pruning, and distributed optimization. It provides examples of successful applications in competitions and introduces the use of LightGBM hyperparameter tuning. Additionally, it outlines the installation procedure, key components of Optuna, and the introduction of the lightgbmtuner for automated optimization.
This document discusses using ARIMA models with BigQuery ML to analyze time series data. It provides an overview of time series data and ARIMA models, including how ARIMA models incorporate AR and MA components as well as differencing. It also demonstrates how to create an ARIMA prediction model and visualize results using BigQuery ML and Google Data Studio. The document concludes that ARIMA models in BigQuery ML can automatically select the optimal order for time series forecasting and that multi-variable time series are not yet supported.
This document discusses using BigQuery as the central part of an ML data pipeline from ETL to model creation to visualization. It introduces BigQuery and BigQuery ML, showing how ETL jobs can load data from Cloud Storage into BigQuery for analysis and model training. Finally, it demonstrates this process by loading a sample CSV dataset into BigQuery and using BigQuery ML to create and evaluate a prediction model.
8. Vertex AI
MLOps with Vertex AI
https://cloud.google.com/vertex-ai?hl=ja#section-7
*各セルのサービス名称,関係変わっているかも
- Google I/O 21にて発表(一般提供)
- 機械学習ワークフローの統合環境
- AutoML含む各種MLツールへのアクセス
- AutoML
- Vertex AI Workbench
- Deep Learning VM Image
- Vertex AI Data Labeling
- Vertex Explainable AI
- Vertex AI Model Monitoring
- …
15. 参考文献?リンク
コードはこちら
https://github.com/kootr/ml-study-session
Vertex AI AutoML ForecastingでDNNの強力な多変量時系列予測を自動モデリングしてしまおう
Vertex AI ではじめる時系列分析入門
How to build forecasting models with Vertex AI
ML 入門: Vertex AI のラーニングパス
Using AutoML for Time Series Forecasting
GCP: AutoML
AutoMLで破産しないように気をつけたいポイント
GCPのAutoMLを使っていたら12万の請求?がきてしまった話
Summary of rules for identifying ARIMA models