# Tag: Hyperparameter (40)

**Fast AutoML with FLAML + Ray Tune**- Sep 6, 2021.

Microsoft Researchers have developed FLAML (Fast Lightweight AutoML) which can now utilize Ray Tune for distributed hyperparameter tuning to scale up FLAML’s resource-efficient & easily parallelizable algorithms across a cluster.**High-Performance Deep Learning: How to train smaller, faster, and better models – Part 4**- Jul 9, 2021.

With the right software, hardware, and techniques at your fingertips, your capability to effectively develop high-performing models now hinges on leveraging automation to expedite the experimental process and building with the most efficient model architectures for your data.**The 7 Best Open Source AI Libraries You May Not Have Heard Of**- Jun 9, 2021.

AI researchers today have many exciting options for working with specialized tools. Although starting original projects from scratch is often not necessary, knowing which existing library to leverage remains a challenge. This list of generally unknown yet awesome, open-source libraries offers an interesting collection to consider for state-of-the-art research that spans from automatic machine learning to differentiable quantum circuits.**Automating Machine Learning Model Optimization**- Mar 17, 2021.

This articles presents an overview of using Bayesian Tuning and Bandits for machine learning.**4 Machine Learning Concepts I Wish I Knew When I Built My First Model**- Mar 9, 2021.

Diving into building your first machine learning model will be an adventure -- one in which you will learn many important lessons the hard way. However, by following these four tips, your first and subsequent models will be put on a path toward excellence.**Bayesian Hyperparameter Optimization with tune-sklearn in PyCaret**- Mar 5, 2021.

PyCaret, a low code Python ML library, offers several ways to tune the hyper-parameters of a created model. In this post, I'd like to show how Ray Tune is integrated with PyCaret, and how easy it is to leverage its algorithms and distributed computing to achieve results superior to default random search method.**How to Speed up Scikit-Learn Model Training**- Feb 11, 2021.

Scikit-Learn is an easy to use a Python library for machine learning. However, sometimes scikit-learn models can take a long time to train. The question becomes, how do you create the best scikit-learn model in the least amount of time?**My machine learning model does not learn. What should I do?**- Feb 10, 2021.

This article presents 7 hints on how to get out of the quicksand.**Algorithms for Advanced Hyper-Parameter Optimization/Tuning**- Nov 17, 2020.

In informed search, each iteration learns from the last, whereas in Grid and Random, modelling is all done at once and then the best is picked. In case for small datasets, GridSearch or RandomSearch would be fast and sufficient. AutoML approaches provide a neat solution to properly select the required hyperparameters that improve the model’s performance.**How I Consistently Improve My Machine Learning Models From 80% to Over 90% Accuracy**- Sep 23, 2020.

Data science work typically requires a big lift near the end to increase the accuracy of any model developed. These five recommendations will help improve your machine learning models and help your projects reach their target goals.**6 Common Mistakes in Data Science and How To Avoid Them**- Sep 10, 2020.

As a novice or seasoned Data Scientist, your work depends on the data, which is rarely perfect. Properly handling the typical issues with data quality and completeness is crucial, and we review how to avoid six of these common scenarios.**13 must-read papers from AI experts**- May 20, 2020.

What research articles do top AI experts in the field recommend? Find out which ones and why, then be sure to add each to your reading to do list.**Hyperparameter Optimization for Machine Learning Models**- May 7, 2020.

Check out this comprehensive guide to model optimization techniques.**KDnuggets™ News 20:n15, Apr 15: How to Do Hyperparameter Tuning on Any Python Script; 10 Must-read Machine Learning Articles**- Apr 15, 2020.

Learn how to do hyperparameter tuning on python ML scripts; Read 10 must-read Machine Learning Articles; Understand the process for Data Science project review; see how data science is used to understand COVID-19; and stay safe and healthy!**How to Do Hyperparameter Tuning on Any Python Script in 3 Easy Steps**- Apr 8, 2020.

With your machine learning model in Python just working, it's time to optimize it for performance. Follow this guide to setup automated tuning using any optimization library in three steps.**Practical Hyperparameter Optimization**- Feb 13, 2020.

An introduction on how to fine-tune Machine and Deep Learning models using techniques such as: Random Search, Automated Hyperparameter Tuning and Artificial Neural Networks Tuning.**Automate Hyperparameter Tuning for Your Models**- Sep 20, 2019.

When we create our machine learning models, a common task that falls on us is how to tune them. So that brings us to the quintessential question: Can we automate this process?**How to Automate Hyperparameter Optimization**- Jun 12, 2019.

A step-by-step guide into performing a hyperparameter optimization task on a deep learning model by employing Bayesian Optimization that uses the Gaussian Process. We used the gp_minimize package provided by the Scikit-Optimize (skopt) library to perform this task.**Building AI to Build AI: The Project That Won the NeurIPS AutoML Challenge**- Jan 23, 2019.

This is an overview of designing a computer program capable of developing predictive models without any manual intervention that are trained & evaluated in a lifelong machine learning setting in NeurIPS 2018 AutoML3 Challenge.**Keras Hyperparameter Tuning in Google Colab Using Hyperas**- Dec 12, 2018.

In this post, I will show you how you can tune the hyperparameters of your existing keras models using Hyperas and run everything in a Google Colab Notebook.**Mastering the Learning Rate to Speed Up Deep Learning**- Nov 6, 2018.

Figuring out the optimal set of hyperparameters can be one of the most time consuming portions of creating a machine learning model, and that’s particularly true in deep learning.**Implementing Automated Machine Learning Systems with Open Source Tools**- Oct 25, 2018.

What if you want to implement an automated machine learning pipeline of your very own, or automate particular aspects of a machine learning pipeline? Rest assured that there is no need to reinvent any wheels.**The Intuitions Behind Bayesian Optimization with Gaussian Processes**- Oct 19, 2018.

Bayesian Optimization adds a Bayesian methodology to the iterative optimizer paradigm by incorporating a prior model on the space of possible target functions. This article introduces the basic concepts and intuitions behind Bayesian Optimization with Gaussian Processes.**Beginners Ask “How Many Hidden Layers/Neurons to Use in Artificial Neural Networks?”**- Jul 16, 2018.

By the end of this article, you could at least get the idea of how these questions are answered and be able to test yourself based on simple examples.**Deep Learning Tips and Tricks**- Jul 4, 2018.

This post is a distilled collection of conversations, messages, and debates on how to optimize deep models. If you have tricks you’ve found impactful, please share them in the comments below!**Deep Quantile Regression**- Jul 3, 2018.

Most Deep Learning frameworks currently focus on giving a best estimate as defined by a loss function. Occasionally something beyond a point estimate is required to make a decision. This is where a distribution would be useful. This article will purely focus on inferring quantiles.**Batch Normalization in Neural Networks**- Jun 26, 2018.

This article explains batch normalization in a simple way. I wrote this article after what I learned from Fast.ai and deeplearning.ai.**Deep Learning Best Practices – Weight Initialization**- Jun 21, 2018.

In this blog I am going to talk about the issues related to initialization of weight matrices and ways to mitigate them. Before that, let’s just cover some basics and notations that we will be using going forward.**Improving the Performance of a Neural Network**- May 30, 2018.

There are many techniques available that could help us achieve that. Follow along to get to know them and to build your own accurate neural network.**Understanding Learning Rates and How It Improves Performance in Deep Learning**- Feb 1, 2018.

Furthermore, the learning rate affects how quickly our model can converge to a local minima (aka arrive at the best accuracy). Thus getting it right from the get go would mean lesser time for us to train the model.**Using AutoML to Generate Machine Learning Pipelines with TPOT**- Jan 29, 2018.

This post will take a different approach to constructing pipelines. Certainly the title gives away this difference: instead of hand-crafting pipelines and hyperparameter optimization, and performing model selection ourselves, we will instead automate these processes.**Managing Machine Learning Workflows with Scikit-learn Pipelines Part 3: Multiple Models, Pipelines, and Grid Searches**- Jan 24, 2018.

In this post, we will be using grid search to optimize models built from a number of different types estimators, which we will then compare and properly evaluate the best hyperparameters that each model has to offer.**Managing Machine Learning Workflows with Scikit-learn Pipelines Part 2: Integrating Grid Search**- Jan 19, 2018.

Another simple yet powerful technique we can pair with pipelines to improve performance is grid search, which attempts to optimize model hyperparameter combinations.**Is Learning Rate Useful in Artificial Neural Networks?**- Jan 15, 2018.

This article will help you understand why we need the learning rate and whether it is useful or not for training an artificial neural network. Using a very simple Python code for a single layer perceptron, the learning rate value will get changed to catch its idea.**Estimating an Optimal Learning Rate For a Deep Neural Network**- Nov 21, 2017.

This post describes a simple and powerful way to find a reasonable learning rate for your neural network.**Stop Doing Fragile Research**- Nov 17, 2017.

If you develop methods for data analysis, you might only be conducting gentle tests of your method on idealized data. This leads to “fragile research,” which breaks when released into the wild. Here, I share 3 ways to make your methods robust.**A Vision for Making Deep Learning Simple**- Sep 5, 2017.

This post introduces Deep Learning Pipelines from Databricks, a new open-source library aimed at enabling everyone to easily integrate scalable deep learning into their workflows, from machine learning practitioners to business analysts.**The Current State of Automated Machine Learning**- Jan 18, 2017.

What is automated machine learning (AutoML)? Why do we need it? What are some of the AutoML tools that are available? What does its future hold? Read this article for answers to these and other AutoML questions.**Contest Winner: Winning the AutoML Challenge with Auto-sklearn**- Aug 5, 2016.

This post is the first place prize recipient in the recent KDnuggets blog contest. Auto-sklearn is an open-source Python tool that automatically determines effective machine learning pipelines for classification and regression datasets. It is built around the successful scikit-learn library and won the recent AutoML challenge.**TPOT: A Python Tool for Automating Data Science**- May 13, 2016.

TPOT is an open-source Python data science automation tool, which operates by optimizing a series of feature preprocessors and models, in order to maximize cross-validation accuracy on data sets.