Call Us

Home / Blog / Data Science / Auto-WEKA: Automatic Model Selection and Hyperparameter Optimization in WEKA

Auto-WEKA: Automatic Model Selection and Hyperparameter Optimization in WEKA

  • October 13, 2022
  • 4540
  • 76
Author Images

Meet the Author : Mr. Bharani Kumar

Bharani Kumar Depuru is a well known IT personality from Hyderabad. He is the Founder and Director of Innodatatics Pvt Ltd and 360DigiTMG. Bharani Kumar is an IIT and ISB alumni with more than 17 years of experience, he held prominent positions in the IT elites like HSBC, ITC Infotech, Infosys, and Deloitte. He is a prevalent IT consultant specializing in Industrial Revolution 4.0 implementation, Data Analytics practice setup, Artificial Intelligence, Big Data Analytics, Industrial IoT, Business Intelligence and Business Management. Bharani Kumar is also the chief trainer at 360DigiTMG with more than Ten years of experience and has been making the IT transition journey easy for his students. 360DigiTMG is at the forefront of delivering quality education, thereby bridging the gap between academia and industry.

Read More >

Auto-WEKA, which is created to help users by automatically fetching through the space of WEKA’s learning algorithms and respective hyperparameter techniques to improve the performance, using different optimization methods.

Progressively, researchers of Machine Learning tools who  are non-experts requiring off-the-rack solutions?

The Machine Learning community has brought available a wide variety of powerful learning algorithms and feature selection methods through open-source packages, such as WEKA and mlr. These packages give an option for the user to make choices such as: selecting a learning algorithm and setting hyperparameters to the model. It is challenging to pick the right choice, so many times users leave to select algorithms based on reputation or set hyperparameters to default values. 

Learn the core concepts of Data Science Course video on Youtube:

Certainly, adopting such an approach will yield performance far worse than that of the best method and hyperparameter settings.

A likely explanation is that it is very challenging to search the combined space of learning algorithms and their hyperparameters: the response function is noisy and the space can be highly dimensional, involving both categorical and continuous choices, and on top of it contains hierarchical dependencies.

Another related concept of work is on meta-learning procedures that extract characteristics of the dataset, such as the computation of so-called landmarking algorithms, to predict which algorithm or hyperparameter configuration will perform well. 

 The most preferable is Bayesian optimization procedures.

To demonstrate the feasibility of an automatic approach to solving many problems, we built Auto-WEKA, which solves problems for the learners and feature engineers which are implemented in the WEKA Machine Learning package. 

Meta-methods take a single base classifier and its parameters as input, and ensemble methods take any number of base learners as input. These ensemble methods are amazing where we have many settings related to tuning hyperparameters. All ML algorithms do not apply to all datasets as reasons can be anything. For any given dataset, our Auto-WEKA implementation automatically only considers the subset of applicable learners.

From: Auto-WEKA: Automatic Model Selection and Hyperparameter Optimization in WEKA

Auto-WEKA which works with tuning hyperparameter values moves straight up to the fidelity of the machine. We can highlight that this combined hyperparameter space is far bigger than an easy union of the bottom learner's hyperparameter spaces since the ensemble methods allow 5 independent base learners. The meta, ensemble methods together with feature selection contribute to moving forward to the entire size of AutoWEKA’s hyperparameter space. 

CIFAR-10-Small is a subset of CIFAR-10, where we have 50,000 training data points.

An approach that the user might take is to perform a 5-fold cross-validation on the training set for every technique with unmodified hyperparameters, and choose the classifier with the smallest average misclassification error across folds.

An Auto-WEKA is amazing at optimizing its given objective function however, this cannot be enough to conclude that it fits models that work well. Because the number of hyperparameters of a machine learning algorithm grows, so does its potential for overfitting. The cross-validation will considerably increase the robustness of Autoweka against overfitting, but as its hyperparameter tuning factor is more far-reaching than that of any other normal classification algorithms, it's very important to analyze whether overfitting poses a controversy in such a scenario.


The automated process of picking algorithms, regulating the hyperparameters, monotonous modeling building, and evaluation of models built are going to be the next phase of Auto ML. We can say that " it can't be neither an automatic data science nor automated development in AI", it is concluded as “transforming model building” Currently, selecting the “best” algorithm to use as per dataset requires a level of intuition or expertise about the information. Data scientists leverage their experience to experiment with different combinations of models and hyperparameter values to attain the best accuracy.Click Here Data Science Course

AutoML will lessen our dependence on intuition by iteratively trying out an algorithm, scoring its performance, and selecting and refining other models. In other words, it'll automate the machine learning process of the information science workflow.

AutoML will become mainstream and help to accelerate the model-building process.

Data Science Placement Success Story

Data Science Training Institutes in Other Locations

Agra, Ahmedabad, Amritsar, Anand, Anantapur, Bangalore, Bhopal, Bhubaneswar, Chengalpattu, Chennai, Cochin, Dehradun, Malaysia, Dombivli, Durgapur, Ernakulam, Erode, Gandhinagar, Ghaziabad, Gorakhpur, Gwalior, Hebbal, Hyderabad, Jabalpur, Jalandhar, Jammu, Jamshedpur, Jodhpur, Khammam, Kolhapur, Kothrud, Ludhiana, Madurai, Meerut, Mohali, Moradabad, Noida, Pimpri, Pondicherry, Pune, Rajkot, Ranchi, Rohtak, Roorkee, Rourkela, Shimla, Shimoga, Siliguri, Srinagar, Thane, Thiruvananthapuram, Tiruchchirappalli, Trichur, Udaipur, Yelahanka, Andhra Pradesh, Anna Nagar, Bhilai, Borivali, Calicut, Chandigarh, Chromepet, Coimbatore, Dilsukhnagar, ECIL, Faridabad, Greater Warangal, Guduvanchery, Guntur, Gurgaon, Guwahati, Hoodi, Indore, Jaipur, Kalaburagi, Kanpur, Kharadi, Kochi, Kolkata, Kompally, Lucknow, Mangalore, Mumbai, Mysore, Nagpur, Nashik, Navi Mumbai, Patna, Porur, Raipur, Salem, Surat, Thoraipakkam, Trichy, Uppal, Vadodara, Varanasi, Vijayawada, Visakhapatnam, Tirunelveli, Aurangabad

Data Analyst Courses in Other Locations

ECIL, Jaipur, Pune, Gurgaon, Salem, Surat, Agra, Ahmedabad, Amritsar, Anand, Anantapur, Andhra Pradesh, Anna Nagar, Aurangabad, Bhilai, Bhopal, Bhubaneswar, Borivali, Calicut, Cochin, Chengalpattu , Dehradun, Dombivli, Durgapur, Ernakulam, Erode, Gandhinagar, Ghaziabad, Gorakhpur, Guduvanchery, Gwalior, Hebbal, Hoodi , Indore, Jabalpur, Jaipur, Jalandhar, Jammu, Jamshedpur, Jodhpur, Kanpur, Khammam, Kochi, Kolhapur, Kolkata, Kothrud, Ludhiana, Madurai, Mangalore, Meerut, Mohali, Moradabad, Pimpri, Pondicherry, Porur, Rajkot, Ranchi, Rohtak, Roorkee, Rourkela, Shimla, Shimoga, Siliguri, Srinagar, Thoraipakkam , Tiruchirappalli, Tirunelveli, Trichur, Trichy, Udaipur, Vijayawada, Vizag, Warangal, Chennai, Coimbatore, Delhi, Dilsukhnagar, Hyderabad, Kalyan, Nagpur, Noida, Thane, Thiruvananthapuram, Uppal, Kompally, Bangalore, Chandigarh, Chromepet, Faridabad, Guntur, Guwahati, Kharadi, Lucknow, Mumbai, Mysore, Nashik, Navi Mumbai, Patna, Pune, Raipur, Vadodara, Varanasi, Yelahanka


Navigate to Address

360DigiTMG - Data Science, Data Scientist Course Training in Bangalore

No 23, 2nd Floor, 9th Main Rd, 22nd Cross Rd, 7th Sector, HSR Layout, Bengaluru, Karnataka 560102


Get Direction: Data Science Course

Success Stories
Make an Enquiry