10th Indian Delegation to Dubai, Gitex & Expand North Star – World’s Largest Startup Investor Connect
All News

XGBoost 2.0 is Here

The best solution for making sense of tabular data, XGBoost, has just been upgraded. XGBoost 2.0 brings forth a plethora of new features and enhancements aimed at revolutionising the machine learning landscape. 

Click here to check out the full release.

XGBoost 2.0 introduces a novel feature under development, focusing on vector-leaf tree models for multi-target regression, multi-label classification, and multi-class classification. Unlike the previous approach of building separate models for each target, this feature allows XGBoost to construct a single tree for all targets, offering several advantages, including prevention of overfitting, smaller model sizes, and the ability to consider correlations between targets. 

Users can combine vector leaf and scalar leaf trees during training through a callback. It’s important to note that this feature is a work in progress, and some aspects are still under development.

You thought XGBoost was powerful enough?

Introducing: XGBoost 2.

– Multi-target trees (vector-leafs)
– Finally: 𝚍𝚎𝚟𝚒𝚌𝚎 parameter.
– 𝚊𝚙𝚙𝚛𝚘𝚡 for GPU devices.
– Controllable histogram bound (to optimize RAM)
– Batch by batch training (External Memory)
– Learning to…

— Yam Peleg (@Yampeleg) October 3, 2023

Read: XGBoost is All You Need 

New Device Parameter 

A significant change is the introduction of a new ‘device’ parameter, replacing existing parameters like ‘gpu_id,’ ‘gpu_hist,’ ‘gpu_predictor,’ ‘cpu_predictor,’ ‘gpu_coord_descent,’ and the PySpark-specific ‘use_gpu.’ Users can now use the ‘device’ parameter to select their preferred device for computation, simplifying the configuration process.

Default Tree Method

Starting from XGBoost 2.0, the ‘hist’ tree method is set as the default. In previous versions, XGBoost would automatically choose between ‘approx’ and ‘exact’ based on input data and the training environment. The new default method aims to improve model training efficiency and consistency.

GPU-Based Approximate Tree Method 

XGBoost 2.0 offers initial support for the ‘approx’ tree method on GPU. While performance optimisation is ongoing, the feature is considered feature-complete, except for the JVM packages. 

Users can access this capability by specifying ‘device=”cuda”‘ and ‘tree_method=”approx”.’ It’s important to note that the Scala-based Spark interface is not yet supported.

Memory Footprint Optimization 

This release also introduces a new parameter, ‘max_cached_hist_node,’ allowing users to limit CPU cache size for histograms. This helps prevent aggressive caching of histograms, especially in deep trees. Additionally, memory usage for ‘hist’ and ‘approx’ tree methods on distributed systems is reduced by half.

Improved External Memory Support 

External memory support receives a significant boost in XGBoost 2.0. The default ‘hist’ tree method now utilises memory mapping, enhancing performance and reducing CPU memory usage. Users are encouraged to try this feature, particularly when memory savings are required.

Learning-to-Rank Enhancements XGBoost 2.0 introduces a new implementation for learning-to-rank tasks, offering a range of new features and parameters to improve ranking performance. 

Notable additions include parameters for pair construction strategy, control over the number of samples per group, experimental unbiased learning-to-rank support, and custom gain functions with NDCG.

Column-Based Split and Federated Learning Significant progress has been made in column-based split for federated learning, with support for various tree methods and vertical federated learning. GPU support for this feature is still in development.

PySpark Enhancements 

The PySpark interface in XGBoost 2.0 has received numerous new features and optimisations, including GPU-based prediction, data initialisation improvements, support for predicting feature contributions, Python typing support, and improved logs for training.

The post XGBoost 2.0 is Here appeared first on Analytics India Magazine.

by Siliconluxembourg

Would-be entrepreneurs have an extra helping hand from Luxembourg’s Chamber of Commerce, which has published a new practical guide. ‘Developing your business: actions to take and mistakes to avoid’, was written to respond to  the needs and answer the common questions of entrepreneurs.  “Testimonials, practical tools, expert insights and presentations from key players in our ecosystem have been brought together to create a comprehensive toolkit that you can consult at any stage of your journey,” the introduction… Source link

by WIRED

B&H Photo is one of our favorite places to shop for camera gear. If you’re ever in New York, head to the store to check out the giant overhead conveyor belt system that brings your purchase from the upper floors to the registers downstairs (yes, seriously, here’s a video). Fortunately B&H Photo’s website is here for the rest of us with some good deals on photo gear we love. Save on the Latest Gear at B&H Photo B&H Photo has plenty of great deals, including Nikon’s brand-new Z6III full-frame… Source link

by Gizmodo

Long before Edgar Wright’s The Running Man hits theaters this week, the director of Shaun of the Dead and Hot Fuzz had been thinking about making it. He read the original 1982 novel by Stephen King (under his pseudonym Richard Bachman) as a boy and excitedly went to theaters in 1987 to see the film version, starring Arnold Schwarzenegger. Wright enjoyed the adaptation but was a little let down by just how different it was from the novel. Years later, after he’d become a successful… Source link