Horovod, Uber's open source distributed deep learning system, enables NVIDIA to scale model training from one to eight GPUs for their self-driving sensing and perception technologies.
Uber built Michelangelo, our machine learning platform, in 2015. Three years later, we reflect our journey to scaling ML at Uber and lessons learned along the way.
Uber developed Michelangelo PyML to run identical copies of machine learning models locally in both real time experiments and large-scale offline prediction jobs.
To improve our maps, Uber Engineering analyzes customer support tickets with natural language processing and deep learning to identify and correct inaccurate map data.
One-click chat, the Uber driver app's smart reply system, leverages machine learning to make in-app messaging between driver-partners and riders more seamless.
Uber Eats engineers describe how they surface restaurant recommendations in the app using multi-objective optimization to give eaters the most satisfying experience while maintaining the health of the Uber Eats marketplace.
Uber built the next generation of COTA by leveraging deep learning models, thereby scaling the system to provide more accurate customer support ticket predictions.
As powerful and widespread as convolutional neural networks are in deep learning, AI Labs’ latest research reveals both an underappreciated failing and a simple fix.
With a solid margin, Uber senior data scientist Slawek Smyl won the M4 Competition with his hybrid Exponential Smoothing-Recurrent Neural Networks (ES-RNN) forecasting method.
Curious about what it is like to traverse the high-dimensional loss landscapes of modern neural networks? Check out Uber AI Labs’ latest research on measuring intrinsic dimension to find out.
Applying hardware acceleration to deep neuroevolution in what is now an open source project, Uber AI Labs was able to train a neural network to play Atari in just a few hours on a single personal computer, making this type of research accessible to a far greater number of people.
Differentiable Plasticity is a new machine learning method for training neural networks to change their connection weights adaptively even after training is completed, allowing a form of learning inspired by the lifelong plasticity of biological brains.
Uber AI Labs introduces Visual Inspector for Neuroevolution (VINE), an open source interactive data visualization tool to help neuroevolution researchers better understand this family of algorithms.
Uber Engineering created Omphalos, our new backtesting framework, to enable efficient and reliable comparison of forecasting models across languages.
Uber ATG Toronto developed Sparse Blocks Network (SBNet), an open source algorithm for TensorFlow, to speed up inference of our 3D vehicle detection systems while lowering computational costs.
In this article, Uber Engineering introduces our Customer Obsession Ticket Assistant (COTA), a new tool that puts machine learning and natural language processing models in the service of customer care to help agents deliver improved support experiences.
As we approach the New Year, Uber Open Source revisits some of Uber Engineering's most popular projects from 2017.
To ring in the New Year, the Uber Engineering Blog shares some of our editor's picks for 2017.
12Page 1 of 2