RM20e: Hyundais E-Rennauto beschleunigt in 10 Sekunden auf 200 km/h
Hyundai hat mit dem RM20e ein Auto vorgestellt, das zeigen soll, was mit Elektroantrieb und 810 PS im Rennsport möglich ist. (Hyundai, Technologie)
Quelle: Golem
Hyundai hat mit dem RM20e ein Auto vorgestellt, das zeigen soll, was mit Elektroantrieb und 810 PS im Rennsport möglich ist. (Hyundai, Technologie)
Quelle: Golem
Polestar wird das Elektroauto Precept in Serie bauen. Mit an Bord sind Android Automotive, Fischernetze und Getränkeflaschen. (Elektromobilität, GreenIT)
Quelle: Golem
Falls Nutzer dabei unachtsam sind, kann dies alle Daten auf dem Nord-Smartphone von Oneplus löschen. (Oneplus Nord, Smartphone)
Quelle: Golem
medium.com – I was one of the first people that did the initial beta CKA exam three years ago in 2017. Recently in Novemeber 2019, I have passed my CKAD “Certified Kubernetes Application Developer”. Just last wee…
Quelle: news.kubernauts.io
Die elektrische Rennserie E1 wird wie die Formel E, nur auf dem Wasser. (Elektromobilität, Technologie)
Quelle: Golem
Porsche hat einen Karosseriebausatz aus pflanzlichen Fasern für Rennwagen hergestellt, bei dem Teile auf Flachs- und Balsaholzbasis gebaut wurden. (Porsche, GreenIT)
Quelle: Golem
2019 hat Airbus das Aus für den A380 angekündigt. Jetzt ist das letzte Flugzeug dieses Typs gebaut. (Airbus, Technologie)
Quelle: Golem
Wollte Microsoft Windows XP wie Apples Oberfläche Aqua aussehen lassen? (Windows XP, Mac OS X)
Quelle: Golem
Machine learning (ML) is transforming businesses and lives alike. Whether it be finding rideshare partners, recommending products or playlists, identifying objects in images, or optimizing marketing campaigns, ML and prediction is at the heart of these experiences. To support businesses like yours that are revolutionizing the world using ML, AI Platform is committed to providing a world-class, enterprise-ready platform for hosting all of your transformative ML models. As a part of our continued commitment, we are pleased to announce the general availability of AI Platform Prediction based on a Google Kubernetes Engine (GKE) backend. The new backend architecture is designed for improved reliability, more flexibility via new hardware options (Compute Engine machine types and NVIDIA accelerators), reduced overhead latency, and improved tail latency. In addition to standard features such as autoscaling, access logs, and request/response logging available during our Beta period, we’ve introduced several updates that improve robustness, flexibility, and usability:XGBoost / scikit learn models on high-mem/high-cpu machine types: Many data scientists like the simplicity and power of XGBoost and scikit learn models for predictions in production. AI Platform makes it simple to deploy models trained using these frameworks with just a few clicks — we’ll handle the complexity of the serving infrastructure on the hardware of your choice. Resource Metrics: An important part of maintaining models in production is understanding their performance characteristics such as GPU, CPU, RAM, and network utilization. These metrics can help make decisions about what hardware to use to minimize latencies and optimize performance. For example, you can view your model’s replica count over time to help understand how your autoscaling model responds to changes in traffic and alter minReplicas to optimize cost and/or latency. Resource metrics are now visible for models deployed on GCE machine types from Cloud Console and Stackdriver Metrics.Regional Endpoints: We have introduced new endpoints in three regions (us-central1, europe-west4, and asia-east1) with better regional isolation for improved reliability. Models deployed on the regional endpoints stay within the specified region.VPC-Service Controls (Beta): Users can define a security perimeter and deploy Online Prediction models that have access only to resources and services within the perimeter, or within another bridged perimeter. Calls to the CAIP Online Prediction APIs are made from within the perimeter. Private IP will allow VMs and Services within the restricted networks or security perimeters to access the CMLE APIs without having to traverse the public internet.But prediction doesn’t just stop with serving trained models. Typical ML workflows involve analyzing and understanding models and predictions. Our platform integrates with other important AI technologies to simplify your ML workflows and make you more productive:Explainable AI. To better understand your business, you need to better understand your model. Explainable AI provides information about the predictions from each request and is available exclusively on AI Platform.What-if tool. Visualize your datasets and better understand the output of your models deployed on the platform.Continuous Evaluation. Obtain metrics about the performance of your live model based on ground-truth labelling of requests sent to your model. Make decisions to retrain or improve the model based on performance over time.”[AI Platform Prediction] greatly increases our velocity by providing us with an immediate, managed and robust serving layer for our models and allows us to focus on improving our features and modelling,” said Philippe Adjiman, data scientist tech lead at Waze. Read more about Waze’s experience adopting the platform here.All of these features are available in a fully managed, cluster-less environment with enterprise support — no need to stand up or manage your own highly available GKE clusters. We also take care of the quota management and protecting your model from overload from clients sending too much traffic. These features of our managed platform allow your data scientists and engineers to focus on business problems instead of managing infrastructure.Related ArticleHow Waze predicts carpools with Google Cloud’s AI PlatformHow Waze predicts carpools using Google Cloud AI Platform.Read Article
Quelle: Google Cloud Platform
Waze’s mission is to eliminate traffic and we believe our carpool feature is a cornerstone that will help us achieve it. In our carpool apps, a rider (or a driver) is presented with a list of users that are relevant for their commute (see below). From there, the rider or the driver can initiate an offer to carpool, and if the other side accepts it, it’s a match and a carpool is born.Let’s consider a rider who is commuting from somewhere in Tel-Aviv to Google’s offices, as an example that we’ll use throughout this post. Our goal will be to present to that rider a list of drivers that are geographically relevant to her commute, and to rank that list by the highest likelihood of the carpool between that rider and any driver on the list to actually happen. Finding all the relevant candidates in a few seconds involves a lot of engineering and algorithmic challenges, and we’ve dedicated a full team of talented engineers to the task. In this post we’ll focus on the machine learning part of the system responsible for ranking those candidates. In particular:If hundreds (or more) drivers could be a good match for our rider (in our example), how can we build a ML model that would decide which ones to show her first?How can we build the system in a way that allows us to iterate quickly on complex models in production while guaranteeing a low latency online in order to keep the overall user experience fast and delightful?ML models to rank lists of drivers and ridersSo, the rider in our example sees a list of potential drivers. For each such driver, we need to answer two questions:What is the probability that our rider will send this driver a request to carpool?What is the probability that the driver will actually accept the rider’s request?We solve this using machine learning: we build models that estimate those two probabilities based on aggregated historical data of drivers and riders sending and accepting requests to carpool. We use the models to sort drivers from highest to lowest likelihood of the carpool to actually happen.The models we’re using combine close to 90 signals to estimate those probabilities. Below are a few of the most important signals to our models:Star Ratings: higher rated drivers tend to get more requestsWalking distance from pickup and dropoff: riders want to start and end their rides as close as possible to the driver’s route. But, the total walking distance (as seen in the screenshot above) isn’t everything: riders also care about how the walking distance compares to their overall commute length. Consider the two plans below of two different riders: both have 15 minutes walking, but the second one looks much more acceptable given that the commute length is larger to start with, while in the first one, the rider needs to walk as much as the actual carpool length, and is thus much less likely to be interested. The signal that is capturing this in the model and that came up as one of the most important signals, is the ratio between the walking and carpool distance.The same kind of consideration is valid on the driver side, when considering the length of the detour compared to the driver’s full drive from origin to destination.Driver’s intent: One of the most important factors impacting the probability of a driver to accept a request to carpool (sent by a rider) is her intent to actually carpool. We have several signals indicating a driver’s intent, but the one that came up as the most important (as captured by the model) is the last time the driver was seen in the app. The more recent it is, the more likely the driver is to accept a request to carpool sent by a rider.Model vs. Serving complexityIn the early stage of our product, we started with simple logistic regression models to estimate the likelihood of users sending/accepting offers. The models were trained offline using scikit learn. The training set was obtained using a “log and learn” approach (logging signals exactly as they were during serving time) over ~90 different signals, and the learned weights were injected into our serving layer. Although those models were doing a pretty good job, we observed via offline experiments the great potential of more advanced non linear models such as gradient boosted regression classifiers for our ranking task. Implementing an in-memory fast serving layer supporting such advanced models would require non-trivial effort, as well as an on-going maintenance cost. A much simpler option was to delegate the serving layer to an external managed service that can be called via a REST API. However, we needed to be sure that it wouldn’t add too much latency to the overall flow. In order to make our decision, we decided to do a quick POC using the AI Platform Online Prediction service, which sounded like a potential great fit for our needs at the serving layer.A quick (and successful) POCWe trained our gradient boosted models over our ~90 signals using scikit learn, serialized it as a pickle file, and simply deployed it as-is to the Google Cloud AI Platform. Done. We get a fully managed serving layer for our advanced model through a REST API. From there, we just had to connect it to our java serving layer (a lot of important details to make it work, but unrelated to the pure model serving layer). Below is a very high level schema of what our offline/online training/serving architecture looks like. The carpool serving layer is responsible for a lot of logic around computing/fetching the relevant candidates to score, but we focus here on the pure ranking ML part. Google Cloud AI Platform plays a key role in that architecture. It greatly increases our velocity by providing us with an immediate, managed and robust serving layer for our models and allows us to focus on improving our features and modelling.Click to enlargeIncreased velocity and the peace of mind to focus on our core model logic was great, but a core constraint was around the latency added by an external REST API call at the serving layer. We performed various latency checks/load tests against the online prediction API for different models and input sizes. AI Platform provided the low double digit millisecond latency that was necessary for our application. In just a couple of weeks, we were able to implement and connect the components together and deploy the model in production for AB testing. Even though our previous models (a set of logistic regression classifiers) were performing well, we were thrilled to observe significant improvements on our core KPIs in the AB test. But what mattered even more for us, was having a platform to iterate quickly over even more complex models, without having to deal with the training/serving implementation and deployment headaches. The tip of the (Google Cloud AI Platform) icebergIn the future we plan to explore more sophisticated models using Tensorflow, along with Google Cloud’s Explainable AI component that will simplify the development of these sophisticated models by providing deeper insights into how they are performing. AI Platform Prediction’s recent GA release of support for GPUs and multiple high-memory and high-compute instance types will make it easy for us to deploy more sophisticated models in a cost effective way.Based on our early success with the AI Platform Prediction service, we plan to aggressively leverage other compelling components offered by GCP’s AI Platform, such as the Training service w/ hyper parameter tuning, Pipelines, etc. In fact, multiple data science teams and projects (ads, future drive predictions, ETA modelling) at Waze are already using or started exploring other existing (or upcoming) components of the AI Platform. More on that in future posts.Related ArticleAI Platform Prediction goes GA with improved reliability & ML workflow integrationAI Platform Prediction goes GA with enhanced reliability & ML workflow integration.Read Article
Quelle: Google Cloud Platform